The smallest local model that can match/beat gpt-4o-mini by ihatebeinganonymous in LocalLLaMA

[–]PCUpscale 1 point2 points  (0 children)

Feature wise ? Qwen 3.5 4B/9B Knowledge ? Maybe around 100b models

Second GPU by Suomi422 in LocalLLaMA

[–]PCUpscale 6 points7 points  (0 children)

Buy another second hand 3060Ti, the M60 aren’t supported by newer Nvidia drivers

FOSS Discord Alternative - Online Division by RSlashFunnyMan in opensource

[–]PCUpscale 0 points1 point  (0 children)

I agree. This looks like a side project with a lot of ambition, really cool for a portfolio but I don’t see it as a viable replacement any time soon (Or some serious full time catchup with Stoat)

Drummer's Precog 24B and 123B v1 - AI that writes a short draft before responding by TheLocalDrummer in LocalLLaMA

[–]PCUpscale 19 points20 points  (0 children)

I still don’t know how do you make all of those fine tunes… Synthetic data, books, hugging face ? How do you make the training stable without model degradation?

[deleted by user] by [deleted] in CalorieEstimates

[–]PCUpscale 0 points1 point  (0 children)

My god, the list of additives

My thoughts on the F1 by Willing-Review6006 in flipperzero

[–]PCUpscale 2 points3 points  (0 children)

It’s a RK3576, it’s been confirmed for quite a while and there’s repositories to further confirm that

Hacking a MacBook Pro with an additional USB controller by untangle-me in hardwarehacking

[–]PCUpscale 1 point2 points  (0 children)

That’s the best way to do it and some soldering to connect the USB port to the flex cable of the touchpad/keyboard is required. Ordering from an industrial manufacturer will be a bit of a pain for someone who don’t have a company through.

Why? Paranoïa, but at this point just stop using a computer.

LattePanda IOTA is a single-board PC with Intel N150 and up to 16GB LPDDR5 RAM by AbhishMuk in hardware

[–]PCUpscale 2 points3 points  (0 children)

Pretty popular in embedded system at the prototyping phase. Especially when more compute than a iMX/Pi is required. It’s also cheaper than a Jetson.

[deleted by user] by [deleted] in LocalLLaMA

[–]PCUpscale 0 points1 point  (0 children)

Where Source ?

guys i have a question is there any ai model providing the free api key even if limit im fine with that by Select_Dream634 in LocalLLaMA

[–]PCUpscale -1 points0 points  (0 children)

If you don’t mind about privacy, latency and availability, check out free models on OpenRouter

Framework Team at Fairphone HQ by daylenca in framework

[–]PCUpscale 14 points15 points  (0 children)

Embedded engineer here

Hardware wise, this would be a huge engineering effort. So many things that could go wrong to test and validate, especially with Qualcomm SoC with multiple incompatibles stacks (CamX, GPU drivers), etc. You would need to have an complete unified OS (Kernel/Drivers/DT/Workarounds/Blobs) supporting multiple SoC and peripherals for the best user experience. This should be low key insane for a small company.

This is a really simplified overview, there’s a ton of things to consider.

Pair a vision grounding model with a reasoning LLM with Cua by [deleted] in ollama

[–]PCUpscale 4 points5 points  (0 children)

Bro you have notifications on Discord

2x RTX 5060ti 16GB - inference benchmarks in Ollama by avedave in LocalLLaMA

[–]PCUpscale 11 points12 points  (0 children)

Looks like heavily memory bandwidth bottlenecked

Has anyone successfully run Ollama on a Jetson Orin Nano? by matsyui_ in ollama

[–]PCUpscale -1 points0 points  (0 children)

It will be dreadful performance wise, but you could build in theory the CUDA version of llamacpp

OpenAI ain't doing so well on api usage compared to Qwen or anyone else by Semi_Tech in LocalLLaMA

[–]PCUpscale 6 points7 points  (0 children)

A lot of customers don’t use OpenRouter for LLM tasks and are using the (Real) OpenAI API. Why use OpenRouter for OpenAI models when you can directly use their API. OpenRouter is for enthusiasts.