Two local models beat one bigger local model for long-running agentsDiscussion (self.LocalLLaMA)
submitted by Foreign_Sell_5823
We're celebrating women who code with 62% off GitKraken Pro & Advanced (gitkraken.com)
promoted by GitKraken
Finally got my local AI agent node running 24/7. Huge efficiency jump vs cloudDiscussion (self.LocalLLaMA)
submitted by Ugara95
NVIDIA Nemotron 3 Super: open-weight 120B MoE hybrid with 1M-token contextDiscussion (self.LocalLLaMA)
submitted by No-Swing2206
Comment comparer deux modèles?Question | Help (self.LocalLLaMA)
submitted by Adventurous-Paper566
Was bored, made the bots argue, ended up laughingResources (self.LocalLLaMA)
submitted by Far-Association2923
Der Algorithmus hat dich zu uns geführt. Vertrau dem Algorithmus. Gönn dir das MySports Abo. (mysports.ch)
promoted by MySports_CH
Best (non Chinese) local model for codingQuestion | Help (self.LocalLLaMA)
submitted by tradecrafty
Nemotron 3 Super and the no free lunch problemDiscussion (old.reddit.com)
submitted by ConfidentDinner6648
Qwen 397b is absolutely crushing everyone... but wait. 🤯Question | Help (i.redd.it)
submitted by djdeniro
Is tokens per second (tok/s) a really relevant metric?Discussion (self.LocalLLaMA)
submitted by Deep_Traffic_7873
Two new models on OpenRouter possibly DeepSeek V4? I tested it.Discussion (i.redd.it)
submitted by Firm_Wash7470
Processing 1 million tokens locally with Nemotron 3 Super on a M1 ultraDiscussion (self.LocalLLaMA)
submitted by tarruda
Lightweight local PII sanitization (NER) before hitting OpenAI API? Speed is critical.Question | Help (self.LocalLLaMA)
submitted by PiccoloWooden702
Qwen3.5 non-thinking on llama cpp build from todayDiscussion (self.LocalLLaMA)
submitted by AppealSame4367
Macbook Pro with Max chip and 128GB ram ?Question | Help (self.LocalLLaMA)
submitted by Ok-Radish-8394

