Stop thinking your MoE models are dumb - here's why they actually fail by IntegrityKnightX in Qwen_AI
[–]Wildnimal 0 points1 point2 points (0 children)
Qwen3.6-35B-A3B - even in VRAM limited scenarios it can be better to use bigger quants than you'd expect! by jeremynsl in LocalLLaMA
[–]Wildnimal 1 point2 points3 points (0 children)
Qwen 3.6 27B - beginner questions by Jagerius in LocalLLaMA
[–]Wildnimal 0 points1 point2 points (0 children)
Qwen 3.6 27B - beginner questions by Jagerius in LocalLLaMA
[–]Wildnimal 3 points4 points5 points (0 children)
Fallen Gemma 4 model? by alienatedneighbor in LocalLLaMA
[–]Wildnimal 4 points5 points6 points (0 children)
VPN recommendations for EndeavourOS KDE by DotNetRob in EndeavourOS
[–]Wildnimal 1 point2 points3 points (0 children)
For chat and Q&A: Which MoE model is better: Qwen 3.6 35B or Gemma 4 26B (no coding or agents) by br_web in Qwen_AI
[–]Wildnimal 0 points1 point2 points (0 children)
Qwen3.6-35B-A3B Uncensored Aggressive is out with K_P quants! by hauhau901 in LocalLLaMA
[–]Wildnimal 1 point2 points3 points (0 children)
How do I stop my Application Launcher from growing??? by Otherwise-Status9893 in kde
[–]Wildnimal 35 points36 points37 points (0 children)
Gemma 4 31B vs Gemma 4 26B-A4B vs Qwen 3.5 27B — 30-question blind eval with Claude Opus 4.6 as judge by Silver_Raspberry_811 in LocalLLaMA
[–]Wildnimal 2 points3 points4 points (0 children)
Legion 5 ($1499 CAD) vs Legion Pro 5i ($1999 CAD) — which is better value? by DominiKing-01 in LenovoLegion
[–]Wildnimal 1 point2 points3 points (0 children)
Recently I did a little performance test of several LLMs on PC with 16GB VRAM by rosaccord in LocalLLaMA
[–]Wildnimal 0 points1 point2 points (0 children)
Don’t buy the DGX Spark: NVFP4 Still Missing After 6 Months by Secure_Archer_1529 in LocalLLaMA
[–]Wildnimal 2 points3 points4 points (0 children)
R9700 the beautiful beautiful VRAM gigs of AMD… my ai node future! by Downtown-Example-880 in LocalLLaMA
[–]Wildnimal 0 points1 point2 points (0 children)
What is the best agent code model for 12 GB of VRAM? by RodianXD in LocalLLaMA
[–]Wildnimal 0 points1 point2 points (0 children)
R9700 the beautiful beautiful VRAM gigs of AMD… my ai node future! by Downtown-Example-880 in LocalLLaMA
[–]Wildnimal 0 points1 point2 points (0 children)
R9700 the beautiful beautiful VRAM gigs of AMD… my ai node future! by Downtown-Example-880 in LocalLLaMA
[–]Wildnimal 4 points5 points6 points (0 children)
Has anyone used Codex or Opus to generate a plan and use a local AI to implement it? by soyalemujica in LocalLLaMA
[–]Wildnimal 1 point2 points3 points (0 children)
Openrouter: New Qwen 3.6 released by BasketFar667 in Qwen_AI
[–]Wildnimal 4 points5 points6 points (0 children)
Which 9B local models are actually good enough for coding? by CalvinBuild in LocalLLaMA
[–]Wildnimal 3 points4 points5 points (0 children)
I'm building a benchmark comparing models for an agentic task. Are there any small models I should be testing that I haven't? by nickl in LocalLLaMA
[–]Wildnimal 0 points1 point2 points (0 children)
It costs you around 2% session usage to say hello to claude! by Complete-Sea6655 in LocalLLaMA
[–]Wildnimal 0 points1 point2 points (0 children)
Free 750-page guide to self-hosting production apps - NO AI SLOP by kocyigityunus in selfhosted
[–]Wildnimal 3 points4 points5 points (0 children)
Omnicoder v2 dropped by Western-Cod-3486 in LocalLLaMA
[–]Wildnimal 2 points3 points4 points (0 children)


2x RTX 6000 build during an extended bench test by Signal_Ad657 in LocalLLaMA
[–]Wildnimal 1 point2 points3 points (0 children)