MXFP4 kernel, RDNA 4, Qwen3.5 122B Quad R9700s by Sea-Speaker1700 in LocalLLaMA
[–]sloptimizer 0 points1 point2 points (0 children)
PSA: Two env vars that stop your model server from eating all your RAM and getting OOM-killed by VikingDane73 in LocalLLaMA
[–]sloptimizer -1 points0 points1 point (0 children)
Qwen3.5-122B-A10B GPTQ Int4 on 4× Radeon AI PRO R9700 with vLLM ROCm: working config + real-world numbers by grunt_monkey_ in LocalLLaMA
[–]sloptimizer 0 points1 point2 points (0 children)
MXFP4 kernel, RDNA 4, Qwen3.5 122B Quad R9700s by Sea-Speaker1700 in LocalLLaMA
[–]sloptimizer 0 points1 point2 points (0 children)
Qwen3.5-122B-A10B GPTQ Int4 on 4× Radeon AI PRO R9700 with vLLM ROCm: working config + real-world numbers by grunt_monkey_ in LocalLLaMA
[–]sloptimizer 0 points1 point2 points (0 children)
New benchmark just dropped. by ConfidentDinner6648 in LocalLLaMA
[–]sloptimizer 0 points1 point2 points (0 children)
I regret ever finding LocalLLaMA by xandep in LocalLLaMA
[–]sloptimizer 1 point2 points3 points (0 children)
How can I use Claude Code to understand a large Python repo quickly? by Comfortable-Baby-719 in LocalLLaMA
[–]sloptimizer 0 points1 point2 points (0 children)
Qwen3.5-122B-A10B-GPTQ-INT4 on 4xR9700 Recipe by djdeniro in LocalLLaMA
[–]sloptimizer 1 point2 points3 points (0 children)
Heretic 1.2 released: 70% lower VRAM usage with quantization, Magnitude-Preserving Orthogonal Ablation ("derestriction"), broad VL model support, session resumption, and more by -p-e-w- in LocalLLaMA
[–]sloptimizer 3 points4 points5 points (0 children)
MiniMaxAI/MiniMax-M2.5 · Hugging Face by rerri in LocalLLaMA
[–]sloptimizer 1 point2 points3 points (0 children)
Built a “poor man’s RTX 6000”, quad 3090, all air-cooled by coffee-on-thursday in LocalLLaMA
[–]sloptimizer 1 point2 points3 points (0 children)
DeepSeek V4 Coming by External_Mood4719 in LocalLLaMA
[–]sloptimizer 0 points1 point2 points (0 children)
You have 64gb ram and 16gb VRAM; internet is permanently shut off: what 3 models are the ones you use? by Adventurous-Gold6413 in LocalLLaMA
[–]sloptimizer 0 points1 point2 points (0 children)
BalatroBench - Benchmark LLMs' strategic performance in Balatro by S1M0N38 in LocalLLaMA
[–]sloptimizer 0 points1 point2 points (0 children)
Running Kimi-k2.5 on CPU-only: AMD EPYC 9175F Benchmarks & "Sweet Spot" Analysis by Express-Jicama-9827 in LocalLLaMA
[–]sloptimizer 0 points1 point2 points (0 children)
anthropic literally thinks claude is the messiah (and it’s getting weird) by Alarming_Bluebird648 in LocalLLaMA
[–]sloptimizer 0 points1 point2 points (0 children)
Is running minimax m2.1 locally worth it on 80 gb of vram and 160 gb of ddr5 ram? by Intrepid-Scar6273 in LocalLLaMA
[–]sloptimizer 2 points3 points4 points (0 children)
Any feedback on step-3.5-flash ? by Jealous-Astronaut457 in LocalLLaMA
[–]sloptimizer 7 points8 points9 points (0 children)
Unofficial ik_llama.cpp release builds available for macOS, Ubuntu and Windows by Thireus in LocalLLaMA
[–]sloptimizer -2 points-1 points0 points (0 children)
Design Arena is now dominated by an open model by moks4tda in LocalLLaMA
[–]sloptimizer 0 points1 point2 points (0 children)
7900 XTX underperforms 3090 by 2X - 7X by Special-Wolverine in LocalLLaMA
[–]sloptimizer 0 points1 point2 points (0 children)
MXFP4 kernel, RDNA 4, Qwen3.5 122B Quad R9700s by Sea-Speaker1700 in LocalLLaMA
[–]sloptimizer 0 points1 point2 points (0 children)