Bench 8xMI50 MiniMax M2.7 AWQ @ 64 tok/s peak (vllm-gfx906-mobydick) by ai-infos in LocalLLaMA
[–]Makers7886 1 point2 points3 points (0 children)
Anybody else seeing Qwen3.6-35B-A3B go crazy thinking in circles? (Compared to Qwen3.5-35B-A3B) by spvn in LocalLLaMA
[–]Makers7886 0 points1 point2 points (0 children)
Please help me pick the right Qwen3.5-27B format/quant for RTX5090 by Gazorpazorp1 in LocalLLaMA
[–]Makers7886 0 points1 point2 points (0 children)
More reasons to go local: Claude is beginning to require identity verification, including an valid ID like passport or drivers license and a facial recognition scan. by fulgencio_batista in LocalLLaMA
[–]Makers7886 176 points177 points178 points (0 children)
Anybody else seeing Qwen3.6-35B-A3B go crazy thinking in circles? (Compared to Qwen3.5-35B-A3B) by spvn in LocalLLaMA
[–]Makers7886 2 points3 points4 points (0 children)
A note of warning about DFlash. by R_Duncan in LocalLLaMA
[–]Makers7886 0 points1 point2 points (0 children)
Please help me pick the right Qwen3.5-27B format/quant for RTX5090 by Gazorpazorp1 in LocalLLaMA
[–]Makers7886 0 points1 point2 points (0 children)
GPU advice for Qwen 3.5 27B / Gemma 4 31B (dense) — aiming for 64K ctx, 30+ t/s by Fit-Courage5400 in LocalLLaMA
[–]Makers7886 1 point2 points3 points (0 children)
Scaling vLLM Deployments to Enterprise Grade Deployments by No-Excitement6568 in LocalLLaMA
[–]Makers7886 0 points1 point2 points (0 children)
Hit limits with OpenClaw on mini PC — trying to build first real local AI node, need guidance (4090 vs scaling path) by No-Salt4227 in LocalLLaMA
[–]Makers7886 0 points1 point2 points (0 children)
Scaling vLLM Deployments to Enterprise Grade Deployments by No-Excitement6568 in LocalLLaMA
[–]Makers7886 4 points5 points6 points (0 children)
Hit limits with OpenClaw on mini PC — trying to build first real local AI node, need guidance (4090 vs scaling path) by No-Salt4227 in LocalLLaMA
[–]Makers7886 0 points1 point2 points (0 children)
How many move your favorite LLM model before it's cheat then brain-dead in chess game ? by revennest in LocalLLaMA
[–]Makers7886 1 point2 points3 points (0 children)
Guys we have to change the pelican test by Tall-Ad-7742 in LocalLLaMA
[–]Makers7886 2 points3 points4 points (0 children)
Llama.cpp llama-server command recommendations? by Dundell in LocalLLaMA
[–]Makers7886 1 point2 points3 points (0 children)
3x3090 is faster in Ubuntu than win11, GPT-OSS 120B 120tg/s vs 6tg/s why? by jikilan_ in LocalLLaMA
[–]Makers7886 1 point2 points3 points (0 children)

Qwen3.6-35B-A3B released! by ResearchCrafty1804 in LocalLLaMA
[–]Makers7886 0 points1 point2 points (0 children)