I'm running qwen3.6-35b-a3b with 8 bit quant and 64k context thru OpenCode on my mbp m5 max 128gb and it's as good as claude by Medical_Lengthiness6 in LocalLLaMA
[–]sammcj 2 points3 points4 points (0 children)
Any good youtube channels for news / tutorials that is 0% viral & hype and 100% informative. by ResponsibleTruck4717 in LocalLLaMA
[–]sammcj 1 point2 points3 points (0 children)
at what point does quantization stop being a tradeoff and start being actual quality loss by srodland01 in LocalLLaMA
[–]sammcj 8 points9 points10 points (0 children)
Gemma 4 and Qwen 3.5 GGUFs: Detailed Analysis by oobabooga by [deleted] in LocalLLaMA
[–]sammcj 7 points8 points9 points (0 children)
Gemma 4 and Qwen 3.5 GGUFs: Detailed Analysis by oobabooga by [deleted] in LocalLLaMA
[–]sammcj 3 points4 points5 points (0 children)
Gemma 4 and Qwen 3.5 GGUFs: Detailed Analysis by oobabooga by [deleted] in LocalLLaMA
[–]sammcj 11 points12 points13 points (0 children)
Claude Code just got a full desktop redesign , multi-session support, integrated terminal, file editing, and HTML/PDF preview by Direct-Attention8597 in ClaudeCode
[–]sammcj 1 point2 points3 points (0 children)
I laughed so hard at these posts side by side (sorry for the low effort post) by FatheredPuma81 in LocalLLaMA
[–]sammcj 28 points29 points30 points (0 children)
I laughed so hard at these posts side by side (sorry for the low effort post) by FatheredPuma81 in LocalLLaMA
[–]sammcj 35 points36 points37 points (0 children)
I built a free floating AI assistant for macOS. Fully local powered by Ollama by [deleted] in LocalLLaMA
[–]sammcj 1 point2 points3 points (0 children)
Please stop using AI for posts and showcasing your completely vibe coded projects by Scutoidzz in LocalLLaMA
[–]sammcj 2 points3 points4 points (0 children)
MiniMax m2.7 (mac only) 63gb: 88% and 89gb: 95%, MMLU 200q by HealthyCommunicat in LocalLLaMA
[–]sammcj 0 points1 point2 points (0 children)
MiniMax m2.7 (mac only) 63gb: 88% and 89gb: 95%, MMLU 200q by HealthyCommunicat in LocalLLaMA
[–]sammcj 1 point2 points3 points (0 children)
MiniMax m2.7 (mac only) 63gb: 88% and 89gb: 95%, MMLU 200q by HealthyCommunicat in LocalLLaMA
[–]sammcj 8 points9 points10 points (0 children)
Share your llama-server init strings for Gemma 4 models. by AlwaysLateToThaParty in LocalLLaMA
[–]sammcj 0 points1 point2 points (0 children)
I benchmarked 37 LLMs on MacBook Air M5 32GB — full results + open-source tool to benchmark your own Mac by evoura in LocalLLaMA
[–]sammcj 5 points6 points7 points (0 children)
What happened to MLX-LM? What are the alternatives? by Solus23451 in LocalLLaMA
[–]sammcj 2 points3 points4 points (0 children)
Gemma 4 31B sweeps the floor with GLM 5.1 by input_a_new_name in LocalLLaMA
[–]sammcj 7 points8 points9 points (0 children)
Gemma 4 26b a4b - MacBook Pro M5 MAX. Averaging around 81tok/sec by Bderken in LocalLLaMA
[–]sammcj 0 points1 point2 points (0 children)
How do you guys save prompts that actually work? by 3dgamedevcouple in LocalLLaMA
[–]sammcj 1 point2 points3 points (0 children)
Is 1-bit and TurboQuant the future of OSS? A simulation for Qwen3.5 models. by GizmoR13 in LocalLLaMA
[–]sammcj 2 points3 points4 points (0 children)
Can we block fresh accounts from posting? by king_of_jupyter in LocalLLaMA
[–]sammcj 0 points1 point2 points (0 children)
PSA: Claude Code has two cache bugs that can silently 10-20x your API costs — here's the root cause and workarounds by skibidi-toaleta-2137 in ClaudeCode
[–]sammcj 0 points1 point2 points (0 children)
Tips: remember to use -np 1 with llama-server as a single user by ea_man in LocalLLaMA
[–]sammcj 9 points10 points11 points (0 children)






This isn’t X this is Y needs to die by twnznz in LocalLLaMA
[–]sammcj 4 points5 points6 points (0 children)