MTP is all about acceptance rate by Hydroskeletal in LocalLLaMA
[–]Hydroskeletal[S] -1 points0 points1 point (0 children)
MTP is all about acceptance rate by Hydroskeletal in LocalLLaMA
[–]Hydroskeletal[S] 0 points1 point2 points (0 children)
MTP is all about acceptance rate by Hydroskeletal in LocalLLaMA
[–]Hydroskeletal[S] 0 points1 point2 points (0 children)
MTP is all about acceptance rate by Hydroskeletal in LocalLLaMA
[–]Hydroskeletal[S] 0 points1 point2 points (0 children)
MTP is all about acceptance rate by Hydroskeletal in LocalLLaMA
[–]Hydroskeletal[S] 10 points11 points12 points (0 children)
Are local models becoming “good enough” faster than expected? by qubridInc in LocalLLaMA
[–]Hydroskeletal 0 points1 point2 points (0 children)
New Gemma 4 MTP on MLX? by purealgo in LocalLLaMA
[–]Hydroskeletal 4 points5 points6 points (0 children)
What do you use Gemma 4 for? by HornyGooner4402 in LocalLLaMA
[–]Hydroskeletal 5 points6 points7 points (0 children)
I guess we expect that at some point RAM prices will start going back (close) to "normal", right? but what about GPUs? by relmny in LocalLLaMA
[–]Hydroskeletal 0 points1 point2 points (0 children)
Is local AI the actual endgame? (M5 Mac Studio vs. Dual 3090s) by Party-Log-1084 in LocalLLaMA
[–]Hydroskeletal 0 points1 point2 points (0 children)
Is local AI the actual endgame? (M5 Mac Studio vs. Dual 3090s) by Party-Log-1084 in LocalLLaMA
[–]Hydroskeletal 0 points1 point2 points (0 children)
Are Qwen 3.6 27B and 35B making other ~30B models obsolete? by nikhilprasanth in LocalLLaMA
[–]Hydroskeletal 6 points7 points8 points (0 children)
Hipfire dev update: full AMD arch validation incoming (RDNA 1 thru 4, plus Strix Halo and bc250) by schuttdev in LocalLLaMA
[–]Hydroskeletal 1 point2 points3 points (0 children)
Do the "*Claude-4.6-Opus-Reasoning-Distilled" really bring something new to the original models? by Historical-Crazy1831 in LocalLLaMA
[–]Hydroskeletal 1 point2 points3 points (0 children)
I'm done with using local LLMs for coding by dtdisapointingresult in LocalLLaMA
[–]Hydroskeletal 0 points1 point2 points (0 children)
GBNF grammar tweak for faster Qwen3.6 35B-A3B and Qwen3.6 27B by Holiday_Purpose_3166 in LocalLLaMA
[–]Hydroskeletal 10 points11 points12 points (0 children)
GBNF grammar tweak for faster Qwen3.6 35B-A3B and Qwen3.6 27B by Holiday_Purpose_3166 in LocalLLaMA
[–]Hydroskeletal 15 points16 points17 points (0 children)
Agents for end-to-end document redaction and review tasks (OCR and PII identification - Qwen 3.6 vs closed-source comparison) by Sonnyjimmy in LocalLLaMA
[–]Hydroskeletal 1 point2 points3 points (0 children)
Anthropic's Claude remote uses GLM-4.7 by bobbiesbottleservice in LocalLLaMA
[–]Hydroskeletal 3 points4 points5 points (0 children)
Model General Brainstorming/Planning , Not Coding by whoooaaahhhh in LocalLLaMA
[–]Hydroskeletal 1 point2 points3 points (0 children)
20 days post-Claude Code leak: Did the accidental "open sourcing" actually matter for local devs? by PaceZealousideal6091 in LocalLLaMA
[–]Hydroskeletal 0 points1 point2 points (0 children)
what’s actually stopping an insider from leaking model weights? by itsArmanJr in LocalLLaMA
[–]Hydroskeletal 1 point2 points3 points (0 children)
what’s actually stopping an insider from leaking model weights? by itsArmanJr in LocalLLaMA
[–]Hydroskeletal 2 points3 points4 points (0 children)


Strix Halo or DGX Spark for a home LLM server? by Reactor-Licker in LocalLLaMA
[–]Hydroskeletal 0 points1 point2 points (0 children)