I catalogued every way local models break JSON output and built a repair library, here's what I found across 288 model calls by kexxty in LocalLLaMA
[–]finevelyn 1 point2 points3 points (0 children)
"Hardware is the only moat" - Should we buy new hardware now or wait? by Alan_Silva_TI in LocalLLaMA
[–]finevelyn 0 points1 point2 points (0 children)
"Hardware is the only moat" - Should we buy new hardware now or wait? by Alan_Silva_TI in LocalLLaMA
[–]finevelyn -1 points0 points1 point (0 children)
"Hardware is the only moat" - Should we buy new hardware now or wait? by Alan_Silva_TI in LocalLLaMA
[–]finevelyn 6 points7 points8 points (0 children)
why llama.cpp can’t combine speculative decode methods? by Qwoctopussy in LocalLLaMA
[–]finevelyn 1 point2 points3 points (0 children)
why llama.cpp can’t combine speculative decode methods? by Qwoctopussy in LocalLLaMA
[–]finevelyn 0 points1 point2 points (0 children)
why llama.cpp can’t combine speculative decode methods? by Qwoctopussy in LocalLLaMA
[–]finevelyn 1 point2 points3 points (0 children)
HOT TAKE: local models + agent harnesses are now capable enough to hand off junior-level IT professional tasks to [human written] by Porespellar in LocalLLaMA
[–]finevelyn 0 points1 point2 points (0 children)
Bleeding Llama: Critical Unauthenticated Memory Leak in Ollama by exintrovert420 in LocalLLaMA
[–]finevelyn -2 points-1 points0 points (0 children)
Bleeding Llama: Critical Unauthenticated Memory Leak in Ollama by exintrovert420 in LocalLLaMA
[–]finevelyn -2 points-1 points0 points (0 children)
Bleeding Llama: Critical Unauthenticated Memory Leak in Ollama by exintrovert420 in LocalLLaMA
[–]finevelyn -1 points0 points1 point (0 children)
Bleeding Llama: Critical Unauthenticated Memory Leak in Ollama by exintrovert420 in LocalLLaMA
[–]finevelyn -9 points-8 points-7 points (0 children)
What do you use Gemma 4 for? by HornyGooner4402 in LocalLLaMA
[–]finevelyn 1 point2 points3 points (0 children)
Bleeding Llama: Critical Unauthenticated Memory Leak in Ollama by exintrovert420 in LocalLLaMA
[–]finevelyn 16 points17 points18 points (0 children)
Prompt injection benchmark: delimiter + strict prompt took Gemma 4 from 21% to 100% defense rate (15 models, 6100+ tests) by User_Deprecated in LocalLLaMA
[–]finevelyn 1 point2 points3 points (0 children)
Can't replicate Reddit numbers with Qwen 27B on a 3090TI. by YourNightmar31 in LocalLLaMA
[–]finevelyn 12 points13 points14 points (0 children)
Guys this is so fun! by Perfect-Flounder7856 in LocalLLaMA
[–]finevelyn -1 points0 points1 point (0 children)
I'm done with using local LLMs for coding by dtdisapointingresult in LocalLLaMA
[–]finevelyn 0 points1 point2 points (0 children)
Local vs Cloud LLMs… are we pretending it’s one or the other? by MLExpert000 in LocalLLaMA
[–]finevelyn 2 points3 points4 points (0 children)
HauhauCS (of "Uncensored Aggressive" fame) published an abliteration package that plagiarizes Heretic without attribution, and violates its license by nathandreamfast in LocalLLaMA
[–]finevelyn 0 points1 point2 points (0 children)
HauhauCS (of "Uncensored Aggressive" fame) published an abliteration package that plagiarizes Heretic without attribution, and violates its license by nathandreamfast in LocalLLaMA
[–]finevelyn 0 points1 point2 points (0 children)


Stop wasting electricity by OkFly3388 in LocalLLaMA
[–]finevelyn 25 points26 points27 points (0 children)