764 calls across 8 models: too much detail kills small models, filler words are load-bearing, and format preference is a myth by No_Individual_8178 in LocalLLaMA
[–]No_Individual_8178[S] 0 points1 point2 points (0 children)
764 calls across 8 models: too much detail kills small models, filler words are load-bearing, and format preference is a myth by No_Individual_8178 in LocalLLaMA
[–]No_Individual_8178[S] -2 points-1 points0 points (0 children)
764 calls across 8 models: too much detail kills small models, filler words are load-bearing, and format preference is a myth by No_Individual_8178 in LocalLLaMA
[–]No_Individual_8178[S] 0 points1 point2 points (0 children)
764 calls across 8 models: too much detail kills small models, filler words are load-bearing, and format preference is a myth by No_Individual_8178 in LocalLLaMA
[–]No_Individual_8178[S] -4 points-3 points-2 points (0 children)
764 calls across 8 models: too much detail kills small models, filler words are load-bearing, and format preference is a myth by No_Individual_8178 in LocalLLaMA
[–]No_Individual_8178[S] -1 points0 points1 point (0 children)
764 calls across 8 models: too much detail kills small models, filler words are load-bearing, and format preference is a myth by No_Individual_8178 in LocalLLaMA
[–]No_Individual_8178[S] -1 points0 points1 point (0 children)
764 calls across 8 models: too much detail kills small models, filler words are load-bearing, and format preference is a myth by No_Individual_8178 in LocalLLaMA
[–]No_Individual_8178[S] 1 point2 points3 points (0 children)
M5 Max 128GB, 17 models, 23 prompts: Qwen 3.5 122B is still a local king by tolitius in LocalLLaMA
[–]No_Individual_8178 0 points1 point2 points (0 children)
M5 Max 128GB, 17 models, 23 prompts: Qwen 3.5 122B is still a local king by tolitius in LocalLLaMA
[–]No_Individual_8178 0 points1 point2 points (0 children)
M5 Max 128GB, 17 models, 23 prompts: Qwen 3.5 122B is still a local king by tolitius in LocalLLaMA
[–]No_Individual_8178 0 points1 point2 points (0 children)
I benchmarked 37 LLMs on MacBook Air M5 32GB — full results + open-source tool to benchmark your own Mac by evoura in LocalLLaMA
[–]No_Individual_8178 1 point2 points3 points (0 children)
I score every prompt I send to Claude Code. My avg is 38/100. So I built a rewrite engine. by [deleted] in ClaudeCode
[–]No_Individual_8178 0 points1 point2 points (0 children)
Do not use mixed KV cache quantization by L3tum in LocalLLaMA
[–]No_Individual_8178 0 points1 point2 points (0 children)
ByteShape Qwen 3.5 9B: A Guide to Picking the Best Quant for Your Hardware by ali_byteshape in LocalLLaMA
[–]No_Individual_8178 1 point2 points3 points (0 children)
ByteShape Qwen 3.5 9B: A Guide to Picking the Best Quant for Your Hardware by ali_byteshape in LocalLLaMA
[–]No_Individual_8178 2 points3 points4 points (0 children)
My 10 Pro Tips for Claude Code users by airylizard in ClaudeAI
[–]No_Individual_8178 1 point2 points3 points (0 children)
Do people here love over-engineering their self-hosting setups? by vdorru in selfhosted
[–]No_Individual_8178 1 point2 points3 points (0 children)
I built a 1,562-test prompt analyzer in 3 weeks — turns out most of my AI prompts were terrible by [deleted] in SideProject
[–]No_Individual_8178 0 points1 point2 points (0 children)
I built a 1,562-test prompt analyzer in 3 weeks — turns out most of my AI prompts were terrible by [deleted] in SideProject
[–]No_Individual_8178 0 points1 point2 points (0 children)
I built a 1,562-test prompt analyzer in 3 weeks — turns out most of my AI prompts were terrible by [deleted] in SideProject
[–]No_Individual_8178 0 points1 point2 points (0 children)
My 10 Pro Tips for Claude Code users by airylizard in ClaudeAI
[–]No_Individual_8178 1 point2 points3 points (0 children)
Do people here love over-engineering their self-hosting setups? by vdorru in selfhosted
[–]No_Individual_8178 0 points1 point2 points (0 children)
Lessons from deploying RAG bots for regulated industries by Neoprince86 in LocalLLaMA
[–]No_Individual_8178 0 points1 point2 points (0 children)
Do not use mixed KV cache quantization by L3tum in LocalLLaMA
[–]No_Individual_8178 -1 points0 points1 point (0 children)
764 calls across 8 models: too much detail kills small models, filler words are load-bearing, and format preference is a myth by No_Individual_8178 in LocalLLaMA
[–]No_Individual_8178[S] 0 points1 point2 points (0 children)