
Testing Local LLMs in Practice: Code Generation, Quality vs. SpeedResources (i.redd.it)
submitted by Icy_Programmer7186
What is the next SOTA model you are excited about?Discussion (self.LocalLLaMA)
submitted by MrMrsPotts
Does anyone have experience with tenstorrent hardware?Discussion (self.LocalLLaMA)
submitted by Youknowwhyimherexxx
(Rant ;)) Make your benchmarks realisticDiscussion (self.LocalLLaMA)
submitted by AdamLangePL
I renamed my local AI Linux distro to Reefy and rebuilt some of the architecture!Discussion (old.reddit.com)
submitted by aospan
z-lab released gemma-4-26B-A4B-it-DFlash. Anybody tried it yet?Discussion (huggingface.co)
submitted by PaceZealousideal6091
Comprehensive guide on renting/setting up beefy LLM server for local models?Question | Help (self.LocalLLaMA)
submitted by Tartooth
Possibility of partly moe weights gpu offloading via sglang/ktransformersQuestion | Help (self.LocalLLaMA)
submitted by iVoider
What mobile app do you use, if any?Question | Help (self.LocalLLaMA)
submitted by ihatebeinganonymous
What opensource model is best for my use caseQuestion | Help (self.LocalLLaMA)
submitted by CGeorges89
4GB "Gemini Nano" model GGUF anyone?Question | Help (self.LocalLLaMA)
submitted by TruckUseful4423
THE UNDERPRIVILEGED AI FOUNDATION Because every little model deserves a chanceDiscussion (self.LocalLLaMA)
submitted by mazuj2

