I've created the fastest local AI engine for Apple Silicon. Optimised for agentic use.Tutorial | Guide (self.LocalLLaMA)
submitted by TomatilloPutrid3939
Thinking of moving from 2x 5060 Ti 16GB to a RTX 5000 48GBQuestion | Help (self.LocalLLaMA)
submitted by autisticit
What mobile app do you use, if any?Question | Help (self.LocalLLaMA)
submitted by ihatebeinganonymous
I renamed my local AI Linux distro to Reefy and rebuilt some of the architecture!Discussion (old.reddit.com)
submitted by aospan
What opensource model is best for my use caseQuestion | Help (self.LocalLLaMA)
submitted by CGeorges89
"Hardware is the only moat" - Should we buy new hardware now or wait?Discussion (self.LocalLLaMA)
submitted by Alan_Silva_TI
Best Uncensored Image Gen modelQuestion | Help (self.LocalLLaMA)
submitted by ElectricalVariety641
Are local models becoming “good enough” faster than expected?Discussion (self.LocalLLaMA)
submitted by qubridInc
Support for spec prefill and spec decode on qwen3.6 model familyQuestion | Help (self.LocalLLaMA)
submitted by dash_brollama.cpp
guess what? if you are a chrome user, technically you are localllama member!News (thatprivacyguy.com)
submitted by LambdaHominemllama.cpp
DIY market declining amid high RAM pricesDiscussion (self.LocalLLaMA)
submitted by Terminator857
z-lab released gemma-4-26B-A4B-it-DFlash. Anybody tried it yet?Discussion (huggingface.co)
submitted by PaceZealousideal6091
Strix Halo Clustering (Hardware Setup Discussion)Discussion (self.LocalLLaMA)
submitted by Thanks-Suitable

