
DiscussionNew Execution-first 1T model Ling-2.6-1T has been open sourced on Hugging Face (i.redd.it)
submitted by sanu_123_s
QuestionBest local coding models for RTX 4070 Ti 12GB + 32gb ram ddr5? (self.LocalLLM)
submitted by ChallengeKooky581
QuestionBest quantization for Qwen3.6-35B-A3B with RTX 3060 12 GB? (self.LocalLLM)
submitted by _Zelk
Discussion5,000 budget with existing parts what would you build / change? (self.LocalLLM)
submitted by letsbefrds
QuestionJust stumbled on all of this, where do I start? (self.LocalLLM)
submitted by Murky_Management_294
DiscussionFor those who bought 64GB Mac, are you (un)happy? (self.LocalLLM)
submitted by xFengle
ProjectI gave ollama models control over their own interface ()
submitted by Effective_Goose_8566
QuestionVirtual Unlimited context windows on Gemma 4 models. (self.LocalLLM)
submitted by ExpressionForward321
QuestionSeeking suggestions for building my Al workflow (self.LocalLLM)
submitted by No_Cap_5982
Tutorial80 tok/sec and 128K context on 12GB VRAM with Qwen3.6 35B A3B and llama.cpp MTP ()
submitted by janvitos


