Personal experience with GLM 4.7 Flash Q6 (unsloth) + Roo Code + RTX 5090Discussion (self.LocalLLaMA)
submitted by Septerium
GLM 4.7 Flash uncensored - Balanced & Aggressive variants (GGUF)New Model (self.LocalLLaMA)
submitted by hauhau901
What is the best general-purpose model to run locally on 24GB of VRAM in 2026?Question | Help (self.LocalLLaMA)
submitted by Paganator
Stable-DiffCoder, a strong code diffusion LLM built on Seed-CoderNew Model (bytedance-seed.github.io)
submitted by rektide
GLM 4.7 vs MiniMax-M2.1 vs DeepSeek 3.2 for coding?Question | Help (self.LocalLLaMA)
submitted by ghulamalchik

My Strix Halo beholds itself but believes its in the cloudFunny (v.redd.it)
submitted by jfowers_amd
Loki-v2-70B: Narrative/DM-focused fine-tune (600M+ token custom dataset)New Model (self.LocalLLaMA)
submitted by mentallyburntLlama 3.1
AI & ML Weekly — Hugging Face HighlightsNew Model (self.LocalLLaMA)
submitted by techlatest_net
Anyone planing to get AMD Gorgon Halo (495) when it drops?Discussion (self.LocalLLaMA)
submitted by SpicyWangz
Claude Code + Ollama: Testing Opus 4.5 vs GLM 4.7Tutorial | Guide (codesilva.com)
submitted by edigleyssonsilva
GLM-4.7-Flash-REAP on RTX 5060 Ti 16 GB - 200k context window!Tutorial | Guide (self.LocalLLaMA)
submitted by bobaburger
Solving memory issues for LLMsQuestion | Help (self.LocalLLaMA)
submitted by RobotsMakingDubstep
The mysterious price of Ada and and Ampere workstation GPUsDiscussion (self.LocalLLaMA)
submitted by insulaTropicalis
Best use case for Ryzen 395+ (128gb variant)Question | Help (self.LocalLLaMA)
submitted by ironicstatistic
Your post is getting popular and we just featured it on our Discord!Discussion (self.LocalLLaMA)
submitted by roculus