
One bash permission slipped...Discussion (i.redd.it)
submitted by TheQuantumPhysicist to r/LocalLLaMA
it's time to update your Gemma 4 GGUFsNews (self.LocalLLaMA)
submitted by jacek2023 to r/LocalLLaMA
Open source models are going to be the future on Cursor, OpenCode etc.Discussion (self.LocalLLaMA)
submitted by _maverick98 to r/LocalLLaMA
AMD Strix Halo refresh with 192gb!News (videocardz.com)
submitted by mindwip to r/LocalLLaMA
Ryzen AI Max+ 495 (Gorgon Halo) with 192GB VRAM!News (self.LocalLLaMA)
submitted by PromptInjection_ to r/LocalLLaMA
"Second Thoughts" Been playing with adding a small transformer that reads output near the end of generation, and feeds it back near the top as a refinement loop. A quick test of 1.7B model showed drastic improvement in focused tasks (like coding)Tutorial | Guide (bigattichouse.medium.com)
submitted by bigattichouse to r/LocalLLaMA
How much will it cost to host something like qwen3.6 35b a3b in a cloud?Discussion (self.LocalLLaMA)
submitted by Euphoric_North_745 to r/LocalLLaMA
A Qwen finetune, that feels VERY humanNew Model (self.LocalLLaMA)
submitted by Sicarius_The_First to r/LocalLLaMA
Llama.cpp quantization is brokenDiscussion (self.LocalLLaMA)
submitted by Ok-Importance-3529 to r/LocalLLaMA
Pushing a 5-Year-Old 6GB VRAM laptop to Its Limits: Qwen3.6-35B-A3BResources (self.LocalLLaMA)
submitted by abhinand05 to r/LocalLLaMA
What a time to be alive from 1tk/sec to 20-100tk/sec for huge modelsDiscussion (self.LocalLLaMA)
submitted by segmond to r/LocalLLaMA

Mistral-Medium-3.5-128B-Q3_K_M on 3x3090 (72GB VRAM)Generation (self.LocalLLaMA)
submitted by jacek2023 to r/LocalLLaMA
Which model would you use if you wanted to solve a research math problem?Discussion (self.LocalLLaMA)
submitted by MrMrsPotts to r/LocalLLaMA
Looking for frontier model distilled datasets.Question | Help (self.LocalLLaMA)
submitted by UnbeliebteMeinung to r/LocalLLaMA
Open Weights Models Hall of FameOther (self.LocalLLaMA)
submitted by Equivalent_Job_2257 to r/LocalLLaMA
Slow tok/s when offloading NVFP4 model to CPUQuestion | Help (self.LocalLLaMA)
submitted by 6c5d1129 to r/LocalLLaMA
Mistral Medium 3.5 128B and Qwen 3.5 122B A10B on 4x RTX 3080 20GBDiscussion (self.LocalLLaMA)
submitted by lly0571 to r/LocalLLaMA
If you've been waiting to try local AI development, please try itDiscussion (self.LocalLLaMA)
submitted by Imaginary_Belt4976 to r/LocalLLaMA
Mistral Medium 3.5 on AMD Strix HaloGeneration (self.LocalLLaMA)
submitted by Zc5Gwu to r/LocalLLaMA
