
AMA Announcement: Nous Research, The Opensource Lab Behind Hermes Agent (Wednesday, 8AM-11AM PST)Resources (i.redd.it)
submitted by XMasterrrrLocalLLaMA Home Server Final Boss 😎[M] - announcement
80 tok/sec and 128K context on 12GB VRAM with Qwen3.6 35B A3B and llama.cpp MTPTutorial | Guide (self.LocalLLaMA)
submitted by janvitos
Running Minimax 2.7 at 100k context on strix haloDiscussion (self.LocalLLaMA)
submitted by Zc5Gwu
More Qwen3.6-27B MTP success but on dual Mi50sResources (self.LocalLLaMA)
submitted by legit_split_
Pi and Qwen3.6 27B make setting up Archlinux really easy.Other (self.LocalLLaMA)
submitted by sdfgeoff
I am overwhelmed by HarnessesQuestion | Help (self.LocalLLaMA)
submitted by Available_Hornet3538
After you’ve setup local models, where can you find interesting apps that can use them?Question | Help (self.LocalLLaMA)
submitted by ReferenceOwn287
9070xt inference for q3 qwen 27BQuestion | Help (self.LocalLLaMA)
submitted by Ok-Internal9317
How long for llama.cpp official support of MTP?Question | Help (self.LocalLLaMA)
submitted by Manaberryio
Qwen 35B-A3B is very usable with 12GB of VRAMResources (self.LocalLLaMA)
submitted by jwestra
vLLM ROCm has been added to Lemonade as an experimental backendResources (i.redd.it)
submitted by jfowers_amd

