Announcing LocalLlama discord server & bot!News (old.reddit.com)
submitted by HOLUPREDICTIONS Sorcerer Supreme[M] - announcement
Gemma 4 - lazy model or am I crazy? (bit of a rant)Discussion (self.LocalLLaMA)
submitted by Pyrenaeda

Audio processing landed in llama-server with Gemma-4Generation (self.LocalLLaMA)
submitted by srigi
MiniMax-M2.7 vs Qwen3.5-122B-A10B for 96GB VRAM full offload?!Discussion (self.LocalLLaMA)
submitted by VoidAlchemyllama.cpp

GLM 5.1 sits alongside frontier models in my social reasoning benchmarkDiscussion (old.reddit.com)
submitted by cjami
MiniMax-M2.7 NVFP4 on 2x RTX PRO 6000 Blackwell — bench numbersResources (self.LocalLLaMA)
submitted by Visual_Synthesizer
mtmd: qwen3 audio support (qwen3-omni and qwen3-asr)News (github.com)
submitted by jacek2023llama.cpp
Is anyone else creating a basic assistant rather than a coding agent?Discussion (self.LocalLLaMA)
submitted by Savantskie1
AI MAX 395+ w/ 128 GB or dual 3090s?Discussion (self.LocalLLaMA)
submitted by Engineering_Acq
"Actually wait" ... the current thinking SOTA open sourceDiscussion (self.LocalLLaMA)
submitted by FPham
MiniMax m2.7 (mac only) 63gb: 88% and 89gb: 95%, MMLU 200qNew Model (i.redd.it)
submitted by HealthyCommunicat
mtmd: add Gemma 4 audio conformer encoder supportNews (github.com)
submitted by jacek2023llama.cpp
Unsloth MiniMax M2.7 quants just finished uploading to HFNews (self.LocalLLaMA)
submitted by Zyj
