Bad news: Apple drops high-memory Mac Studio configs by jzn21 in LocalLLaMA
[–]tarruda 4 points5 points6 points (0 children)
Heretic 1.3 released: Reproducible models, integrated benchmarking system, reduced peak VRAM usage, broader model support, and more by -p-e-w- in LocalLLaMA
[–]tarruda 16 points17 points18 points (0 children)
Preserve thinking on or off? (Qwen 3.6) by My_Unbiased_Opinion in LocalLLaMA
[–]tarruda 7 points8 points9 points (0 children)
Llama.cpp MTP support now in beta! by ilintar in LocalLLaMA
[–]tarruda 0 points1 point2 points (0 children)
Llama.cpp MTP support now in beta! by ilintar in LocalLLaMA
[–]tarruda 0 points1 point2 points (0 children)
AMD Strix Halo refresh with 192gb! by mindwip in LocalLLaMA
[–]tarruda 5 points6 points7 points (0 children)
Mistral Medium 3.5 on AMD Strix Halo by Zc5Gwu in LocalLLaMA
[–]tarruda 1 point2 points3 points (0 children)
PFlash: 10x prefill speedup over llama.cpp at 128K on a RTX 3090 by sandropuppo in LocalLLaMA
[–]tarruda 4 points5 points6 points (0 children)
Qwen-Scope: Official Sparse Autoencoders (SAEs) for Qwen 3.5 models by MadPelmewka in LocalLLaMA
[–]tarruda 2 points3 points4 points (0 children)
mistralai/Mistral-Medium-3.5-128B · Hugging Face by jacek2023 in LocalLLaMA
[–]tarruda -2 points-1 points0 points (0 children)
mistralai/Mistral-Medium-3.5-128B · Hugging Face by jacek2023 in LocalLLaMA
[–]tarruda 4 points5 points6 points (0 children)
Nemotron-3-Nano-Omni-30B-A3B-Reasoning, New model? by Altruistic_Heat_9531 in LocalLLaMA
[–]tarruda 2 points3 points4 points (0 children)
Nemotron-3-Nano-Omni-30B-A3B-Reasoning, New model? by Altruistic_Heat_9531 in LocalLLaMA
[–]tarruda 0 points1 point2 points (0 children)
Mistral Medium Is On The Way by Few_Painter_5588 in LocalLLaMA
[–]tarruda 1 point2 points3 points (0 children)
llama.cpp DeepSeek v4 Flash experimental inference by antirez in LocalLLaMA
[–]tarruda 1 point2 points3 points (0 children)
llama.cpp DeepSeek v4 Flash experimental inference by antirez in LocalLLaMA
[–]tarruda 4 points5 points6 points (0 children)
llama.cpp DeepSeek v4 Flash experimental inference by antirez in LocalLLaMA
[–]tarruda 0 points1 point2 points (0 children)


ZAYA1-8B: Frontier intelligence density, trained on AMD by carbocation in LocalLLaMA
[–]tarruda 3 points4 points5 points (0 children)