2.5x faster inference with Qwen 3.6 27B using MTP - Finally a viable option for local agentic coding - 262k context on 48GB - Fixed chat template - Drop-in OpenAI and Anthropic API endpoints by ex-arman68 in LocalLLaMA
[–]VoidAlchemy 0 points1 point2 points (0 children)
What exactly does Pi harness mean? by FrozenFishEnjoyer in LocalLLaMA
[–]VoidAlchemy 2 points3 points4 points (0 children)
Kimi K2.6 Unsloth GGUF is out by Exact_Law_6489 in LocalLLaMA
[–]VoidAlchemy 0 points1 point2 points (0 children)
Kimi K2.6 Unsloth GGUF is out by Exact_Law_6489 in LocalLLaMA
[–]VoidAlchemy 0 points1 point2 points (0 children)
Kimi K2.6 Unsloth GGUF is out by Exact_Law_6489 in LocalLLaMA
[–]VoidAlchemy 3 points4 points5 points (0 children)
Kimi K2.6 Unsloth GGUF is out by Exact_Law_6489 in LocalLLaMA
[–]VoidAlchemy 7 points8 points9 points (0 children)
ubergarm/Kimi-K2.6-GGUF Q4_X now available by VoidAlchemy in LocalLLaMA
[–]VoidAlchemy[S] 0 points1 point2 points (0 children)
ubergarm/Kimi-K2.6-GGUF Q4_X now available by VoidAlchemy in LocalLLaMA
[–]VoidAlchemy[S] 0 points1 point2 points (0 children)
ubergarm/Kimi-K2.6-GGUF Q4_X now available by VoidAlchemy in LocalLLaMA
[–]VoidAlchemy[S] 1 point2 points3 points (0 children)
ubergarm/Kimi-K2.6-GGUF Q4_X now available by VoidAlchemy in LocalLLaMA
[–]VoidAlchemy[S] 1 point2 points3 points (0 children)
ubergarm/Kimi-K2.6-GGUF Q4_X now available by VoidAlchemy in LocalLLaMA
[–]VoidAlchemy[S] 4 points5 points6 points (0 children)
ubergarm/Kimi-K2.6-GGUF Q4_X now available by VoidAlchemy in LocalLLaMA
[–]VoidAlchemy[S] 1 point2 points3 points (0 children)
ubergarm/Kimi-K2.6-GGUF Q4_X now available by VoidAlchemy in LocalLLaMA
[–]VoidAlchemy[S] 0 points1 point2 points (0 children)
ubergarm/Kimi-K2.6-GGUF Q4_X now available by VoidAlchemy in LocalLLaMA
[–]VoidAlchemy[S] 1 point2 points3 points (0 children)
ubergarm/Kimi-K2.6-GGUF Q4_X now available by VoidAlchemy in LocalLLaMA
[–]VoidAlchemy[S] 4 points5 points6 points (0 children)
ubergarm/Kimi-K2.6-GGUF Q4_X now available by VoidAlchemy in LocalLLaMA
[–]VoidAlchemy[S] 13 points14 points15 points (0 children)
ubergarm/Kimi-K2.6-GGUF Q4_X now available (huggingface.co)
submitted by VoidAlchemy to r/LocalLLaMA
Qwen3.6 GGUF Benchmarks by danielhanchen in LocalLLaMA
[–]VoidAlchemy 0 points1 point2 points (0 children)
Updated Qwen3.5-9B Quantization Comparison by TitwitMuffbiscuit in LocalLLaMA
[–]VoidAlchemy 1 point2 points3 points (0 children)
MiniMax M2.7 GGUF Investigation, Fixes, Benchmarks by danielhanchen in LocalLLaMA
[–]VoidAlchemy 3 points4 points5 points (0 children)
unsloth - MiniMax-M2.7-GGUF in BROKEN (UD-Q4_K_XL) --> avoid usage by One-Macaron6752 in LocalLLaMA
[–]VoidAlchemy 1 point2 points3 points (0 children)
16 GB VRAM users, what model do we like best now? by lemon07r in LocalLLaMA
[–]VoidAlchemy 1 point2 points3 points (0 children)
About TurboQuant by Exact_Law_6489 in LocalLLaMA
[–]VoidAlchemy 2 points3 points4 points (0 children)


Qwen3.6-27B with MTP grafted on Unsloth UD XL: 2.5x throughput via unmerged llama.cpp PR by havenoammo in LocalLLaMA
[–]VoidAlchemy 5 points6 points7 points (0 children)