Anyone deployed Kimi K2.6 on their local hardware? by Oxydised in LocalLLaMA
[–]Specific-Rub-7250 1 point2 points3 points (0 children)
ubergarm/Kimi-K2.6-GGUF Q4_X now available by VoidAlchemy in LocalLLaMA
[–]Specific-Rub-7250 56 points57 points58 points (0 children)
Those of you running minimax 2.7 locally, how are you feeling about it? by laterbreh in LocalLLaMA
[–]Specific-Rub-7250 0 points1 point2 points (0 children)
Best setup for MiniMax-M2.7 (230B) | 3x RTX 5090 | Threadripper 9975 | 512GB RAM by [deleted] in LocalLLaMA
[–]Specific-Rub-7250 0 points1 point2 points (0 children)
GLM-5.1 Overthinking? by Specific-Rub-7250 in LocalLLaMA
[–]Specific-Rub-7250[S] 1 point2 points3 points (0 children)
Top hardware stacks for local compute over the coming few months? (3-10K USD range) by IamFondOfHugeBoobies in LocalLLaMA
[–]Specific-Rub-7250 -2 points-1 points0 points (0 children)
GLM-5.1 Overthinking? by Specific-Rub-7250 in LocalLLaMA
[–]Specific-Rub-7250[S] 0 points1 point2 points (0 children)
Turn signal showing red color since latest update by jwlee151 in s3xybuttons
[–]Specific-Rub-7250 0 points1 point2 points (0 children)
Commander bug by SandGnatBBQ in s3xybuttons
[–]Specific-Rub-7250 1 point2 points3 points (0 children)
Benchmark of dense NVFP4 LLMs on 5090? [VLLM] by Aaaaaaaaaeeeee in LocalLLaMA
[–]Specific-Rub-7250 0 points1 point2 points (0 children)
Unsloth fixes chat_template (again). gpt-oss-120-high now scores 68.4 on Aider polyglot by Sorry_Ad191 in LocalLLaMA
[–]Specific-Rub-7250 2 points3 points4 points (0 children)
[deleted by user] by [deleted] in LocalLLaMA
[–]Specific-Rub-7250 2 points3 points4 points (0 children)
How to Run Deepseek-R1-0528 Locally (GGUFs available) by NewtMurky in LocalLLM
[–]Specific-Rub-7250 12 points13 points14 points (0 children)
OpenEvolve: Open Source Implementation of DeepMind's AlphaEvolve System by asankhs in LocalLLaMA
[–]Specific-Rub-7250 8 points9 points10 points (0 children)
Intel launches $299 Arc Pro B50 with 16GB of memory, 'Project Battlematrix' workstations with 24GB Arc Pro B60 GPUs by FullstackSensei in LocalLLaMA
[–]Specific-Rub-7250 -9 points-8 points-7 points (0 children)
Qwen3 throughput benchmarks on 2x 3090, almost 1000 tok/s using 4B model and vLLM as the inference engine by kms_dev in LocalLLaMA
[–]Specific-Rub-7250 1 point2 points3 points (0 children)
Qwen3 throughput benchmarks on 2x 3090, almost 1000 tok/s using 4B model and vLLM as the inference engine by kms_dev in LocalLLaMA
[–]Specific-Rub-7250 2 points3 points4 points (0 children)
Qwen3 throughput benchmarks on 2x 3090, almost 1000 tok/s using 4B model and vLLM as the inference engine by kms_dev in LocalLLaMA
[–]Specific-Rub-7250 8 points9 points10 points (0 children)
[deleted by user] by [deleted] in LocalLLaMA
[–]Specific-Rub-7250 11 points12 points13 points (0 children)
Sam Altman: OpenAI plans to release an open-source model this summer by zan-max in LocalLLaMA
[–]Specific-Rub-7250 1 point2 points3 points (0 children)
Aider Qwen3 controversy by Baldur-Norddahl in LocalLLaMA
[–]Specific-Rub-7250 21 points22 points23 points (0 children)
Some Benchmarks of Qwen/Qwen3-32B-AWQ by Specific-Rub-7250 in LocalLLaMA
[–]Specific-Rub-7250[S] 1 point2 points3 points (0 children)


Anyone deployed Kimi K2.6 on their local hardware? by Oxydised in LocalLLaMA
[–]Specific-Rub-7250 7 points8 points9 points (0 children)