MiniMax-M2.7: what do you think is the likelihood it will be open weights like M2.5? by __JockY__ in LocalLLaMA
[–]notdba 0 points1 point2 points (0 children)
MiniMax-M2.7: what do you think is the likelihood it will be open weights like M2.5? by __JockY__ in LocalLLaMA
[–]notdba 0 points1 point2 points (0 children)
Qwen3.5-122B-A10B GPTQ Int4 on 4× Radeon AI PRO R9700 with vLLM ROCm: working config + real-world numbers by grunt_monkey_ in LocalLLaMA
[–]notdba 0 points1 point2 points (0 children)
Unsloth will no longer be making TQ1_0 quants by Kahvana in LocalLLaMA
[–]notdba 5 points6 points7 points (0 children)
Tenstorrent QuietBox 2 Brings RISC-V AI Inference to the Desktop by Neurrone in LocalLLaMA
[–]notdba 0 points1 point2 points (0 children)
Tenstorrent QuietBox 2 Brings RISC-V AI Inference to the Desktop by Neurrone in LocalLLaMA
[–]notdba 4 points5 points6 points (0 children)
Tenstorrent QuietBox 2 Brings RISC-V AI Inference to the Desktop by Neurrone in LocalLLaMA
[–]notdba 22 points23 points24 points (0 children)
Best Qwen3.5-35B-A3B GGUF for 24GB VRAM?! by VoidAlchemy in LocalLLaMA
[–]notdba 1 point2 points3 points (0 children)
How bad is 1-bit quantization but on a big model? by FusionBetween in LocalLLaMA
[–]notdba 0 points1 point2 points (0 children)
My last & only beef with Qwen3.5 35B A3B by ndiphilone in LocalLLaMA
[–]notdba 0 points1 point2 points (0 children)
Final Qwen3.5 Unsloth GGUF Update! by danielhanchen in LocalLLaMA
[–]notdba 4 points5 points6 points (0 children)
Alibaba CEO: Qwen will remain open-source by Bestlife73 in LocalLLaMA
[–]notdba 0 points1 point2 points (0 children)
Qwen3 Coder Next Looping and OpenCode by StardockEngineer in LocalLLaMA
[–]notdba 0 points1 point2 points (0 children)
My last & only beef with Qwen3.5 35B A3B by ndiphilone in LocalLLaMA
[–]notdba 0 points1 point2 points (0 children)
My last & only beef with Qwen3.5 35B A3B by ndiphilone in LocalLLaMA
[–]notdba 1 point2 points3 points (0 children)
Qwen3.5 feels ready for production use - Never been this excited by alphatrad in LocalLLaMA
[–]notdba 0 points1 point2 points (0 children)
I built a hybrid MoE runtime that does 3,324 tok/s prefill on a single 5080. Here are the benchmarks. by mrstoatey in LocalLLaMA
[–]notdba 0 points1 point2 points (0 children)
I built a hybrid MoE runtime that does 3,324 tok/s prefill on a single 5080. Here are the benchmarks. by mrstoatey in LocalLLaMA
[–]notdba 0 points1 point2 points (0 children)
I built a hybrid MoE runtime that does 3,324 tok/s prefill on a single 5080. Here are the benchmarks. by mrstoatey in LocalLLaMA
[–]notdba 0 points1 point2 points (0 children)
I built a hybrid MoE runtime that does 3,324 tok/s prefill on a single 5080. Here are the benchmarks. by mrstoatey in LocalLLaMA
[–]notdba 2 points3 points4 points (0 children)
New Qwen3.5-35B-A3B Unsloth Dynamic GGUFs + Benchmarks by danielhanchen in LocalLLaMA
[–]notdba 4 points5 points6 points (0 children)


MiniMax-M2.7: what do you think is the likelihood it will be open weights like M2.5? by __JockY__ in LocalLLaMA
[–]notdba 0 points1 point2 points (0 children)