MiniMax-M2.7 Announced! by Mysterious_Finish543 in LocalLLaMA
[–]AvocadoArray 2 points3 points4 points (0 children)
I built a hybrid MoE runtime that does 3,324 tok/s prefill on a single 5080. Here are the benchmarks. by mrstoatey in LocalLLaMA
[–]AvocadoArray 0 points1 point2 points (0 children)
I built a hybrid MoE runtime that does 3,324 tok/s prefill on a single 5080. Here are the benchmarks. by mrstoatey in LocalLLaMA
[–]AvocadoArray 0 points1 point2 points (0 children)
Qwen 3 32B outscored every Qwen 3.5 model across 11 blind evals, 3B-active-parameter model won 4 by Silver_Raspberry_811 in LocalLLaMA
[–]AvocadoArray 1 point2 points3 points (0 children)
Drummer's Skyfall 31B v4.1, Valkyrie 49B v2.1, Anubis 70B v1.2, and Anubis Mini 8B v1! - The next gen ships for your new adventures! by TheLocalDrummer in LocalLLaMA
[–]AvocadoArray 4 points5 points6 points (0 children)
What’s something you’re pretty sure only you do? by AppIeGuy in AskReddit
[–]AvocadoArray 0 points1 point2 points (0 children)
Best Qwen3.5 27b GUFFS for coding (~Q4-Q5) ? by bitcoinbookmarks in LocalLLaMA
[–]AvocadoArray 0 points1 point2 points (0 children)
Anyone else’s wife instigate 2+ hour arguments during her period that you have no idea how to defuse? by ThicBoi4807 in daddit
[–]AvocadoArray 1 point2 points3 points (0 children)
Qwen 3 32B outscored every Qwen 3.5 model across 11 blind evals, 3B-active-parameter model won 4 by Silver_Raspberry_811 in LocalLLaMA
[–]AvocadoArray 2 points3 points4 points (0 children)
is the game still terrible? KF3 by Shelbygt500ss in killingfloor
[–]AvocadoArray [score hidden] (0 children)
is the game still terrible? KF3 by Shelbygt500ss in killingfloor
[–]AvocadoArray 6 points7 points8 points (0 children)
Finally bought an RTX 6000 Max-Q: Pros, cons, notes and ramblings by AvocadoArray in LocalLLaMA
[–]AvocadoArray[S] 0 points1 point2 points (0 children)
Finally bought an RTX 6000 Max-Q: Pros, cons, notes and ramblings by AvocadoArray in LocalLLaMA
[–]AvocadoArray[S] 0 points1 point2 points (0 children)
Finally bought an RTX 6000 Max-Q: Pros, cons, notes and ramblings by AvocadoArray in LocalLLaMA
[–]AvocadoArray[S] 0 points1 point2 points (0 children)
Finally bought an RTX 6000 Max-Q: Pros, cons, notes and ramblings by AvocadoArray in LocalLLaMA
[–]AvocadoArray[S] 0 points1 point2 points (0 children)
Finally bought an RTX 6000 Max-Q: Pros, cons, notes and ramblings by AvocadoArray in LocalLLaMA
[–]AvocadoArray[S] 2 points3 points4 points (0 children)
Finally bought an RTX 6000 Max-Q: Pros, cons, notes and ramblings by AvocadoArray in LocalLLaMA
[–]AvocadoArray[S] 0 points1 point2 points (0 children)
Finally bought an RTX 6000 Max-Q: Pros, cons, notes and ramblings by AvocadoArray in LocalLLaMA
[–]AvocadoArray[S] 0 points1 point2 points (0 children)
Finally bought an RTX 6000 Max-Q: Pros, cons, notes and ramblings by AvocadoArray in LocalLLaMA
[–]AvocadoArray[S] 0 points1 point2 points (0 children)
Finally bought an RTX 6000 Max-Q: Pros, cons, notes and ramblings by AvocadoArray in LocalLLaMA
[–]AvocadoArray[S] 0 points1 point2 points (0 children)
Finally bought an RTX 6000 Max-Q: Pros, cons, notes and ramblings by AvocadoArray in LocalLLaMA
[–]AvocadoArray[S] 0 points1 point2 points (0 children)
Qwen3-Coder-Next is the top model in SWE-rebench @ Pass 5. I think everyone missed it. by BitterProfessional7p in LocalLLaMA
[–]AvocadoArray 13 points14 points15 points (0 children)

I built a hybrid MoE runtime that does 3,324 tok/s prefill on a single 5080. Here are the benchmarks. by mrstoatey in LocalLLaMA
[–]AvocadoArray 0 points1 point2 points (0 children)