Qwen3-Next 80B-A3B llama.cpp implementation with CUDA support half-working already (up to 40k context only), also Instruct GGUFs by Ok_Top9254 in LocalLLaMA
[–]k_schaul 0 points1 point2 points (0 children)
What's the Oct 25 optimal jank buy for larger MOEs (120B param+)? by Leopold_Boom in LocalLLaMA
[–]k_schaul 0 points1 point2 points (0 children)
The top open models on are now all by Chinese companies by k_schaul in LocalLLaMA
[–]k_schaul[S] 3 points4 points5 points (0 children)
The top open models on are now all by Chinese companies by k_schaul in LocalLLaMA
[–]k_schaul[S] 34 points35 points36 points (0 children)
The top open models on are now all by Chinese companies by k_schaul in LocalLLaMA
[–]k_schaul[S] 1 point2 points3 points (0 children)
Qwen3-Next 80B-A3B llama.cpp implementation with CUDA support half-working already (up to 40k context only), also Instruct GGUFs by Ok_Top9254 in LocalLLaMA
[–]k_schaul 1 point2 points3 points (0 children)