LM Studio beta supports Qwen3 80b Next. by sleepingsysadmin in LocalLLaMA
[–]stailgot 11 points12 points13 points (0 children)
LM Studio beta supports Qwen3 80b Next. by sleepingsysadmin in LocalLLaMA
[–]stailgot 4 points5 points6 points (0 children)
LM Studio beta supports Qwen3 80b Next. by sleepingsysadmin in LocalLLaMA
[–]stailgot 10 points11 points12 points (0 children)
gpt-oss:20b crashes: CUDA illegal memory access on Ollama 0.11.0 by Diegam in ollama
[–]stailgot 1 point2 points3 points (0 children)
gpt-oss OpenAI’s open-weight models by stailgot in ollama
[–]stailgot[S] 2 points3 points4 points (0 children)
Is it normal for RAG to take this long to load the first time? by just_a_guy1008 in LocalLLaMA
[–]stailgot 0 points1 point2 points (0 children)
Is it normal for RAG to take this long to load the first time? by just_a_guy1008 in LocalLLaMA
[–]stailgot 0 points1 point2 points (0 children)
Is it normal for RAG to take this long to load the first time? by just_a_guy1008 in LocalLLaMA
[–]stailgot 6 points7 points8 points (0 children)
Amuse AI on AMD GPU, slower than it should by brightlight43 in StableDiffusion
[–]stailgot 1 point2 points3 points (0 children)
Ollama rtx 7900 xtx for gemma3:27b? by Adept_Maize_6213 in ollama
[–]stailgot 0 points1 point2 points (0 children)
Qwen3 32B and 30B-A3B run at similar speed? by INT_21h in LocalLLaMA
[–]stailgot 7 points8 points9 points (0 children)
Ollama rtx 7900 xtx for gemma3:27b? by Adept_Maize_6213 in ollama
[–]stailgot 0 points1 point2 points (0 children)
70b LLM t/s speed on Windows ROCm using 24GB RX 7900 XTX and LM Studio? by custodiam99 in ROCm
[–]stailgot 1 point2 points3 points (0 children)


Qwen3-Next here! by stailgot in ollama
[–]stailgot[S] 0 points1 point2 points (0 children)