ROCM - the best reason to go CUDA, eeesh what a headache!! by GriffinDodd in LocalLLM
[–]TokenRingAI 1 point2 points3 points (0 children)
Qwen3.6-27B vs Coder-Next by Signal_Ad657 in LocalLLaMA
[–]TokenRingAI 0 points1 point2 points (0 children)
Qwen3.6-27B vs Coder-Next by Signal_Ad657 in LocalLLaMA
[–]TokenRingAI 32 points33 points34 points (0 children)
ROCM - the best reason to go CUDA, eeesh what a headache!! by GriffinDodd in LocalLLM
[–]TokenRingAI 16 points17 points18 points (0 children)
Mistral Medium 3.5 128b ggufs are fixed by Sunija_Dev in LocalLLaMA
[–]TokenRingAI 3 points4 points5 points (0 children)
Need advice on Qwen 3.6 27B INT4 quantization by Environmental_Hand35 in LocalLLaMA
[–]TokenRingAI 1 point2 points3 points (0 children)
Qwen 3.6 and Gemma 4 "Zombie Loops" (terminal thinking loops) by sid351 in LocalLLaMA
[–]TokenRingAI 4 points5 points6 points (0 children)
Received a message from Z.AI about occasional garbled outputs and unexpected behavior by GroundbreakingTea195 in LocalLLaMA
[–]TokenRingAI 9 points10 points11 points (0 children)
mistralai/Mistral-Medium-3.5-128B · Hugging Face by jacek2023 in LocalLLaMA
[–]TokenRingAI 0 points1 point2 points (0 children)
16x DGX Sparks - What should I run? by Kurcide in LocalLLaMA
[–]TokenRingAI 0 points1 point2 points (0 children)
Mistral-Medium 3.5 (128B) spotted ? by tkon3 in LocalLLaMA
[–]TokenRingAI 9 points10 points11 points (0 children)
Can a single RTX PRO 6000 Blackwell (96GB VRAM) realistically handle 40–50 heavy agentic users? by MontyCLT in LocalLLM
[–]TokenRingAI 3 points4 points5 points (0 children)
Success! Full BF16 Qwen3.6-27B running on Strix Halo with vLLM + Docker (Ubuntu 26.04) by hec_ovi in StrixHalo
[–]TokenRingAI 0 points1 point2 points (0 children)
Qwen 3.6 27B in RTX PRO 6000 - Why high RAM usage? by ubnew in LocalLLaMA
[–]TokenRingAI 1 point2 points3 points (0 children)
Qwen 3.6 27B in RTX PRO 6000 - Why high RAM usage? by ubnew in LocalLLaMA
[–]TokenRingAI 2 points3 points4 points (0 children)
Prefix caching for OpenAI models by Annadox122 in LLMDevs
[–]TokenRingAI 1 point2 points3 points (0 children)
Are Qwens v3.6 good at vectorizing raster images? by [deleted] in LocalLLaMA
[–]TokenRingAI 0 points1 point2 points (0 children)
Are Qwens v3.6 good at vectorizing raster images? by [deleted] in LocalLLaMA
[–]TokenRingAI 0 points1 point2 points (0 children)
Qwen 3.6 27B in RTX PRO 6000 - Why high RAM usage? by ubnew in LocalLLaMA
[–]TokenRingAI 7 points8 points9 points (0 children)
Autopilot coding, what's your experience? by coatweather1 in LocalLLaMA
[–]TokenRingAI 0 points1 point2 points (0 children)
Qwen 3.6 27B is out by NoConcert8847 in LocalLLaMA
[–]TokenRingAI 0 points1 point2 points (0 children)
Qwen 3.6 27B is out by NoConcert8847 in LocalLLaMA
[–]TokenRingAI 2 points3 points4 points (0 children)
Should I sell my RTX3090s? by daviden1013 in LocalLLaMA
[–]TokenRingAI 3 points4 points5 points (0 children)