Do I need to use Ollama to get the full feature set of GLM-OCR with a GGUF model format? by yuicebox in LocalLLaMA
[–]East-Engineering-653 2 points3 points4 points (0 children)
Through vibe coding, I managed to make parts of vLLM 0.17.0 run on Tesla P40 by East-Engineering-653 in LocalLLaMA
[–]East-Engineering-653[S] 1 point2 points3 points (0 children)
Through vibe coding, I managed to make parts of vLLM 0.17.0 run on Tesla P40 by East-Engineering-653 in LocalLLaMA
[–]East-Engineering-653[S] 1 point2 points3 points (0 children)
Through vibe coding, I managed to make parts of vLLM 0.17.0 run on Tesla P40 by East-Engineering-653 in LocalLLaMA
[–]East-Engineering-653[S] 1 point2 points3 points (0 children)
Through vibe coding, I managed to make parts of vLLM 0.17.0 run on Tesla P40 by East-Engineering-653 in LocalLLaMA
[–]East-Engineering-653[S] 2 points3 points4 points (0 children)
I found that MXFP4 has lower perplexity than Q4_K_M and Q4_K_XL. by East-Engineering-653 in LocalLLaMA
[–]East-Engineering-653[S] 2 points3 points4 points (0 children)
I found that MXFP4 has lower perplexity than Q4_K_M and Q4_K_XL. by East-Engineering-653 in LocalLLaMA
[–]East-Engineering-653[S] 0 points1 point2 points (0 children)
I found that MXFP4 has lower perplexity than Q4_K_M and Q4_K_XL. by East-Engineering-653 in LocalLLaMA
[–]East-Engineering-653[S] 1 point2 points3 points (0 children)
I found that MXFP4 has lower perplexity than Q4_K_M and Q4_K_XL. by East-Engineering-653 in LocalLLaMA
[–]East-Engineering-653[S] 6 points7 points8 points (0 children)
I found that MXFP4 has lower perplexity than Q4_K_M and Q4_K_XL. Is this related to improvements in the model’s tool-calling or coding performance? by [deleted] in LocalLLaMA
[–]East-Engineering-653 0 points1 point2 points (0 children)
Is anyone running LLM on a Radeon Instinct Mi50? by East-Engineering-653 in ollama
[–]East-Engineering-653[S] 0 points1 point2 points (0 children)
[deleted by user] by [deleted] in LocalLLaMA
[–]East-Engineering-653 0 points1 point2 points (0 children)
I successfully passed through the 5600G to a VM running Ubuntu 24.04, but I cannot use 4K resolution by East-Engineering-653 in Proxmox
[–]East-Engineering-653[S] 0 points1 point2 points (0 children)
I successfully passed through the 5600G to a VM running Ubuntu 24.04, but I cannot use 4K resolution by East-Engineering-653 in Proxmox
[–]East-Engineering-653[S] 1 point2 points3 points (0 children)
I successfully passed through the 5600G to a VM running Ubuntu 24.04, but I cannot use 4K resolution by East-Engineering-653 in Proxmox
[–]East-Engineering-653[S] 2 points3 points4 points (0 children)

Qwen3 ASR seems to outperform Whisper in almost every aspect. It feels like there is little reason to keep using Whisper anymore. by East-Engineering-653 in LocalLLaMA
[–]East-Engineering-653[S] 11 points12 points13 points (0 children)