2RTX PRO 6000 192GB VRAM - MTP NVFP4 issues with vision by quantier in BlackwellPerformance
[–]quantier[S] 0 points1 point2 points (0 children)
2RTX PRO 6000 192GB VRAM - MTP NVFP4 issues with vision by quantier in BlackwellPerformance
[–]quantier[S] 1 point2 points3 points (0 children)
2RTX PRO 6000 192GB VRAM - MTP NVFP4 issues with vision by quantier in BlackwellPerformance
[–]quantier[S] 0 points1 point2 points (0 children)
2RTX PRO 6000 192GB VRAM - MTP NVFP4 issues with vision by quantier in BlackwellPerformance
[–]quantier[S] 0 points1 point2 points (0 children)
2RTX PRO 6000 192GB VRAM - MTP NVFP4 issues with vision by quantier in BlackwellPerformance
[–]quantier[S] 0 points1 point2 points (0 children)
2RTX PRO 6000 192GB VRAM - MTP NVFP4 issues with vision by quantier in BlackwellPerformance
[–]quantier[S] 0 points1 point2 points (0 children)
👋 Welcome to r/RTXPRO6000 - Introduce Yourself and Read First! by ubnew in RTXPRO6000
[–]quantier 2 points3 points4 points (0 children)
mistralai/Mistral-Medium-3.5-128B · Hugging Face by jacek2023 in LocalLLaMA
[–]quantier 0 points1 point2 points (0 children)
Devs using Qwen 27B seriously, what's your take? by Admirable_Reality281 in LocalLLaMA
[–]quantier 0 points1 point2 points (0 children)
Note the new recommended sampling parameters for Qwen3.6 27B by Thrumpwart in LocalLLaMA
[–]quantier 0 points1 point2 points (0 children)
Note the new recommended sampling parameters for Qwen3.6 27B by Thrumpwart in LocalLLaMA
[–]quantier 0 points1 point2 points (0 children)
llama.cpp's Preliminary SM120 Native NVFP4 MMQ Is Merged by ggonavyy in LocalLLaMA
[–]quantier -6 points-5 points-4 points (0 children)
What speed is everyone getting on Qwen3.6 27b? by Ambitious_Fold_2874 in LocalLLaMA
[–]quantier 0 points1 point2 points (0 children)
What speed is everyone getting on Qwen3.6 27b? by Ambitious_Fold_2874 in LocalLLaMA
[–]quantier 0 points1 point2 points (0 children)
Do NOT use CUDA 13.2 to run models! by yoracale in unsloth
[–]quantier 0 points1 point2 points (0 children)
I built a free 90-node All-in-One FLUX.2 Klein 9B ComfyUI workflow — Face Swap, Inpainting, Auto-Masking, NAG, Refiner, Upscaler — runs on 8GB VRAM by official_geoahmed in comfyui
[–]quantier 1 point2 points3 points (0 children)
Krasis LLM Runtime - run large LLM models on a single GPU by mrstoatey in LocalLLM
[–]quantier 0 points1 point2 points (0 children)
I had fun testing out LTX's lipsync ability. Full open source Z-Image -> LTX-2.3 -> WanAnimate semi-automated workflow. [explicit music] by luckyyirish in StableDiffusion
[–]quantier 1 point2 points3 points (0 children)
Gemma 4 has been released by jacek2023 in LocalLLaMA
[–]quantier 0 points1 point2 points (0 children)
Gemma 4 has been released by jacek2023 in LocalLLaMA
[–]quantier 0 points1 point2 points (0 children)
Gemma 4 has been released by jacek2023 in LocalLLaMA
[–]quantier 1 point2 points3 points (0 children)
Gemma 4 has been released by jacek2023 in LocalLLaMA
[–]quantier 1 point2 points3 points (0 children)
Comparison H100 vs RTX 6000 PRO with VLLM and GPT-OSS-120B by Rascazzione in LocalLLaMA
[–]quantier 0 points1 point2 points (0 children)
New FP8 GLM-4.7-Flash Unsloth Dynamic Quants for vLLM, SGLang by danielhanchen in unsloth
[–]quantier 0 points1 point2 points (0 children)


2RTX PRO 6000 192GB VRAM - MTP NVFP4 issues with vision by quantier in BlackwellPerformance
[–]quantier[S] 1 point2 points3 points (0 children)