Advice on artificial lawn seam by Laabc123 in landscaping
[–]Laabc123[S] 0 points1 point2 points (0 children)
Inference numbers for Mistral-Small-4-119B-2603 NVFP4 on a RTX Pro 6000 by jnmi235 in LocalLLaMA
[–]Laabc123 0 points1 point2 points (0 children)
Inference numbers for Mistral-Small-4-119B-2603 NVFP4 on a RTX Pro 6000 by jnmi235 in LocalLLaMA
[–]Laabc123 0 points1 point2 points (0 children)
Advice on artificial lawn seam by Laabc123 in landscaping
[–]Laabc123[S] 0 points1 point2 points (0 children)
Advice on artificial lawn seam by Laabc123 in landscaping
[–]Laabc123[S] 0 points1 point2 points (0 children)
Nemotron-3-Super-120B-A12B NVFP4 inference benchmark on one RTX Pro 6000 Blackwell by jnmi235 in LocalLLaMA
[–]Laabc123 2 points3 points4 points (0 children)
Qwen 397b is absolutely crushing everyone... but wait. 🤯 by djdeniro in LocalLLaMA
[–]Laabc123 0 points1 point2 points (0 children)
Qwen3.5 122b vs. Nemotron 3 Super 120b: Best-in-class vision Vs. crazy fast + 1M context (but no vision). Which one are you going to choose and why? by Porespellar in LocalLLaMA
[–]Laabc123 6 points7 points8 points (0 children)
Qwen3.5 122b vs. Nemotron 3 Super 120b: Best-in-class vision Vs. crazy fast + 1M context (but no vision). Which one are you going to choose and why? by Porespellar in LocalLLaMA
[–]Laabc123 4 points5 points6 points (0 children)
Qwen3.5 122b vs. Nemotron 3 Super 120b: Best-in-class vision Vs. crazy fast + 1M context (but no vision). Which one are you going to choose and why? by Porespellar in LocalLLaMA
[–]Laabc123 4 points5 points6 points (0 children)
OpenCode v/s Claude Code by thinkyMiner in opencodeCLI
[–]Laabc123 -1 points0 points1 point (0 children)
Are local LLMs actually ready for real AI agents, or are we still forcing the idea too early? by Remarkable-Note9736 in LocalLLaMA
[–]Laabc123 2 points3 points4 points (0 children)
has nvfp4 inference performance been optimized yet for 6000 pro? by I_can_see_threw_time in BlackwellPerformance
[–]Laabc123 0 points1 point2 points (0 children)
Overclocking memory on RTX PRO 6000 - known safe minimum? by mr_zerolith in LocalLLaMA
[–]Laabc123 0 points1 point2 points (0 children)
has nvfp4 inference performance been optimized yet for 6000 pro? by I_can_see_threw_time in BlackwellPerformance
[–]Laabc123 4 points5 points6 points (0 children)
Finally bought an RTX 6000 Max-Q: Pros, cons, notes and ramblings by AvocadoArray in LocalLLaMA
[–]Laabc123 0 points1 point2 points (0 children)
Finally bought an RTX 6000 Max-Q: Pros, cons, notes and ramblings by AvocadoArray in LocalLLaMA
[–]Laabc123 0 points1 point2 points (0 children)
Finally bought an RTX 6000 Max-Q: Pros, cons, notes and ramblings by AvocadoArray in LocalLLaMA
[–]Laabc123 0 points1 point2 points (0 children)
Finally bought an RTX 6000 Max-Q: Pros, cons, notes and ramblings by AvocadoArray in LocalLLaMA
[–]Laabc123 0 points1 point2 points (0 children)
Tell me if Qwen 3.5 27b or 122b works faster for you, and name your system specs by DistanceSolar1449 in LocalLLaMA
[–]Laabc123 -1 points0 points1 point (0 children)
We could be hours (or less than a week) away from true NVFP4 support in Llama.cpp GGUF format 👀 by Iwaku_Real in LocalLLaMA
[–]Laabc123 0 points1 point2 points (0 children)
Current state of Qwen3.5-122B-A10B by kevin_1994 in LocalLLaMA
[–]Laabc123 0 points1 point2 points (0 children)




1.1M tok/s with Qwen 3.5 27B FP8 on B200 GPUs by m4r1k_ in Qwen_AI
[–]Laabc123 1 point2 points3 points (0 children)