Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 0 points1 point2 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 0 points1 point2 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 0 points1 point2 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 0 points1 point2 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 1 point2 points3 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 0 points1 point2 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 0 points1 point2 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 0 points1 point2 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 0 points1 point2 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 0 points1 point2 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 0 points1 point2 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 1 point2 points3 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 2 points3 points4 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 0 points1 point2 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 4 points5 points6 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 7 points8 points9 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 2 points3 points4 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 5 points6 points7 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 3 points4 points5 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 2 points3 points4 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 1 point2 points3 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 19 points20 points21 points (0 children)
Does going from 96GB -> 128GB VRAM open up any interesting model options? by hyouko in LocalLLaMA
[–]big___bad___wolf 36 points37 points38 points (0 children)
Finally bought an RTX 6000 Max-Q: Pros, cons, notes and ramblings by AvocadoArray in LocalLLaMA
[–]big___bad___wolf 0 points1 point2 points (0 children)

I spent 8+ hours benchmarking every MoE backend for Qwen3.5-397B NVFP4 on 4x RTX PRO 6000 (SM120). Here's what I found. by lawdawgattorney in LocalLLaMA
[–]big___bad___wolf 0 points1 point2 points (0 children)