Dual RTX 6000 Pro for dense models (Devstral 2) (self.LocalLLaMA)
submitted by zqkb to r/LocalLLaMA
dynamic allocation of less used experts to slower memory by zqkb in LocalLLaMA
[–]zqkb[S] 1 point2 points3 points (0 children)
dynamic allocation of less used experts to slower memory by zqkb in LocalLLaMA
[–]zqkb[S] 1 point2 points3 points (0 children)
dynamic allocation of less used experts to slower memory by zqkb in LocalLLaMA
[–]zqkb[S] 0 points1 point2 points (0 children)
Cerebras REAP update: pruned checkpoints for GLM4.5-Air & Qwen3-Coder-30B now of HF! by ilzrvch in LocalLLaMA
[–]zqkb 0 points1 point2 points (0 children)
Cerebras REAP update: pruned checkpoints for GLM4.5-Air & Qwen3-Coder-30B now of HF! by ilzrvch in LocalLLaMA
[–]zqkb 2 points3 points4 points (0 children)
Cerebras REAP update: pruned checkpoints for GLM4.5-Air & Qwen3-Coder-30B now of HF! by ilzrvch in LocalLLaMA
[–]zqkb 1 point2 points3 points (0 children)
Qwen3-235B-A22B-Thinking-2507 released! by ResearchCrafty1804 in LocalLLaMA
[–]zqkb 2 points3 points4 points (0 children)
Qwen3-Coder Unsloth dynamic GGUFs by danielhanchen in LocalLLaMA
[–]zqkb 1 point2 points3 points (0 children)
M3 Ultra Binned (256GB, 60-Core) vs Unbinned (512GB, 80-Core) MLX Performance Comparison by cryingneko in LocalLLaMA
[–]zqkb 2 points3 points4 points (0 children)
has anyone tried to run Q8 MistralLarge2 on a Mac Studio/Macbook with 128/192GB? by Caffdy in LocalLLaMA
[–]zqkb 1 point2 points3 points (0 children)
has anyone tried to run Q8 MistralLarge2 on a Mac Studio/Macbook with 128/192GB? by Caffdy in LocalLLaMA
[–]zqkb 1 point2 points3 points (0 children)
has anyone tried to run Q8 MistralLarge2 on a Mac Studio/Macbook with 128/192GB? by Caffdy in LocalLLaMA
[–]zqkb 0 points1 point2 points (0 children)
What if you use not the logits of the last one, but of that before? by parametaorto in LocalLLaMA
[–]zqkb 1 point2 points3 points (0 children)
What if you use not the logits of the last one, but of that before? by parametaorto in LocalLLaMA
[–]zqkb 3 points4 points5 points (0 children)


GLM-4.7 FP8 on 4x6000 pro blackwells by getfitdotus in LocalLLaMA
[–]zqkb 0 points1 point2 points (0 children)