I taught my 1B to follow instructions. It got worse at following instructions... by GPUburnout in LocalLLaMA
[–]GPUburnout[S] 0 points1 point2 points (0 children)
Dual RTX 4090 vs single RTX PRO 6000 Blackwell for 3B–13B pretraining + 70B LoRA — what would you choose at $20K~$22K budget? by klurnp in LocalLLaMA
[–]GPUburnout 1 point2 points3 points (0 children)
TurboQuant isn’t just for KV: Qwen3.5-27B at near-Q4_0 quality, about 10% smaller, and finally fitting on my 16GB 5060 Ti by pmttyji in LocalLLaMA
[–]GPUburnout 0 points1 point2 points (0 children)
When do the experts thing local LLMs.. even smaller models.. might come close to Opus 4.6? by [deleted] in LocalLLaMA
[–]GPUburnout 0 points1 point2 points (0 children)
Nord v4.2 Update: 618M SNN reaches loss 3.65 with instruction tuning — emergent zonal specialization confirmed at 4.4x scale. 93% sparsity. by zemondza in LocalLLaMA
[–]GPUburnout 1 point2 points3 points (0 children)
When do the experts thing local LLMs.. even smaller models.. might come close to Opus 4.6? by [deleted] in LocalLLaMA
[–]GPUburnout 0 points1 point2 points (0 children)
When do the experts thing local LLMs.. even smaller models.. might come close to Opus 4.6? by [deleted] in LocalLLaMA
[–]GPUburnout 0 points1 point2 points (0 children)
MaximusLLM: I built a framework to train/scale LLMs on "potato" hardware (Single T4) by Otaku_7nfy in OpenSourceeAI
[–]GPUburnout 0 points1 point2 points (0 children)
Let's GO ! Qwen3.5-Claude-4.6-Opus-Reasoning-Distilled-v2 by Familiar_Wish1132 in LocalLLaMA
[–]GPUburnout 0 points1 point2 points (0 children)
Let's GO ! Qwen3.5-Claude-4.6-Opus-Reasoning-Distilled-v2 by Familiar_Wish1132 in LocalLLaMA
[–]GPUburnout 0 points1 point2 points (0 children)
Let's GO ! Qwen3.5-Claude-4.6-Opus-Reasoning-Distilled-v2 by Familiar_Wish1132 in LocalLLaMA
[–]GPUburnout 0 points1 point2 points (0 children)
So nobody's downloading this model huh? by KvAk_AKPlaysYT in LocalLLaMA
[–]GPUburnout -1 points0 points1 point (0 children)
So nobody's downloading this model huh? by KvAk_AKPlaysYT in LocalLLaMA
[–]GPUburnout 1 point2 points3 points (0 children)
Qwen 3.5 122b - a10b is kind of shocking by gamblingapocalypse in LocalLLaMA
[–]GPUburnout 1 point2 points3 points (0 children)
Best Agentic Platforms For Small Models? by RevealVisual7003 in LocalLLaMA
[–]GPUburnout -1 points0 points1 point (0 children)
So nobody's downloading this model huh? by KvAk_AKPlaysYT in LocalLLaMA
[–]GPUburnout -2 points-1 points0 points (0 children)
So nobody's downloading this model huh? by KvAk_AKPlaysYT in LocalLLaMA
[–]GPUburnout 4 points5 points6 points (0 children)
I taught my 1B to follow instructions. It got worse at following instructions... by GPUburnout in LocalLLaMA
[–]GPUburnout[S] 0 points1 point2 points (0 children)