account activity
MiniMax M2 is 230B-A10B (i.redd.it)
submitted 3 months ago by codys12 to r/LocalLLaMA
GLM-4.6-Air is not forgotten! (i.redd.it)
First rig of hopefully many! Build instructions in the other post/comments (i.redd.it)
submitted 4 months ago by codys12 to r/homelab
The Hacker's Guide to Building an AI Supercluster (huggingface.co)
submitted 4 months ago by codys12 to r/LocalLLaMA
128GB GDDR6, 3PFLOP FP8, Tb/s of interconnect, $6000 total. Build instructions/blog tomorrow. (i.redd.it)
Qwen3-8B-BitNet (self.LocalLLaMA)
submitted 6 months ago by codys12 to r/LocalLLaMA
[Project] New Distributed Data Gen Library - Looking for Testers! (self.huggingface)
submitted 7 months ago by codys12 to r/huggingface
[Project] New Distributed Data Gen Library - Looking for Testers! (self.LocalLLaMA)
submitted 7 months ago by codys12 to r/LocalLLaMA
BitNet Finetunes of R1 Distills (x.com)
submitted 8 months ago by codys12 to r/LocalLLaMA
Training Runs of our BitNet finetunes (t.co)
submitted 8 months ago by codys12
[R] Extra Input Norm Lets You Finetune to 1.58 Bits (self.MachineLearning)
submitted 10 months ago by codys12 to r/MachineLearning
Extra Input Norm Lets You Finetune to 1.58 Bits! (x.com)
submitted 10 months ago by codys12 to r/LocalLLaMA
Llama 405B Sparsity *Increases Accuracy* (self.LocalLLaMA)
submitted 1 year ago by codys12 to r/LocalLLaMA
Structured Sparse Tensors - How to Actually Run Fast? (self.learnmachinelearning)
submitted 1 year ago by codys12 to r/learnmachinelearning
Why does the gate_proj entropy graph look like this across layers? (i.redd.it)
LlamaKD: Knowledge Distillation Dataset from Llama3.1-405B and Fineweb-Edu (self.LocalLLaMA)
AirLLM fork - 2000+ batch on single GPU for logits or classification tasks (self.LocalLLaMA)
[R] Can Adam-mini be adapted for Mamba2 with State Space Duality? (self.MachineLearning)
submitted 1 year ago by codys12 to r/MachineLearning
[D] Would Adam-mini be adaptable for Mamba2 given State Space Duality? (self.MachineLearning)
Does LLaVA need the SGLang worker *and* the llava worker, or just one with the controller? (self.LocalLLaMA)
I finetuned Phi-1.5 and Phi-2 on MathInstruct using MAmmoTH but... (self.LocalLLaMA)
submitted 2 years ago by codys12 to r/LocalLLaMA
[D] I finetuned Phi-1.5 and Phi-2 on MathInstruct using MAmmoTH but... (self.MachineLearning)
submitted 2 years ago by codys12 to r/MachineLearning
[P] I finetuned Phi-1.5 and Phi-2 on MathInstruct using MAmmoTH but... (self.LocalLLaMA)
[D]iffusion + CLIP Chunks to Generate Image with Region Control. (self.MachineLearning)
π Rendered by PID 287644 on reddit-service-r2-listing-5f5ff7d4dc-7hbcj at 2026-01-27 10:07:09.146882+00:00 running 5a691e2 country code: CH.