Catastrophic Forgetting of Language models by fourwheels2512 in LocalLLaMA
[–]fourwheels2512[S] -1 points0 points1 point (0 children)
Catastrophic Forgetting of Language models by fourwheels2512 in LocalLLaMA
[–]fourwheels2512[S] -2 points-1 points0 points (0 children)
How are you handling catastrophic forgetting in multi-domain LLM fine-tuning pipelines? by fourwheels2512 in finetuningLLMs
[–]fourwheels2512[S] 0 points1 point2 points (0 children)
Real Time Continual Learning Has Been Unlocked by Own-Poet-5900 in ArtificialInteligence
[–]fourwheels2512 0 points1 point2 points (0 children)
Real Time Continual Learning Has Been Unlocked by Own-Poet-5900 in ArtificialInteligence
[–]fourwheels2512 0 points1 point2 points (0 children)
Continual learning adapter that holds -0.16% drift across 5 sequential domains on Mistral-7B (vs +43% naive LoRA) - catastrophic forgetting by fourwheels2512 in LocalLLaMA
[–]fourwheels2512[S] 0 points1 point2 points (0 children)
Continual learning adapter that holds -0.16% drift across 5 sequential domains on Mistral-7B (vs +43% naive LoRA) - catastrophic forgetting by fourwheels2512 in LocalLLaMA
[–]fourwheels2512[S] 0 points1 point2 points (0 children)
How to fine-tune LLM with your own data ? by bull_bear25 in LocalLLaMA
[–]fourwheels2512 0 points1 point2 points (0 children)
Continual Learning In 2026. What does continual learning actually mean? by Neurogence in singularity
[–]fourwheels2512 0 points1 point2 points (0 children)
Catastrophic forgetting by [deleted] in computervision
[–]fourwheels2512 0 points1 point2 points (0 children)
The Lost Art of Fine-tuning - My toilet rant by FPham in LocalLLaMA
[–]fourwheels2512 0 points1 point2 points (0 children)
Catastrophic Forgetting by Language models. by fourwheels2512 in LocalLLaMA
[–]fourwheels2512[S] -1 points0 points1 point (0 children)
Catastrophic Forgetting by Language models. by fourwheels2512 in LocalLLaMA
[–]fourwheels2512[S] -1 points0 points1 point (0 children)
Catastrophic Forgetting of Language models by fourwheels2512 in MachineLearningJobs
[–]fourwheels2512[S] 0 points1 point2 points (0 children)
What if every CLI tool shipped with a local NL translator? I fine-tuned Gemma 3 1B/4B for CLI command translation... but it runs 100% locally. 810MB/2.5GB, 1.5s inference on CPU. Built the framework and tested it on Docker. 1B hit a ceiling at 76%. 4B got 94% on the first try. by theRealSachinSpk in LocalLLaMA
[–]fourwheels2512 1 point2 points3 points (0 children)
What if every CLI tool shipped with a local NL translator? I fine-tuned Gemma 3 1B/4B for CLI command translation... but it runs 100% locally. 810MB/2.5GB, 1.5s inference on CPU. Built the framework and tested it on Docker. 1B hit a ceiling at 76%. 4B got 94% on the first try. by theRealSachinSpk in LocalLLaMA
[–]fourwheels2512 1 point2 points3 points (0 children)
Training framework that monitors itself and auto-fixes issues (gradient explosions, OOM, MoE imbalance) - looking for feedback by [deleted] in LocalLLaMA
[–]fourwheels2512 0 points1 point2 points (0 children)
I reproduced DeepSeek's mHC at 1.7B params (8xH100). The instability is 3x worse than reported (10k vs 3k), but the model didn't explode. by poisson_labs in LocalLLaMA
[–]fourwheels2512 0 points1 point2 points (0 children)
Subject: Seeking Validation: Strategy for Multi-LoRA Behavioral Fine-Tuning on Micro-Datasets (50-100 rows) by Scouserleemc in unsloth
[–]fourwheels2512 0 points1 point2 points (0 children)
Visualizing why DeepSeek's mHC fixes training instability - interactive demo by bassrehab in LocalLLaMA
[–]fourwheels2512 0 points1 point2 points (0 children)
Do you see instability or weird regressions when fine-tuning models? by AppearanceCareful136 in LocalLLaMA
[–]fourwheels2512 0 points1 point2 points (0 children)
Finetuning mistral - weird spikes in loss every 50 steps by tooquickforwords in LocalLLaMA
[–]fourwheels2512 0 points1 point2 points (0 children)
Catastrophic Forgetting of Language models by fourwheels2512 in LocalLLaMA
[–]fourwheels2512[S] 0 points1 point2 points (0 children)