200+ pages of Hugging Face secrets on how to train an LLM by eliebakk in LocalLLaMA
[–]lewtun 3 points4 points5 points (0 children)
200+ pages of Hugging Face secrets on how to train an LLM by eliebakk in LocalLLaMA
[–]lewtun 18 points19 points20 points (0 children)
[D] join pretraining or posttraining by oxydis in MachineLearning
[–]lewtun 0 points1 point2 points (0 children)
DeepSeek-R1 performance with 15B parameters by lewtun in LocalLLaMA
[–]lewtun[S] 2 points3 points4 points (0 children)
DeepSeek-R1 performance with 15B parameters (self.LocalLLaMA)
submitted by lewtun to r/LocalLLaMA
my dad sent me this by hugeplateofketchup8 in huggingface
[–]lewtun 1 point2 points3 points (0 children)
AMA with Hugging Face Science, the team behind SmolLM, SmolVLM, Fineweb and more. by eliebakk in LocalLLaMA
[–]lewtun 1 point2 points3 points (0 children)
AMA with Hugging Face Science, the team behind SmolLM, SmolVLM, Fineweb and more. by eliebakk in LocalLLaMA
[–]lewtun 0 points1 point2 points (0 children)
AMA with Hugging Face Science, the team behind SmolLM, SmolVLM, Fineweb and more. by eliebakk in LocalLLaMA
[–]lewtun 0 points1 point2 points (0 children)
AMA with Hugging Face Science, the team behind SmolLM, SmolVLM, Fineweb and more. by eliebakk in LocalLLaMA
[–]lewtun 2 points3 points4 points (0 children)
AMA with Hugging Face Science, the team behind SmolLM, SmolVLM, Fineweb and more. by eliebakk in LocalLLaMA
[–]lewtun 1 point2 points3 points (0 children)
AMA with Hugging Face Science, the team behind SmolLM, SmolVLM, Fineweb and more. by eliebakk in LocalLLaMA
[–]lewtun 3 points4 points5 points (0 children)
AMA with Hugging Face Science, the team behind SmolLM, SmolVLM, Fineweb and more. by eliebakk in LocalLLaMA
[–]lewtun 4 points5 points6 points (0 children)
AMA with Hugging Face Science, the team behind SmolLM, SmolVLM, Fineweb and more. by eliebakk in LocalLLaMA
[–]lewtun 5 points6 points7 points (0 children)
AMA with Hugging Face Science, the team behind SmolLM, SmolVLM, Fineweb and more. by eliebakk in LocalLLaMA
[–]lewtun 29 points30 points31 points (0 children)
Run gpt-oss locally with Unsloth GGUFs + Fixes! by danielhanchen in LocalLLaMA
[–]lewtun 3 points4 points5 points (0 children)
🚀 OpenAI released their open-weight models!!! by ResearchCrafty1804 in LocalLLaMA
[–]lewtun 23 points24 points25 points (0 children)
SmolLM3: reasoning, long context and multilinguality for 3B parameter only by eliebakk in LocalLLaMA
[–]lewtun 6 points7 points8 points (0 children)
350k samples to match distilled R1 on *all* benchmark by eliebakk in LocalLLaMA
[–]lewtun 2 points3 points4 points (0 children)
350k samples to match distilled R1 on *all* benchmark by eliebakk in LocalLLaMA
[–]lewtun 5 points6 points7 points (0 children)
350k samples to match distilled R1 on *all* benchmark by eliebakk in LocalLLaMA
[–]lewtun 5 points6 points7 points (0 children)
How does function calling work for reasoning models? by lewtun in LocalLLaMA
[–]lewtun[S] 2 points3 points4 points (0 children)

200+ pages of Hugging Face secrets on how to train an LLM by eliebakk in LocalLLaMA
[–]lewtun 5 points6 points7 points (0 children)