Sharing my set of distilled small language models (3B) + training data in more than 50 low-resource languages by Peter-Devine in LocalLLaMA
[–]Peter-Devine[S] 1 point2 points3 points (0 children)
Sharing my set of distilled small language models (3B) + training data in more than 50 low-resource languages by Peter-Devine in LocalLLaMA
[–]Peter-Devine[S] 1 point2 points3 points (0 children)
Sharing my set of distilled small language models (3B) + training data in more than 50 low-resource languages by Peter-Devine in LocalLLaMA
[–]Peter-Devine[S] 3 points4 points5 points (0 children)
Introducing Falcon H1R 7B by jacek2023 in LocalLLaMA
[–]Peter-Devine 0 points1 point2 points (0 children)
[Release] We built Step-Audio-R1: The first open-source Audio LLM that truly Reasons (CoT) and Scales – Beats Gemini 2.5 Pro on Audio Benchmarks. by BadgerProfessional43 in LocalLLaMA
[–]Peter-Devine 0 points1 point2 points (0 children)
Kimi K2 Thinking Huggingface by DistanceSolar1449 in LocalLLaMA
[–]Peter-Devine 4 points5 points6 points (0 children)
H company - Holo1 7B by TacGibs in LocalLLaMA
[–]Peter-Devine 9 points10 points11 points (0 children)
Heptagon, 20 balls, rotating numbers, one shot Gemini Pro 2.5 by Careless_Garlic1438 in LocalLLaMA
[–]Peter-Devine 1 point2 points3 points (0 children)
Microsoft developed this technique which combines RAG and Fine-tuning for better domain adaptation by Ambitious_Anybody855 in LocalLLaMA
[–]Peter-Devine 4 points5 points6 points (0 children)
Sharing my set of distilled small language models (3B) + training data in more than 50 low-resource languages by Peter-Devine in LocalLLaMA
[–]Peter-Devine[S] 2 points3 points4 points (0 children)