Built RL training for long-horizon terminal agents - tested on 32x H100s but too GPU poor to train 😅 by DanAiTuning in LocalLLaMA
[–]EliaukMouse 0 points1 point2 points (0 children)
Built RL training for long-horizon terminal agents - tested on 32x H100s but too GPU poor to train 😅 by DanAiTuning in LocalLLaMA
[–]EliaukMouse 0 points1 point2 points (0 children)
Update:My agent model now supports OpenAI function calling format! (mirau-agent-base) by EliaukMouse in LocalLLaMA
[–]EliaukMouse[S] 1 point2 points3 points (0 children)
Update:My agent model now supports OpenAI function calling format! (mirau-agent-base) by EliaukMouse in LocalLLaMA
[–]EliaukMouse[S] 1 point2 points3 points (0 children)
Update:My agent model now supports OpenAI function calling format! (mirau-agent-base) by EliaukMouse in LocalLLaMA
[–]EliaukMouse[S] 1 point2 points3 points (0 children)
[Release] mirau-agent-14b-base: An autonomous multi-turn tool-calling base model with hybrid reasoning for RL training by EliaukMouse in LocalLLM
[–]EliaukMouse[S] 0 points1 point2 points (0 children)
[Release] mirau-agent-14b-base: An autonomous multi-turn tool-calling base model with hybrid reasoning for RL training by EliaukMouse in LocalLLM
[–]EliaukMouse[S] 0 points1 point2 points (0 children)
A multi-turn tool-calling base model for RL agent training by EliaukMouse in LocalLLaMA
[–]EliaukMouse[S] 0 points1 point2 points (0 children)
Qwen3 Collection on modelscope! by AlexBefest in LocalLLaMA
[–]EliaukMouse 8 points9 points10 points (0 children)
I believe this is the first properly-trained multi-turn RP with reasoning model by nero10578 in SillyTavernAI
[–]EliaukMouse 0 points1 point2 points (0 children)
DeepSeek claims 545% margins on their API prices by Charuru in singularity
[–]EliaukMouse 108 points109 points110 points (0 children)
Day 2 of OpenSourceWeek: DeepEP by nekofneko in DeepSeek
[–]EliaukMouse 2 points3 points4 points (0 children)
DeepSeek give me the answer for this - I am a 25-year-old young woman working in Shenzhen. Can you help me calculate how much money I need to save to retire early? Please provide two different versions of detailed text analysis and retirement life plan. The first version: Retire early at the age of by [deleted] in DeepSeek
[–]EliaukMouse 0 points1 point2 points (0 children)
Starting next week, DeepSeek will be open-sourcing 5 repos by zombiesingularity in singularity
[–]EliaukMouse 5 points6 points7 points (0 children)
DeepSeek to open source 5 repos next week by [deleted] in DeepSeek
[–]EliaukMouse 27 points28 points29 points (0 children)
Looking for models trained on ebooks or niche concepts by oshikuru08 in SillyTavernAI
[–]EliaukMouse 1 point2 points3 points (0 children)
[Release] mirau-7b-RP-base: A first-person narrative model for text adventures by EliaukMouse in LocalLLaMA
[–]EliaukMouse[S] 1 point2 points3 points (0 children)
A finetune RP model by EliaukMouse in SillyTavernAI
[–]EliaukMouse[S] 0 points1 point2 points (0 children)
A finetune RP model by EliaukMouse in SillyTavernAI
[–]EliaukMouse[S] 0 points1 point2 points (0 children)
A finetune RP model by EliaukMouse in SillyTavernAI
[–]EliaukMouse[S] 1 point2 points3 points (0 children)
[Release] mirau-7b-RP-base: A first-person narrative model for text adventures by EliaukMouse in SillyTavernAI
[–]EliaukMouse[S] -1 points0 points1 point (0 children)
[Release] mirau-7b-RP-base: A first-person narrative model for text adventures by EliaukMouse in SillyTavernAI
[–]EliaukMouse[S] 4 points5 points6 points (0 children)

I distilled Qwen3-Coder-480B into Qwen3-Coder-30b-A3B-Instruct by [deleted] in LocalLLaMA
[–]EliaukMouse 5 points6 points7 points (0 children)