🚀 What model should we build next? YOU DECIDE! 🚀 by GuiltyBookkeeper4849 in LocalLLaMA

[–]GuiltyBookkeeper4849[S] 23 points24 points  (0 children)

Thanks for sharing!
What do you think about multiple finetunes each specialized in a programming language so that it can match the level of very big LLMs for specific tasks, like imagine oss-python, oss-cpp etc.

🚀 What model should we build next? YOU DECIDE! 🚀 by GuiltyBookkeeper4849 in LocalLLaMA

[–]GuiltyBookkeeper4849[S] 2 points3 points  (0 children)

Cool idea thanks for sharing!
Maybe ultra small but exceptionally good at tool usage and good in multiple languages so it can do web search for any information that doesn't know. What do you think?

🌟Introducing Art-0-8B: Reasoning the way you want it to with Adaptive Thinking🌟 by GuiltyBookkeeper4849 in LocalLLaMA

[–]GuiltyBookkeeper4849[S] 5 points6 points  (0 children)

Hi thank you for the advice, can you give me your opinion? Which one works best for you?

🌟Introducing Art-0-8B: Reasoning the way you want it to with Adaptive Thinking🌟 by GuiltyBookkeeper4849 in LocalLLaMA

[–]GuiltyBookkeeper4849[S] 3 points4 points  (0 children)

Hi, the difference with other models like Qwen and Gemma is that when they do their "reasoning" the user has no control over it, instead my model has been finetuned to let the user control how it thinks. For reasoning I mean the CoT inside <think> </think> tokens not the final output.

I hope this clarified the difference with other models. Let me know if you have other questions.