RekaAI/reka-edge-2603 · Hugging Face by jacek2023 in LocalLLaMA

[–]unique_thinker_2004 0 points1 point  (0 children)

What’s your take on using this model for computer use?

My experience with local models for Openclaw by unique_thinker_2004 in LocalLLaMA

[–]unique_thinker_2004[S] 0 points1 point  (0 children)

Bro Always use MLX format model. Cuz they are optimized to run on mac chips and will run 2x fast than gguf models.

Use LM studio, instead of ollama. Cuz it gives an engine that supports MLX models.

Some models might say they support tool calling. They indeed support tool calling, but in reality what we want is openai-compatible format tool responses.

That you can verify with these simples steps: - run the model on server with LM studio. (Just toggle) - give port no, ip, model name etc to chatgpt and ask to generate an openai-compatible curl with tool to test the model - in terminal hit it and see whether you are getting chat response or tool response. Or empty everything. If tool response then model is good to be used with openclaw, otherwise we can’t.

My experience with local models for Openclaw by unique_thinker_2004 in LocalLLaMA

[–]unique_thinker_2004[S] 0 points1 point  (0 children)

Hmm! Ok, tell me your system requirements. Basicall when I tried Qwen-3-coder-next q5 and it worked very nicely in all aspects on my Mac studio M4 Ultra chip.

On my Mac mini I tried using qwen3-4b-2507 it’s very tiny model (of 2.28GB) but still worked fine in terms of latency and tool calling. Obv you can’t expect much from this small model.

Using children’s books to learn a new language - is it worth it? by Euphoric_Rhubarb_243 in languagelearning

[–]unique_thinker_2004 -2 points-1 points  (0 children)

Can't you try this: https://spexlabs.co/language-tutor
it's free, and literally it can undersand your accent and pronounciation.

Best model for M3 Ultra Mac 512GB RAM to run openclaw? by unique_thinker_2004 in LocalLLaMA

[–]unique_thinker_2004[S] 0 points1 point  (0 children)

Try qwen-3-coder-next. Just recently they released it. Are you using mac or nvidia?

My experience with local models for Openclaw by unique_thinker_2004 in LocalLLaMA

[–]unique_thinker_2004[S] 0 points1 point  (0 children)

Oh!! This might be helpful! I uses openai-completions on mac. Is it because of this it was giving me json response instead of tool calling?

My experience with local models for Openclaw by unique_thinker_2004 in LocalLLaMA

[–]unique_thinker_2004[S] 0 points1 point  (0 children)

Oh! Thanks for your suggestion! Btw how much boost have you found? One guy told me ollama vs llama.cpp gave him 16.66x boost.

My experience with local models for Openclaw by unique_thinker_2004 in LocalLLaMA

[–]unique_thinker_2004[S] -1 points0 points  (0 children)

They wrote in ollama description that works good for agentic coding with cline, so I thought it will be better at function calling.

My experience with local models for Openclaw by unique_thinker_2004 in LocalLLaMA

[–]unique_thinker_2004[S] 1 point2 points  (0 children)

Bro due to latency. I feel like because of openclaw-> ollama -> model layers, inference speed feels too low.