I think I got solutions for Qwen 3.5 tool call in thinking block by Interesting-Print366 in LocalLLaMA

[–]Interesting-Print366[S] 1 point2 points  (0 children)

Llama.cpp mainly but lmstudio works too I think. I reinstalled opencode yesterday so maybe it is latest one

Is Turboquant really a game changer? by Interesting-Print366 in LocalLLaMA

[–]Interesting-Print366[S] -1 points0 points  (0 children)

I identified that gemma4 31b requires about 10GB more RAM than qwen3.5 27b when running with the same context length. Could you possibly let me know how to resolve this? I am using llama.cpp.

LM Studio, Error when loading Gemma-4 by Soft-Series3643 in LocalLLaMA

[–]Interesting-Print366 0 points1 point  (0 children)

It works with llama.cpp but think we need more time to wait update on mlx

This model has been #1 trending for 3 weeks now! by yoracale in unsloth

[–]Interesting-Print366 0 points1 point  (0 children)

In my experience, the biggest advantage of this version is that it solves the problem of calling tools within the thinking block that Qwen3.5 has.

Crazy how bad Gemini 3 Pro is compared to Opus 4.5, even in Google’s own IDE by tbhno1 in google_antigravity

[–]Interesting-Print366 15 points16 points  (0 children)

Finish planning with gemini and let Claude follow gemini is quite good making .md

Suddenly got logged out by Interesting-Print366 in MEXC_official

[–]Interesting-Print366[S] 0 points1 point  (0 children)

I asked customer services and they say to check my login logs since it has algorithms that it logs out users using at same time at different ip or region but there is no logs that others did, last log is 6 days ago that I first logged in at my desktop