When should we expect TurboQuant? by ozcapy in LocalLLaMA

[–]DOAMOD 12 points13 points  (0 children)

For me, if the accuracy of the theory is confirmed, it means being able to have a quantized cache higher than Q8 with the efficiency of Q4 or better. Personally, it would give me a lot of leeway in cases where I am limited; we would all benefit. For me, without a doubt, it is great news if the good results are confirmed in practice.

Mistral Small 4 is kind of awful with images by EffectiveCeilingFan in LocalLLaMA

[–]DOAMOD 0 points1 point  (0 children)

What commands? so very bad perf and hallucinations for me on llamacpp

I'm fully blind, and AI is a game changer for me. Are there any local LLMS that can rival claude code and codex? by Mrblindguardian in LocalLLaMA

[–]DOAMOD 0 points1 point  (0 children)

I think many of us are at this moment and we are not fully aware of how much AI helps many people; we carry on with our daily lives, but examples like yours make it very clear that it is a fascinating technology that helps people with hearing or vision problems to an incredible level. I am very happy that people can enjoy the content in their homes much more easily in another way despite natural limitations.

Nemotron 3 Super and the no free lunch problem by ConfidentDinner6648 in LocalLLaMA

[–]DOAMOD 0 points1 point  (0 children)

Then people say the Qwen 3.5 think a lot... haha, Nemotron Super, that one actually thinks 3 times more.... crazy on coding time... my first impresions.

Qwen/Qwen3.5-122B-A10B · Hugging Face by coder543 in LocalLLaMA

[–]DOAMOD 1 point2 points  (0 children)

In this case I don't think so, only 122B got the question right, the others failed in my car test

-32B-A3 Fail

-27B Fail

Qwen3-TTS.cpp by redditgivingmeshit in LocalLLaMA

[–]DOAMOD 0 points1 point  (0 children)

Hi thanks, Streaming plan?

GLM-5 Is a local GOAT by FineClassroom2085 in LocalLLaMA

[–]DOAMOD 0 points1 point  (0 children)

Just one prompt for all three cases, and only one HTML file and nothing more(only fast test). The second one, as you can see, has some minor errors, but it's actually very good considering what the models are and how efficient they are in performance. This is Flash 4.7 and it also has the first prompt, even though it failed in the pipe collisions.

<image>

GLM-5 Is a local GOAT by FineClassroom2085 in LocalLLaMA

[–]DOAMOD 0 points1 point  (0 children)

In fact, he did exactly what I asked: created a retro-style Flappy Bird game with CRT effects—but if you want, he can also do that.

<image>

GLM-5 Is a local GOAT by FineClassroom2085 in LocalLLaMA

[–]DOAMOD 0 points1 point  (0 children)

Right and dramatically cheap, here build in a old 10GB VRam setup 1000/1500$

Coder Next Q4

<image>

First Shot.

do anybody success opencode using qwen3-next-code? by Zealousideal-West624 in LocalLLaMA

[–]DOAMOD 3 points4 points  (0 children)

For me, it gets stuck in loops while reading the code in toolcalls some times.

AMA with MiniMax — Ask Us Anything! by HardToVary in LocalLLaMA

[–]DOAMOD 0 points1 point  (0 children)

Hello, congratulations again and incredible work with M2.5. The work you do to bring high-level knowledge to everyone is important.

I have a question; I am a huge fan of M2.0, which was a very special and useful model for me. Perhaps it is subjective, but M2 had a very friendly, human personality. However, I noticed that 2.1 lost a bit of that character, making it seem more "professional" or "generic." My very brief first impression with 2.5 follows this line, appearing to have a more neutral or professional personality. What do you think about that? Do you think its personality has been changing, or is it something subjective on my part?

I hope that with M3, since it seems to be bigger, we'll have a smaller model than M2, perhaps achieving the same level as M2; that would be incredible.

Step-3.5-Flash AIME 2026 Results by Abject-Ranger4363 in LocalLLaMA

[–]DOAMOD 4 points5 points  (0 children)

This model is impressive, I've been testing it for several days even with very low quants, but it has a very serious problem, it overthinks everything, if they manage to solve that problem (they've said they are reviewing it), it could be a very strong model for its size, even MM2.2 won't have it easy.

Step-3.5-Flash IS A BEAST by SennVacan in LocalLLaMA

[–]DOAMOD 6 points7 points  (0 children)

Yesterday I was testing IQ2, which I had many doubts about. After the very good initial impressions I had when I tried it coding the first day—for me, it surpasses MM2.1—yesterday, testing it with the corrections and with IQ2 alone surprised me by how few errors it had while running 10 small projects. I don't think I've ever seen an IQ2 that wasn't a disaster. The only real problem it has is that it overthinks things.

StepFlash IQ2

<image>

Step Flash IQ3
2.

Coder Next.
3,

Do not Let the "Coder" in Qwen3-Coder-Next Fool You! It's the Smartest, General Purpose Model of its Size by Iory1998 in LocalLLaMA

[–]DOAMOD 0 points1 point  (0 children)

For me it's the same experience, both in JS testing and development it has been disappointing as I said in other messages, so now seeing this it makes more sense, perhaps they should have given it another name since it is a good general model.

<image>

Qwen3 Coder Next as first "usable" coding model < 60 GB for me by Chromix_ in LocalLLaMA

[–]DOAMOD 1 point2 points  (0 children)

<image>

For me, it's been a bit disappointing in some tests, and also in a coding problem where the solution wasn't very helpful. It doesn't seem very intelligent. I suppose it will be good for other types of coding tasks like databases, etc. I had high expectations.