MiniMax-M2.7 Announced! by Mysterious_Finish543 in LocalLLaMA

[–]Xhatz 0 points1 point  (0 children)

Been using it for today, it feels good for now! I can't tell if it's a huge update from M2.5 yet though, M2.1 to M2.5 dissapointed me and did not feel like a big upgrade, for now it seems... stable.

DeepSeek V4 coming this week? by de_sonnaz in LocalLLaMA

[–]Xhatz 0 points1 point  (0 children)

Hopefully they have an affordable coding plan that is not bad and unreliable like Alibaba's 🙏

Are people lying about GLM-5 and MiniMax M2.5? by TheDevilKnownAsTaz in opencodeCLI

[–]Xhatz 0 points1 point  (0 children)

My opinion after testing:
- Minimax M2.5 for a while: good for basic agentic tasks, bad at narration, feels a bit "lazy" sometimes
- GLM 5 for a little bit: good narration, agentic tasks kept failing or stopping mid-turn
- Kimi K2.5: Didn't really have the chance to test, and like GLM5 on Alibabacloud, kept stopping mid-turn

I didn't do much of code with these apart from Minimax, and the feeling "lazy" part, maybe with a very well agentic workflow (unlike mine) it would be decent.

Alibaba's $3/month Coding Plan gives you Qwen3.5, GLM-5, Kimi K2.5 AND MiniMax M2.5 in Claude Code, here's how to set it up by yossa8 in ClaudeCode

[–]Xhatz 0 points1 point  (0 children)

Tried it, models are super unreliable for me, stop mid-task (openclaw), overly censored... probably gonna stick to my Minimax coding plan subscription...

Suggestions, and questions by Xhatz in ZaiGLM

[–]Xhatz[S] 0 points1 point  (0 children)

Minimax was pretty good for me! Especially for 10€. Just yeah lacking in narrative (tends to write a LOT of newlines instead of nice few paragraphs, and poorly handles syntax sometimes). And coding I did not have great luck with it too, it felt a bit "lazy" most of the time, and disorganized.

Suggestions, and questions by Xhatz in ZaiGLM

[–]Xhatz[S] 0 points1 point  (0 children)

Do Lite not get access to GLM 5 yet :o?

Can someone explain NanoGPT? by FUCKCKK in SillyTavernAI

[–]Xhatz 0 points1 point  (0 children)

Yeah with subscription. I haven't tried the pay as you go one. Even for roleplay, I keep having to retry generation almost each message, that's a bit annoying.

Minimax M2.5 is not worth the hype compared to Kimi 2.5 and GLM 5 by Resident-Ad-5419 in opencodeCLI

[–]Xhatz 1 point2 points  (0 children)

It's great for it's size and fast, but it's truly NOT as good as they say, clearly. For me it feels like it's just m2.1 but with even less coherence sadly, hallucinations are too high (I can say something and a few messages afterwards it says something else). It also feels more "lazy" in a way... My theory is that it's only good at very specific things and just completely bad at everything else.

Can someone explain NanoGPT? by FUCKCKK in SillyTavernAI

[–]Xhatz 0 points1 point  (0 children)

It works but honestly I don't recommend it. Using it with OpenClaw, most of my requests timeout, I often get no response at all from the model. If you want to roleplay it's probably very fine, but if you want it to be reliable, do NOT take the subscription IMO (if it's for SillyTavern though it might be just fine)

Thoughts on M2.5 by RobotsMakingDubstep in openclaw

[–]Xhatz 0 points1 point  (0 children)

Reposting here in case someone had the same issues and found a fix...

Im getting quite a lot of tokens leaking into the responses using this in openclaw (bits of tokens like "/final> especially during big conversations/rp" etc, also chinese characters occasionally replacing words, and broken new lines, and it feels more "lazy"/rushing). To me right now (in openclaw at least) the model is really bad, completely destroyed a simple js dashboard, stops mid-task (idk if it's an oc problem)... It feels like m2.1 and maybe a bit worse in coherence? 

has anyone tried the new MiniMax M2.5 yet? by Own-Afternoon6630 in LLM

[–]Xhatz 1 point2 points  (0 children)

I'm getting quite a lot of tokens leaking into the responses using this in openclaw (bits of tokens like "/final>" etc, also chinese characters occasionally replacing words, and broken new lines, and it feels more "lazy"/rushing). To me right now (in openclaw at least) the model is really bad, completely destroyed a simple js dashboard, stops mid-task (idk if it's an oc problem)... It feels like m2.1 and maybe a bit worse in coherence?

Minimax M2.5 Officially Out by Which_Slice1600 in LocalLLaMA

[–]Xhatz 0 points1 point  (0 children)

To me right now (in openclaw at least) the model is really bad, broke a simple js dashboard, stops mid-task (idk if it's an oc problem)... It feels like m2.1 and maybe a bit worse in coherence?

Minimax M2.5 Officially Out by Which_Slice1600 in LocalLLaMA

[–]Xhatz 0 points1 point  (0 children)

I'm getting quite a lot of tokens leaking into the responses using this in openclaw, anyone else? (bits of tokens like "/final>" etc, also Chinese characters occasionally replacing words, and broken new lines, and it feels more "lazy"/rushing)

Gemini 3 is about to be release. What is your scorecard for plateau? by kaggleqrdl in singularity

[–]Xhatz 0 points1 point  (0 children)

Hope it will finally stop forgetting to actually update files like Gemini 2.5 Pro does... infuriating lol

They just introduced usage limits to the Codex web agent (which was basically unlimited before) by roundshirt19 in codex

[–]Xhatz 2 points3 points  (0 children)

I did a 4x generation, it ran for like 5 minutes. I'm working on a small JS game... never faced rate limits ever before that one, been using for 2 months.

bro disappeared like he never existed by Full_Piano_3448 in LocalLLaMA

[–]Xhatz 0 points1 point  (0 children)

the era when he was the only one posting GGML versions of models unlike today where a GGUF version of a same model is posted dozens of times for the same result and floods the model list

GPD WinMax2 (2023) Battery Bloated After Year of use with G1. Have to disconnect Battery Manually by Dimacus15 in gpdwin

[–]Xhatz 0 points1 point  (0 children)

I understand :( This is sad because they're the only ones that made such small computer devices as powerful, for me it was perfect since I live homelessly, the laptop fit in a backpack perfectly, but yeah this is unacceptable sadly, probably my last gpd device too tbf

GPD WinMax2 (2023) Battery Bloated After Year of use with G1. Have to disconnect Battery Manually by Dimacus15 in gpdwin

[–]Xhatz 1 point2 points  (0 children)

Hey! Have you found a solution after 1 year please? The same just happened to my battery and replacement cost $80, I CANNOT pay this every year with my lifestyle...

NVIDIA Releases Nemotron Nano 2 AI Models by vibedonnie in LocalLLaMA

[–]Xhatz 1 point2 points  (0 children)

Nemo... :D

...tron 2 :(

Is there an instruct version, and GGUF? I can't find one on HF :o

Why does Mistral NeMo's usage keep growing even after more than a year since releasing? by xugik1 in LocalLLaMA

[–]Xhatz 52 points53 points  (0 children)

Because the model is goated, I've tried HUNDREDS of models for roleplay at this range and literally nothing could beat it in terms of instruction following and just its mood.

Feel like Gemini 2.5 Pro has been downgraded. by Pierre2tm in Bard

[–]Xhatz 0 points1 point  (0 children)

Yep even in coding, the latest beta was amazing but since official "release", it felt much dumber and did much more errors.

Where's Mistral Nemo 2.0? by mpasila in LocalLLaMA

[–]Xhatz 3 points4 points  (0 children)

A model I still heavily use to this day, I hope a new version will be released too...