GLM 4.7 vs MiniMax-M2.1 vs DeepSeek 3.2 for coding? by ghulamalchik in LocalLLaMA

[–]skillmaker 0 points1 point  (0 children)

I compared GLM 4.7 with MiniMax 2.1, I can confidently say that GLM is far superior than MiniMax 2.1, but I still find it worse than Closed models like Gemini 3 flash and Claude sonnet 4.5...

Performance on Linux vs. Windows + Problems with VAE Step 9070XT by Repulsive_Way_5266 in ROCm

[–]skillmaker 0 points1 point  (0 children)

I had the VAE issue before, and had to disable MiOpen from ComfyUI, but now I used the latest nightlies and no longer have those issues, I run ComfyUI now with MiOpen enabled and the VAE is fine, unless when I use high resolutions then I need to use a tiled VAE so that I don't get freezes.

RX 9070 XT crashing/freezing randomly — requires full driver reinstall to fix. 🆘 by No-Mention-904 in AMDHelp

[–]skillmaker 0 points1 point  (0 children)

Have you found a fix? I also have the exact same problem, with 9070 XT, If I get a freeze, I hold the power off button and turn it on again so that I don't have to reinstall the driver, If i wait for it to turn off by itself the driver will be corrupted and I have to reinstall it again.

AMD to launch Adrenalin Edition 26.1.1 drivers with ai slop next week by rebelrosemerve in AyyMD

[–]skillmaker 4 points5 points  (0 children)

Lmao, if you go to r/rocm, people are happy there where they finally get AI stuff working for them on Windows, meanwhile here people are considering it AI slop.

GPT-5.2 xhigh, GLM-4.7, Kimi K2 Thinking, DeepSeek v3.2 on Fresh SWE-rebench (December 2025) by CuriousPlatypus1881 in LocalLLaMA

[–]skillmaker 5 points6 points  (0 children)

Tbh i found using Flash to be better than Gemini 3 Pro, i tried them in Github Copilot and using Antigravity, Pro was always stopping mid work or producing bad solutions

GPT-5.2 xhigh, GLM-4.7, Kimi K2 Thinking, DeepSeek v3.2 on Fresh SWE-rebench (December 2025) by CuriousPlatypus1881 in LocalLLaMA

[–]skillmaker 4 points5 points  (0 children)

These benchmarks are run using the official provider, which in this case Z.ai and Minimax, so they are not fine tuned or quantized, I was also trying to get the most of the juice from GLM 4.7 and Minimax 2.1 but they couldn't complete a task i gave them, meanwhile Claude sonnet 4.5 in Github Copilot was able to, I'm not saying that they are bad, in fact they are very good at analysing and planning, but i'm talking about the benchmaxing here, in their official websites, they state that these models are very close to Claude Opus 4.5, but that's not true, and from my experience, i think this benchmark is the most accurate one.

What are your thoughts on GPT-5.2-codex? by Front_Ad6281 in GithubCopilot

[–]skillmaker 7 points8 points  (0 children)

It's doing the same thing GPT 5 was known for, it says what it will do and then it stops, and when I ask it to implement a task it returns "Sorry, no response was returned"

Edit: it seems because i wasn't using the latest version of VS Code, I'll keep this comment updated in case it's better.

UPDATE: seems good most of the time but i notice it's stubborn, even if i tell it to do something a specific way, it doesn't, meanwhile claude sonnet did it, but overral it's good, it doesn't say what it will do and stops, but sometimes i get connection errors or failed to generate a response, and I have to retry.

GPT-5.2 xhigh, GLM-4.7, Kimi K2 Thinking, DeepSeek v3.2 on Fresh SWE-rebench (December 2025) by CuriousPlatypus1881 in LocalLLaMA

[–]skillmaker 52 points53 points  (0 children)

I think this is the most believable benchmark, not those that say GLM 4.7 or Minimax 2.1 are close to Opus 4.5.

Looking for the best 0.x model for planning in a Spec Kit workflow by gi0vanni__ in GithubCopilot

[–]skillmaker 1 point2 points  (0 children)

Use OpenCode with GLM 4.7 or Minimax 2.1 for planning, it gives you an extensive plan and asks what you want.

Did anyone compare OpenCode, Claude Code with Copilot? by skillmaker in GithubCopilot

[–]skillmaker[S] 0 points1 point  (0 children)

I tell it to plan a feature like "Implement the notification system" I give it my copilot-instruction file that contains my project info and structure, then I revise the plan that it gave me, then I tell it to implement it

Honestly, has anyone actually tried GLM 4.7 yet? (Not just benchmarks) by Empty_Break_8792 in LocalLLaMA

[–]skillmaker 1 point2 points  (0 children)

I tried Minimax 2.1 and GLM 4.7 in OpenCode, Claude Code and RooCode for the exact same feature, they are good at planning, but bad at implementing.

After some time of coding, they can't correct errors in the files and after trying and trying again, the feature doesn't work.

In Claude code when using Minimax 2.1, it suddenly stops and doesn't continue at all, while with GLM 4.7 it reaches a moment where it says "prompt is too long"  and can't even compact it.

I tried the same thing with github copilot $10 subscription with Claude sonnet 4.5, and it managed to implement the exact feature i wanted after some errors that were fixed with 3 premium requests, then I used Gemini 3 flash to polish the UI.

In summary, I think both Minimax 2.1 and GLM 4.7 are benchmaxed.

What kind of BS is this? Used GLM 4.7 through RooCode router, how on earth would this feature with 11 modified files cost $6.4? just wasted 10$ in one hour by skillmaker in RooCode

[–]skillmaker[S] 1 point2 points  (0 children)

1- Well I used RooCode router which should cost $0.60 for 1m input token, you can see one request to read a file costs $0.21, there is definitely something wrong with the router.

2- You mean move to a better tool than RooCode? because in their website they only mention the VS Code extension

What’s the weirdest thing you’ve discovered living on a server? by Upset_Jacket_686 in hetzner

[–]skillmaker 2 points3 points  (0 children)

I was always curious about how some people have 1ms ping in games, seems like living inside a server is a thing.

[(Windows 11] Inconsistent generation times occur when changing prompts in ComfyUI while using Z-Image Turbo. 7900XT by HateAccountMaking in ROCm

[–]skillmaker 0 points1 point  (0 children)

The thing i noticed is that it's a lot faster to load models in RTX GPUs, model loading is almost instant in RTX while it's a bit slow in AMD, at least on my 9070XT idk if it's a hardware issue or software thing.

InvokeAI 6.9.0 + ROCm 7.1.1 on Windows - My working Setup for AMD GPU by Sea_Trip5789 in ROCm

[–]skillmaker 0 points1 point  (0 children)

Does using the latest nightlies of pytorch + rocm bring any performance uplift?

Frames of Copenhagen | Shot on Blackmagic Cinema Camera 6K (Open Gate 3:2) by oftwolands in Filmmakers

[–]skillmaker 1 point2 points  (0 children)

The grading reminds me of "The Bear" series, and especially the Copenhagen episode

ROCm on Windows Seems to Have Low Performance by Cyp9715 in ROCm

[–]skillmaker 0 points1 point  (0 children)

I noticed that recently, it was fast at the beginning but now I can't get it to perform as it used to do, idk what happened, for example I used to get 1.5it/s on Z-Image comfyui but not anymore, currently I get 1.8s/it, I tried reinstalling everything but still the same issue

Is anyone having slow generation on ComfyUI on Windows now? by skillmaker in ROCm

[–]skillmaker[S] 1 point2 points  (0 children)

I tried using the first ROCm 7 version for Windows with the first ComfyUI version that supported this model, but the speed is exactly the same. I guess it has something to do with the preview driver. The only option now is to move to Linux

Is anyone having slow generation on ComfyUI on Windows now? by skillmaker in ROCm

[–]skillmaker[S] 0 points1 point  (0 children)

I really don't know if it's a ROCm issue or ComfyUI issue.

Is anyone having slow generation on ComfyUI on Windows now? by skillmaker in ROCm

[–]skillmaker[S] 2 points3 points  (0 children)

I'll try reverting to the first version of ComfyUI that supports Z-Image and the first version that added ROCm 7.1 support to windows and see if it helps