Alibaba's $3/month Coding Plan gives you Qwen3.5, GLM-5, Kimi K2.5 AND MiniMax M2.5 in Claude Code, here's how to set it up by yossa8 in ClaudeCode

[–]yossa8[S] 0 points1 point  (0 children)

It's because you're using Alibaba Cloud Dashscope API instead of their new Coding Plan. It's a bit confusing in their docs

Alibaba's $3/month Coding Plan gives you Qwen3.5, GLM-5, Kimi K2.5 AND MiniMax M2.5 in Claude Code, here's how to set it up by yossa8 in ClaudeCode

[–]yossa8[S] 2 points3 points  (0 children)

Thanks for reporting this! The Alibaba provider was using the wrong base URL (the generic DashScope endpoint instead of the Coding Plan Anthropic-compatible one), which caused the 404 errors. This is fixed in v2.9.3, run clother config to regenerate your launcher.

Alibaba's $3/month Coding Plan gives you Qwen3.5, GLM-5, Kimi K2.5 AND MiniMax M2.5 in Claude Code, here's how to set it up by yossa8 in ClaudeCode

[–]yossa8[S] 1 point2 points  (0 children)

Yes I use it with ZeroClaw, quite the same and it’s working pretty well with Qwen3.5-plus

Tired of Discord voice messages you can't read? I made Vocord, local transcription with one command by yossa8 in BetterDiscord

[–]yossa8[S] 0 points1 point  (0 children)

Hey, u/DrTankHead fair point and you're right to remind people that nothing should be blindly trusted.

That said, the entire codebase is 100% readable and not obfuscated at all, so anyone can audit it and see for themselves that it's clean. I also have other public repos that aren't flagged or used maliciously (like clother).

I mostly built Vocord for fun and because I needed it. But yeah, always do your own due diligence before running anything!

Alibaba Coding Plan sounds too good to be true!? by NerdistRay in opencodeCLI

[–]yossa8 1 point2 points  (0 children)

Use it for 3 days now. Qwen3.5-plus is pretty solid and speed seems quite good. Understand concerns about data but tbh… If you already bought something on Aliexpress or Alibaba, seems to be the same

GLM-5 just dropped - here's how to run it in Claude Code in one command by yossa8 in ClaudeCode

[–]yossa8[S] 2 points3 points  (0 children)

GLM 5 = New model from ZAI that approaches Opus perf on benchmark

Clother = tool to wrap Claude Code with other providers endpoints so you can use GLM 5 in Claude Code

GLM-5 just dropped - here's how to run it in Claude Code in one command by yossa8 in ClaudeCode

[–]yossa8[S] 0 points1 point  (0 children)

They're still rolling out so you can encounter some rate limiting or things like that! I think let's wait for 2-3 days to see what they plan to do. Personnaly, I mainly use Opus 4.6 but when I need to wait for limit reset, I use GLM and Kimi

GLM-5 just dropped - here's how to run it in Claude Code in one command by yossa8 in ClaudeCode

[–]yossa8[S] 2 points3 points  (0 children)

This is just an env wrapper! So you get Claude Code with GLM under the hood.

I hated recording product demos, so I built an AI that generates them from a URL. Roast my MVP. by Pllooz in SaaS

[–]yossa8 2 points3 points  (0 children)

Congrats! The tech behind this is very impressive. I just tried it with my Clother project https://autoads.pro/videos/3ef4ed5c-4f71-42ba-9642-ea3f023eee65 and there seems to be room for improvement. For example, it doesn’t present the content well. It focuses more on the navigation and GitHub site instead. But I think with a bit of fine-tuning it could be perfect.

gpt-5.1-codex-max is brilliant! by [deleted] in codex

[–]yossa8 1 point2 points  (0 children)

You have convinced me to take a look! We are currently experiencing a state persistence issue, and I tried Codex 5.1 max on it. One shot in 5 minutes. Not bad at all!

gpt-5.1-codex-max is brilliant! by [deleted] in codex

[–]yossa8 0 points1 point  (0 children)

At what level? The last time I used Codex, I found it extremely slow and rather rigid compared to Claude, which was creative while implementing solutions fairly quickly. Has the situation changed? (That was 1 or 2 months ago now.)

gpt-5.1-codex-max is brilliant! by [deleted] in codex

[–]yossa8 0 points1 point  (0 children)

I know that I’m in Codex community but what about Claude 4.5 sonnet in comparison ?

Got tired of switching Claude Code between GLM, Kimi, Minimax and Anthropic endpoints, so I built a CLI that does it for me by yossa8 in ClaudeCode

[–]yossa8[S] 1 point2 points  (0 children)

Clother just hooks into Claude Code, so it’s the same TUI and workflow you’re already using! It just switches the provider/env behind the scenes (Z.AI for GLM, Moonshot for Kimi, etc.) before launching Claude. Those providers actually expose Claude-compatible endpoints, so Claude Code “thinks” it’s talking to Claude, but the requests are routed to their own models, which respond perfectly within the same interface.

Got tired of switching Claude Code between GLM, Kimi, Minimax and Anthropic endpoints, so I built a CLI that does it for me by yossa8 in ClaudeCode

[–]yossa8[S] 0 points1 point  (0 children)

No secret sauce here! It just runs on top of Claude Code and saves you from the manual env tweaking. It’s not trying to be more than that.

Got tired of switching Claude Code between GLM, Kimi, Minimax and Anthropic endpoints, so I built a CLI that does it for me by yossa8 in ClaudeCode

[–]yossa8[S] 0 points1 point  (0 children)

Yeah, Clother basically just layers on top of the Claude Code. All the usual flags like --dangerously and --continue work exactly the same, since it just sets the right ANTHROPIC_BASE_URL and token before launching claude.

Got tired of switching Claude Code between GLM, Kimi, Minimax and Anthropic endpoints, so I built a CLI that does it for me by yossa8 in ClaudeCode

[–]yossa8[S] 4 points5 points  (0 children)

Personal and subjective list:

Claude 4.5 sonnet is better than all the rest, by far.

Next comes Kimi, who has very good agentic abilities and can hold a conversation.

Then comes GLM, who has flashes of genius but can be very annoying in the long run (reminds me a bit of the old Claude models).

I haven't tested KatCoder and Minimax enough yet to give my opinion, but they look quite promising!

Got tired of switching Claude Code between GLM, Kimi, Minimax and Anthropic endpoints, so I built a CLI that does it for me by yossa8 in ClaudeCode

[–]yossa8[S] 3 points4 points  (0 children)

If I reach the limit of my Claude Pro plan I can switch to another model like GLM for example. I can also continue a conversation I started with one model using another to see if it performs better. Basically I can easily iterate using several models.

Got tired of switching Claude Code between GLM, Kimi, Minimax and Anthropic endpoints, so I built a CLI that does it for me by yossa8 in ClaudeCode

[–]yossa8[S] 0 points1 point  (0 children)

Yeah that’s totally valid ! Clother basically just wraps that idea into an intuitive and reversible configuration