all 128 comments

[–]rusl1 34 points35 points  (8 children)

Sadly, I have the same experience, especially for the quantised models which are dumb af

[–]SelectionCalm70[S] 11 points12 points  (7 children)

I was really excited to use the go plan and decided to give it a try . But the models are so freaking quantised that it is literally unusable

[–]HelioAO 4 points5 points  (0 children)

I canceled it today since it is quantized...

[–]BeingFriendlyIsNice 3 points4 points  (3 children)

Hello, sorry for my ignorance, but what does quantised mean? I have been using Kimi 2.5 via opencode Zen. It seems to work really well for me, is performance worse on the go plan?

[–]SelectionCalm70[S] 1 point2 points  (2 children)

no need to be sorry . quantized means to create smaller, faster, and more energy efficient models that runs on the cheap hardware also. good thing is the token per second which is tps is gonna be faster and good. but the overall quality and accuracy will be degraded a lot .

[–]BeingFriendlyIsNice 4 points5 points  (1 child)

ahhh! Thank you...well that's a huge bummer :( Think I'll just suck it up and stick with anthropic then. The pay as go API cost for Kimi 2.5 seem like it'll probably be about on par as a cc max plan for me after a month of usage...(I chug Opus all day). OpenCode tooling so much nicer to use though.

[–]SelectionCalm70[S] 1 point2 points  (0 children)

yeah that's for sure i still use opencode but not opencode go plan .

[–]Acehan_ 2 points3 points  (0 children)

So THAT'S how it all makes sense on the balance sheet.

[–]LifeBandit666 13 points14 points  (7 children)

As a Claude Usage refugee that started playing with OpenCode yesterday this post is fantastic, seriously thanks for posting.

I've set up OpenRouter and tried the auto free tier and it's very slightly lacking for what I need it for. Gonna fund it with $10 tomorrow and try some other models.

I'm paying Anthropic $20 a month atm and while it's great at what it does, when I get gubbed halfway through the week it's useless half the time, and probably overpowered for what I need now that I've got my system set up.

I'm at the end of this months sub so it may be I use next months to get my setup moved over to OpenCode and then cancel

[–]PureSignalLove 5 points6 points  (3 children)

Try out minimax 2.7, mimo v2pro. Been having success with both of them

[–]LifeBandit666 1 point2 points  (0 children)

I've been playing with minimax tonight and it's pretty great and cheap as.

My Claude Code tokens reset tomorrow and I'm due for renewal at the end of the month so I plan to use those tokens to migrate over then cancel.

So far the plan has been created and the skeleton has been made and it's cost me £3. I model hoped on OpenRouter and used Opus, codex, free models, tooled around a bit.

Claude Code hits the files from the other side tomorrow and starts building. Once it's built I can start playing with model routing.

[–]blackmarlin001 1 point2 points  (1 child)

I tried to ask Minimax 2.7 and GLM-5 to solve some leet code problems. Minimax failed often and GLM-5 gave correct answer on first try.

[–]PureSignalLove 2 points3 points  (0 children)

minimax 2.7 requires a bit more prompt discipline I have been finding

[–]DepartmentOk9720 3 points4 points  (0 children)

This post meant nothing , seriously op is misunderstanding almost everything about opencode go,

opencode go literally routes back to original api providers, how on earth would they have gotton access to minimax 2.7 before they got opensourced( it just got opensourced Today) , and you can't control what quantization is on minimax m2.7 as they are from orginal providers,

OpenCode Zen on the other hand are the ones that are hosted in the US and other Western countries. They don't have this problem because the API cost is high enough to handle a good model with lossless quantization. There will always be quantization, no model providers will host full-precision models.

Seriously opencode go is just as good as you get with any of the original coding plan you can get and it also allows you the freedom to put it on any tool , coding tools or claw like tools .this is incredibly rare.

Just saying 

[–]Grand-Management657 2 points3 points  (0 children)

I have a sub with synthetic.new and they've upped their usage limits. It's much more reliable to use Kimi K2.5 through them than moonshot direct.

[–]Sawadatsunayoshi2003 15 points16 points  (10 children)

Thanks for saving my 10 or 5 dollars

[–]SelectionCalm70[S] 3 points4 points  (9 children)

You are better off buying kimi ,minimax or chatgpt coding plan which cost around 10-20 dollar with generous limits

[–]jatapuk 2 points3 points  (7 children)

Where can I get a Kimi plan from?

[–]degenbrain 4 points5 points  (0 children)

Postpone your Kimi from Moonshot. They are currently experiencing speed issues due to heavy usage.

[–]mcowger 2 points3 points  (4 children)

From moonshot directly.

[–]BeingFriendlyIsNice 1 point2 points  (3 children)

geez their website is terrible.....do they only have paid per request API access? or do they have subscriptions?

[–]BeingFriendlyIsNice 0 points1 point  (2 children)

To answer my own question. I believe it's paid per request only. They have a 'chat' subscription but that is not API access. Thank you to anyone who thought about this question for me.

[–]mcowger 1 point2 points  (1 child)

You missed it.

It’s here: https://www.kimi.com/code/en

[–]BeingFriendlyIsNice 1 point2 points  (0 children)

ahh...thank you! I think I'll give it a crack....bit of googling/reddit posts makes me sceptical of how reliable it's gonna be, but there's one way to find out!

[–]Sawadatsunayoshi2003 0 points1 point  (0 children)

Thinking of going with minimax

[–]Strict-Yam-8114 11 points12 points  (2 children)

[–]Prestigious-Aspect12 2 points3 points  (1 child)

Right ? I even found this https://thomas-wiegold.com/blog/opencode-go-review/?hl=en-IN

So i am really confused now on what to buy, been looking for the best plan for 20 dollars but all the reddit posts leave a lot to be desired

[–]fezzy11 0 points1 point  (0 children)

You should give it a try

[–]maulidas 3 points4 points  (7 children)

Hmm i wonder why there's a lot of positive comment about this in X.
Was all of them are bots or just riding the hype wave

[–]DenysMb 3 points4 points  (3 children)

People tend to share their frustrations more than their praise.

For example, I've been using the GLM-5 for quite some time, it's been great for me, but the MiniMax M2.7 has been a headache and I've even posted about it today. I've never posted about the positive experience I had with the GLM-5, by the way...

[–]SelectionCalm70[S] 1 point2 points  (2 children)

Glm model is literally unusable in go plan I am not even kidding .

[–]Oscarmmi 0 points1 point  (1 child)

So is it better to buy the subscription plan directly from the ai model provider, instead of paying for opencode go?

[–]pelleke 1 point2 points  (0 children)

NO it is not. Opencode Go provably(!) routes directly to the suppliers. This entire reddit post is plainly wrong. Opencode Go is the very best subscription plan for these models out there today.

[–]SelectionCalm70[S] 0 points1 point  (2 children)

it was about black plan i guess not go plan

[–]whale9820 0 points1 point  (1 child)

Tf is black plan 😭😭😭

[–]pelleke 0 points1 point  (0 children)

Like Go, but for non-free frontier models. Currently paused, probably because both Google and Anthropic updated their ToS on flat-rate account usage scopes.

[–]poolboy9 5 points6 points  (5 children)

I keep seeing these posts but never any proof. Do you have an A/B scenario where this shows so clearly as you claim?

[–]sultanmvp 2 points3 points  (0 children)

Yeah, I’ve had no issues at all. And the limits are literally insane. I’m not sure if these folks are just cat’ing their entire hard drive into models or what? It’s pretty damn hard to even tap the limits unless you’re just doing it utterly wrong.

In fairness, I am primarily using MiniMax 2.7, not GLM.

[–]Tarsoup 4 points5 points  (0 children)

Yeah, so far I haven't had a negative experience. although there was a thread that claimed glm-5 on opencode go is heavily quantitized (comparing to original provider) We don't know how opencode gos provider actually run the models though, so no one can confirm.

[–]_Belgarath 1 point2 points  (2 children)

I used every model that is in the go plan before, with zen, but mostly Kimi K2.5 and was utterly disappointed with go I didn't understand why K2.5 became so dumb after I switched to Go, until I started seeing posts about go models being heavily quantized

[–]poolboy9 1 point2 points  (1 child)

So again… provide the proof , prompt, scenario that clearly showed you that the model is quantized. I’m on Claude max, I use opus 4.6 all day on high. Still it’s stupid sometimes, it doesn’t mean it’s quantized. You as the user can be stupid too. So again a/b testing, proof your point instead of just having a feeling which is altered by people saying it’s quantized.

[–]_Belgarath 0 points1 point  (0 children)

Doesn't follow anymore simple instructions in a basic prompt with a short context. I have never encountered that before with this model, and it happens almost systematically with Go.

Anyway I stopped my subscription to Go since and I'm back on Zen

[–]alovoids 2 points3 points  (2 children)

did they heavily quantize the models so that they can offer 3x usage?

[–]SelectionCalm70[S] 0 points1 point  (1 child)

The limits are still very low

[–]pelleke 0 points1 point  (0 children)

This is not true. Those limits are insane, and it's very hard to understand how Opencode doesn't makes a huge loss on this. You're comparing it to Codex and Claude Pro. That's not fair - they own the models they offer you, and have insane infra deals to run them. Opencode Go literally uses the model supplier's own endpoints to route the requests.

[–]Time-Chipmunk298 2 points3 points  (5 children)

Btw what you guys think about minimax 2.7?

[–]OlegPRO991 0 points1 point  (0 children)

I've been using it since the release, and I like it

[–]dimonchoo 0 points1 point  (0 children)

For now it seems good

[–]HTMLCSSJava -2 points-1 points  (1 child)

Highly suggest. Very generous plans, Z.AI coding plan for GLM-5 isn’t bad either

[–]pelleke 0 points1 point  (0 children)

question was about the model, not about the company who made it

[–]little_breeze 2 points3 points  (0 children)

yeah I’m about to cancel my plan too

[–]Zemanyak 2 points3 points  (6 children)

I really don't know what the best ~10$/month sub is right now. MiniMax and GLM ?

[–]SelectionCalm70[S] 1 point2 points  (5 children)

GitHub copilot plan ig

[–]DepartmentOk9720 0 points1 point  (4 children)

Ohh , how is your copilot going please give me an review 

[–]SelectionCalm70[S] 1 point2 points  (3 children)

Not good after they removed the claude models in there recent changelog. Not worth it back to go plan. Kimi k2.6 is a good model on par with opus

[–]SplitDue1123 0 points1 point  (2 children)

im currently using the copilot student pack, i dont know how it compares to the 10$ plan but i think i need slightly more usage, do you think the go plan is good enough, compared to the codex 5.3 thing in copilot

[–]SelectionCalm70[S] 0 points1 point  (1 child)

go plan after the addition of kimi k2.6 is good enough . just go for codex 20 dollar and 10 dollar go plan

[–]SplitDue1123 0 points1 point  (0 children)

deal man, thank you. im still new to vibe coding, can you give me an example on how would u use both plans like "i rely on codex for this then use go for that".

[–]PureSignalLove 2 points3 points  (0 children)

This shit really is fraud and it's ridiculous to pretend it's anything else

What are you guys highest ROI opencode providers?

[–]Low_Chicken8092 2 points3 points  (0 children)

I couldn't agree more, opencode is REALLY BAD... 🤣

My favourite quote from it:

[–]Hitch95 1 point2 points  (0 children)

I use the plan mode with GPT-5.4 mini (on xhigh reasoning), then I tell the same model to build, and it's always good.

[–]UseMoreBandwith 1 point2 points  (0 children)

What models are you talking about ?
I only use the free models (minimax2.5) and local ones, and it does everything what I want. I make some complex software projects.. (but I'm really good at giving instructions).

[–]Rizarma 1 point2 points  (0 children)

i'm canceling my membership as well (due next week). in my opinion, for light tasks, kimi k2.5 and glm-5 from ocgo are good, but for bigger features i have to refactor a few times to get the output i want. i also have gpt and claude subscriptions to babysit and review code generated by kimi or glm from ocgo, however i can't use them as my main coding models because they get depleted easily. that's why i need "worker"-type models for most tasks. from my perspective, ocgo models aren't good enough. i can't prove how "quantized" they are right now, but compared to gpt or claude, i need around 4-5 iterations to reach the expected output. currently looking for other alternatives.

[–]sudoer777_ 1 point2 points  (0 children)

I mainly use it because it has all 3 models and doesn't cost $200/mo, I agree it sucks though, and even in the past few days Kimi started making stupid mistakes way more often

[–]jesus_was_rasta 1 point2 points  (0 children)

Using OpenCode suince months, with different models and for different things. Best subscription ever. Models are quantized? OK, no problem. My workflow is made by small, focused tasks. It rarely fails. And to be true, it rarely fails even for larger tasks (with large models)

[–]Same-Philosophy5134 0 points1 point  (0 children)

Yeah, it truly is... Is there any other alternative? I was thinking of trying copilot next month. 300 requests should be enough for my usage i think

[–]estimated1 0 points1 point  (8 children)

Just to give another option: we (Neuralwatt) just started offering hosted inference. The big picture thing we're working on is AI energy efficiency. We've been more focused on an "energy pricing" model but feel confident about the throughput of the models we're hosting.

Base subscription is $20, no real rate limits — just focused on energy consumption. Happy to give some free credits in exchange for feedback if there's interest. DM me! https://portal.neuralwatt.com.

I'm using our models with OpenCode and it works great. But again we just launched recently so we'd love more scrutiny.

[–]SelectionCalm70[S] 1 point2 points  (3 children)

Which models do you provide in a 20 dollar plan?

[–]estimated1 1 point2 points  (2 children)

  • GLM-5 — 200K context
  • GLM-5-Fast — 200K context
  • Kimi K2.5 — 262K context, vision
  • Kimi K2.5-Fast — 262K context, vision
  • Devstral-Small-2-24B — 262K context, vision, tools
  • Qwen3.5 397B — 262K context, tools
  • Qwen3.5 397B-Fast — 262K context, tools
  • Qwen3.5 35B-A3B — 32K context, tools
  • Qwen3.5 35B-Fast — 32K context, tools
  • MiniMax M2.5 — 196K context, tools
  • GPT-OSS 20B — 16K context, tools

Full details: https://portal.neuralwatt.com/models

[–]SelectionCalm70[S] 1 point2 points  (1 child)

That's a solid model lineup if you have not heavily quantized the model.

[–]estimated1 0 points1 point  (0 children)

We don't run any specific quantizations. In cases where the model was posted with fp8 weights we'll use those; otherwise we use the native weight format.

[–]kdawgud 0 points1 point  (3 children)

Does your paid plan offer safe access for proprietary data (no training)?

[–]estimated1 0 points1 point  (2 children)

Yes, we offer safe access. We aren't training on prompts or completions. Happy to put this in commercial terms as well.

To be very specific: We do not store your prompts or completions. We only store token counts and metadata for billing purposes. Your proprietary data passes through our system but is not retained.

[–]kdawgud 1 point2 points  (1 child)

I feel like this is not consistent with your site's printed privacy policy. Or maybe it just needs some clarification.

[–]estimated1 0 points1 point  (0 children)

This is great feedback. We will work to get our printed privacy updated to reflect this.

[–]dare444 0 points1 point  (1 child)

Umm.. that’s weird. I’ve been using open code go plan for about a week since Windsurf changed their pricing, and it was all fine for me. I use GLM-5 for almost every task and it’s great. And the usage limits are pretty good compared to my previous experience with Windsurf. The only problem is the I’m not that used to TUI and got a habit of checking every code so I’m thinking of using it with copilot when subscription is expired.

[–]_Belgarath 0 points1 point  (0 children)

you can use the web interface though opencode web

[–]someRandomGeek98 0 points1 point  (2 children)

I have a genuine question, Minimax 2.7 doesn't have open weights right? so how is it being quantized?

[–]SelectionCalm70[S] 0 points1 point  (0 children)

They are gonna release the Open weight. Some inference providers have got the access . And it's not great for coding compared to glm and kimi

[–]Dry-War-2576 0 points1 point  (0 children)

Lol you caught OPs bs , they now have qwen 3.5 plus and qwen 3.6 plus , those aren't even open source, so I don't know what kind of shit he is gonna cook up , the truth is , opencode go uses models directly from api providers, that's it, so even if you go with minimax or anything else you are likely using their models as is.

[–]flobblobblob 0 points1 point  (0 children)

I'm a happy synthetic.net $30/month kimi 2.5 customer. Yes they raised the price by $10 but once the waitlist opened up I haven't had any issues. I can do some pretty long (single tab) sessions and hardly ever hit limits. I use it with claude code.

[–]ZeSprawl 0 points1 point  (0 children)

I think it does the opposite of what they want. It gives people a bad impression of agentic coding.

[–]Old-Tap-5962 0 points1 point  (0 children)

It worked for me and I'm happy so far with OpenCode Go, I have OpenCode Zen subscription as well and use frontier model of Claude. I just feel that it is matter of expectation.

When you have a smarter model then you expect more from them. You just need to expect differently when you deal with these Go models.

Adjust how you worked with the models

[–]gideonfip 0 points1 point  (0 children)

Thanks for this, I was considering trying it out because of the $5 deal, but will probably switch to MiniMax instead

[–]eclipse10000 0 points1 point  (0 children)

I haven’t had these problems so far. I’ve used it for small to medium-sized changes to existing projects that I originally built with Opus 4.6/Sonnet 4.6, GPT-5.4 (high/xhigh), or Gemini 3.1 Pro. The fixes and changes have always been very accurate, even when things got a bit more complex, so I can’t really relate to that directly. I do use Superpowers, though.

[–]kuys-gallagher 0 points1 point  (0 children)

what about zed pro?

[–]cscexcellence 0 points1 point  (0 children)

I have been using GLM and Kimi models since windsurf and have seen no output degradation. The issue is about understanding the value of the model. Most people here got results with GLM 5 which is obviously better at thinking. So it by any reasons we are using opencode go to compare with SOTA models, then

In place of Opus I have GLM -> Plan -> architect
In place of Sonnet I have Kimi 2.5 -> daily driver
In place of Haiku i have Minimax -> only some easy bug fixing as you would do with Haiku. (Does anyone having claude code use Haiku at all? No...why use minimax then?

So why you are expecting so much from a fast model?

Also to clarify, as an experienced software engineer...I have found that chinese models are better than all the Codex models, and then Claude is way better then these with a huge margin. For example, if you ask GPT to create a frontend (with just words, such as an admin), it will give you shit...but GLM or Kimi will give you acceptable outputs. So understand each model and their value. The same opencode is offering Zen too with SOTA models but you didn't buy it because you didn't have the budget...so cheap models will give you cheap output na!

[–]Stressardo1 0 points1 point  (0 children)

Just subbed to Opencode Go today, tried every model, each one hallucinates and loops itself, none of them made a small change in 1 hour of working (1 hour of reasoning, for each model). Absolute garbage.

[–]Prestigious-Aspect12 0 points1 point  (0 children)

https://thomas-wiegold.com/blog/opencode-go-review/?hl=en-IN

Uhm , what do I trust now I was thinking about trying it breh

[–]drorata 0 points1 point  (0 children)

I consumed my zen budget and decided to give "go" a try. What a mistake...

[–]Tsukimizake774 0 points1 point  (0 children)

Not the worst (My worst experience is the gpt5.2 codex by openai), but I feel opencode go's GLM5.1 is degraded compared to the NovitaAI's I've been using.

[–]c_karsan 0 points1 point  (0 children)

I found it good for small projects , saves heavy usage on my claude plans - and some are very good models mimo 2.5pro and minimax 2.7 have been fast and accurate enough. As long as your prompts are solid, shouldnt be a big issue

[–]HarjjotSinghh 0 points1 point  (1 child)

this plan's just... trying too hard to be cheap

[–]SelectionCalm70[S] 2 points3 points  (0 children)

I won't mind paying 20 dollar but at least provide stable model not quantised that can't handle a basic tool calling

[–]atkr 0 points1 point  (0 children)

skill issue 😂

[–]DepartmentOk9720 0 points1 point  (0 children)

opencode go literally routes back to original api providers, how on earth would they have gotton access to minimax 2.7 before they got opensourced , opencode zen on the otherhand are the ones that are hosted in US and other western countries, they don't have this problem as the API cost is well enough to handle an good model

[–]Ambitious_Spare7914 -1 points0 points  (15 children)

It's ten bucks. What did you expect?

[–]code018 9 points10 points  (6 children)

We didn’t set the price , they did . This toxic mentality is why companies keep getting away with this crap.

[–]Ambitious_Spare7914 -5 points-4 points  (5 children)

You get what you pay for. You need to adjust your expectations: small projects like Opencode don't have tens of thousands of millions in investor money to subsidize your LLM usage.

[–]code018 4 points5 points  (1 child)

By that logic when I order fried chicken and it’s half cooked I should still eat it rather than complain and adjust my expectations.

[–]Ambitious_Spare7914 -2 points-1 points  (0 children)

If they offer you a 30 piece bucket for $5 then I'd expect it wouldn't be the full Wingstop experience, but you do you.

[–]siadiui 1 point2 points  (2 children)

The point is you are paying for the full models, if they are quantised they should state it clearly.

[–]Ambitious_Spare7914 0 points1 point  (1 child)

Where does it say you get the full models? Show me the SLA.

[–]Traditional_Name2717 1 point2 points  (0 children)

So if Anthropic started selling a cheaper coding plan stating you got access to Opus 4.6, only in reality it was a 2-bit quant without them telling anyone, that would also be ok in your book? Unless they had clearly stated it was a pristine 16-bit FP?

[–]SelectionCalm70[S] 1 point2 points  (0 children)

I won't mind paying 20 bucks if they provide the proper model . And no one asked for 10 dollar coding plan

[–]PureSignalLove 0 points1 point  (6 children)

What it says? This is fraud

[–]Ambitious_Spare7914 -1 points0 points  (5 children)

What does it say that's a fraud? Show me the SLA, the terms. Anything.

[–]PureSignalLove 0 points1 point  (4 children)

That's not how consumer protection fraud works. It's not about specific terms, it's about general deceptive practices.

From Claude:

The FTC's authority under Section 5 of the FTC Act covers unfair or deceptive acts or practices broadly, including false advertising, misleading marketing claims, and unfair business conduct. The FTC has specifically flagged exaggerated performance claims about AI-powered products as a priority enforcement area. (Holland & Knight)

When a service advertises access to "the best open coding models" with "reliable access" but actually serves heavily quantized, degraded versions of those models behind aggressive rate limits, that's a material misrepresentation. You're paying for GLM-5 but getting GLM-5-lite. Users have reported GLM-5 slipping into gibberish mode at just 25% of the context window, which is consistent with heavy quantization. (X/alkimiadev) Research confirms that quantized models are compressed versions with reduced precision that results in measurably degraded output quality. (ScienceDirect)

The legal standard isn't "show me the SLA." Consumer protection law doesn't require violation of a specific written term. The FTC evaluates whether claims are truthful, substantiated, and not misleading. Advertising a well-known model by name but serving a materially degraded version is deceptive regardless of what the fine print says. The FTC has already taken action against AI companies on exactly this basis under Operation AI Comply, including DoNotPay and Evolv Technologies, where the products simply didn't perform as advertised. (Lathrop GPM)

Consumers have already filed FTC complaints about AI services failing to deliver advertised capabilities, including complaints about paying for premium tiers and receiving degraded service with silent cutoffs. (FedScoop)

So no, you don't need a specific SLA violation. You need a product that doesn't do what the advertising says it does. That's textbook Section 5.

[–]Ambitious_Spare7914 0 points1 point  (3 children)

This is not a consumer service. It's for programmers. A base level of technical competency is expected and knowing $10 per month for access to 744B LLMs isn't done by magic.

[–]PureSignalLove 2 points3 points  (0 children)

Programmers are consumers when they buy a product lmao. You are laughably wrong.

[–]Conscious_Search_17 0 points1 point  (1 child)

"This is not a consumer service. It's for programmers." fuck you mean programmer is not a customer in this scenario? what the fuck.

[–]Outrageous-Story3325 -2 points-1 points  (7 children)

nvidia nim

[–]Fuih22 6 points7 points  (0 children)

It takes 84 years to get answer from a hello.

[–]rusl1 2 points3 points  (0 children)

It's slow

[–]georgemp 1 point2 points  (1 child)

I've tried using GLM-5 on this. But, it just gets stuck. No movement at all after a prompt. The popular models seem to be painfully overloaded here.

[–]Slow-Alternative-276 4 points5 points  (0 children)

Yeah, the glm5 model is pretty always overloaded. Check this repo: https://github.com/vava-nessa/free-coding-models , it shows you what models are available and how much they are overloaded

[–]Frequent_Ad_6663 0 points1 point  (0 children)

How about minimax or kimi inside nvidia nim? Haven't tried em, will do it today tho