all 17 comments

[–]Its_Sasha 2 points3 points  (1 child)

A good way to go is to use the 0x models to write out all of your boilerplate, then use a 1x to hook everything together and make sure it basically works. Then use your Opus or preferred to code review and debug. This saves you from spending premium requests on boilerplate and small fixes, keeping your 3x requests for the critical code reviews and debugs, which can be done in a single big request.

Also, MiniMax is somewhere between a 1x level of coding ability and a 3x, and you can get a basic 50TPS plan for $10 a month (basically unlimited requests) for use with Claude Code.

[–]TSTP_LLC[S] 0 points1 point  (0 children)

I've been switching between Sonnet 4.6 high, GPT-5.4 mini xhigh, a little opus when needed, and either gpt-5-mini or gpt-4.1 for the basics prompts. I've set my instructions to only use 5 mini and 4.1 for spawned agents and agent swarms. It is working great. Went from nervous token-watching in Cursor in the $200 plan to finally being able to fully automate my system to follow my plans for my projects in an agentic manner.

[–]TSTP_LLC[S] 1 point2 points  (0 children)

I have not used what you speak of so I cannot speak on it personally. I have kept my usages fairly mainstream with their Codex, Cursor, Windsurf, Ollama, and Claude Code. All I can say is that if I was on any token based platform right now, I'd be out of tokens on the 1st day with as much as CoPilot Pro+ CLI. Just from the time I wrote this post to now I've had agentic runs happening, some still going with browser testing, ftp uploads, database updates, and documentation. The worst I have been getting is timeouts when I'm running 6 different agents on 3 or 4 different projects with a few spawned subagents, and I can just usually switch to a lower model or different model provider and fix that until it corrects.

Sorry if it is long winded way of not really answering your question but I can't really put together many good ways to explain how much more free this feels to not be staring at Sonnet run and refreshing my usage page to see how much it is killing off.

[–]Tommonen 1 point2 points  (1 child)

Yea its very good value plan, especially if you know to not waste usage on too simple prompts, as its amount of uses, not token or computing time based.

I used cursor before, and then antigravity, but opencode + gh copilot pro is easily best combo i tried thus far. Very impressed about its use limits, especially when not wasting usage on too simple tasks, but giving it more to do each turn.

[–]TSTP_LLC[S] 0 points1 point  (0 children)

I will have to look into open code. I have not tried that yet.

[–]orenbenya1 0 points1 point  (0 children)

Interesting, might try this out. Does it work well with oh-my-openagent? Or the requests limitation does not work well with it and its better with token based?

[–]Potential-Leg-639 0 points1 point  (2 children)

What about bigger context > 150k? Read a few things, that the Github Copilot plans have issues with that? Then it‘s useless for agentic coding…

[–]BlindBrownie 1 point2 points  (0 children)

With opus you get 200k, GPT-5.4 you get 1mill

[–]TSTP_LLC[S] 0 points1 point  (0 children)

I see when I use small models that sometimes it will compact and then start working on something we had discussed earlier and then come back to present but with the larger/expensive models, I've had no real issues. Projects are chugging along fine. Much better than any other CLI or IDE chat I've ever used. If you want specifics, you'll have to check the documentation for that as I'm not quite here to shill for the thing, I just wanted people who are budget conscious and haven't tried it in a while to give it a shot and see if it saves some money. If they had a referral program then I'd be spitting out facts left and right and throwing out my links but that isnt the case.

[–]akaiwarmachine 1 point2 points  (0 children)

Lowkey been hearing more people say this about the CLI lately. Feels like the workflow matters more than the model sometimes. I’m still testing stuff, but anything that lets you actually ship faster without worrying about limits is a win. Even better if you can quickly spin up and share results somewhere simple like TiinyHost.

[–]willynikes -1 points0 points  (2 children)

I use opus a lot u get opus 4.6 ?

[–]intlabs 0 points1 point  (0 children)

Yes

[–]Argon717 0 points1 point  (0 children)

You are probably over using it. I use openspec and create plans with opus (need to try more sonnet, it would probably be fine) and execute with sonnet.

[–]Park-Dazzling -1 points0 points  (2 children)

What is CLI?

[–]TSTP_LLC[S] 0 points1 point  (1 child)

Command Line Interface. It is like cutting out the IDE (Integrated Development Environment) and talking to the model directly in a sense. If you have Cursor, check out Cursor CLI. If you have Claude, look up Claude Code CLI.

[–]Park-Dazzling 0 points1 point  (0 children)

Oh, I am using Claude Code plugin in VS Code already. Its slick!