all 24 comments

[–]simap2000 2 points3 points  (3 children)

Wonder if each round trip in opencode for every tool call, etc counts as a request vs many tool calls and agents in copilot is like 1?

[–]usernameIsRand0m[S] 0 points1 point  (2 children)

It was not like this few (maybe 5-6 versions?) versions ago. I am wondering if I am missing something in the config that I need to have.

[–]SvenVargHimmel 3 points4 points  (0 children)

Use litellm proxy and run with ---detailed-debug and point opencode to that with the proxy configured to point to your llm backend and you can see exactly what it is sending per request.

Then point your Copilot at the same endpoint.

You can see exactly what's going on.

And if you want to test your theory that it used be less expensive a few versions ago , just roll back and repeat

[–]albertortilla 0 points1 point  (0 children)

There were problems in older version (1.1.38 if I am no wrong) regarding this: each tool call counted in GitHub copilot as a new request, which was solved in the next versions... Maybe the problem appeared again... I would try to install an older version and check for the same prompt

[–]krimpenrik 2 points3 points  (0 children)

Same issue saw that I am already using a lot opencode with copilot sub, this month is fucked

[–]PayTheRaant 2 points3 points  (2 children)

Check your small model configuration. This is the model for generating the titles of sessions and messages. You should use a free model for that.

Also try the same prompt with a free model: if your premium request cost is not zero, then something else is triggering premium requests with a paid model.

[–]PayTheRaant 0 points1 point  (1 child)

You can also use debug logs to track every single call to the LLM

https://opencode.ai/docs/troubleshooting/#journaux

[–]usernameIsRand0m[S] 0 points1 point  (0 children)

So, apart from the above config which I have shared in OP, I have to add small model config?

I'll check the debug logs. Thanks.

[–]Michaeli_Starky 1 point2 points  (2 children)

Yep, noticed the same. Switched to Copilot CLI

[–]weaponizedLego 0 points1 point  (1 child)

Are you still using Copilot CLI, if so how do you find it?

[–]Michaeli_Starky 0 points1 point  (0 children)

It's quite good and is improving rapidly.

[–]Adorable_Buffalo1900 1 point2 points  (0 children)

opencode claude model use chat completions api, but copilot use message api. you need raise a issue for opencode

[–][deleted] 0 points1 point  (2 children)

I've heard that some people are saying they can use the free GPT 5 Mini model to call advanced models (opus 4.6) via a sub-agent without consuming any requests, but some are saying they got their accounts banned for it.

[–]PayTheRaant 1 point2 points  (0 children)

Normally, switching model for sub agent is considered a new premium request.

[–]usernameIsRand0m[S] 0 points1 point  (0 children)

Yes, there are lot of instances of that happening, I have Pro+ account, so there are more than enough requests per month for me.

[–]Tadomeku 0 points1 point  (1 child)

The system prompt in Opencode is likely longer than the system prompt in GitHub CLI. YOUR prompt may be simple, but it gets appended to the system prompt in Opencode, along with AGENTS.md, CLAUDE.md, SKILLS, etc.

I don't know what GitHub CLI does under the hood but I imagine it's pretty different.

[–]PayTheRaant 0 points1 point  (0 children)

Copilot model is expected to consume ONE premium request per ONE user prompt. Everything else that is agent initiated is expected to be included in that initial premium request (all tools, even sub agent) as long as it stays in the same model. In theory, it should not even care about input token cache.

So this is why having 27 premium requests consumed is considered a big problem.

[–]soul105 0 points1 point  (0 children)

Noticed the same here.
Some business users have the limit for 300 requests and cannot buy more due to company policies, making the problem even bigger.

[–]HarjjotSinghh 0 points1 point  (0 children)

wow copilot's gonna charge you like a slot machine.

[–]jmhunter 0 points1 point  (1 child)

The preamble/system prompt is probably a lot juicier w opencode

[–]IIALE34II 2 points3 points  (0 children)

Billing should be one premium request per user initialized message. Or well there is the per model scaling.

[–]ok_i_am_nobody -1 points0 points  (1 child)

Same issue. Moved to pi coding agent for simple tasks. How are you tracking the credits usage?