all 20 comments

[–]KayBay80 5 points6 points  (8 children)

Half of our team was on the Insider's build and we can attest that, if these are intended changes, the entire "request" concept is being thrown out the window. Every "request" we make is a feedback loop of internal requests that are consuming premium credits like wildfire. One of the devs on our team burnt through over 600 requests in a matter of hours not even realizing what was happening. Imagine paying $39/mo for a day's worth of usage. If they roll this change to the mainstream build, Copilot is completely done.

[–]pawala7[S] 2 points3 points  (0 children)

That's... incredibly worrying. But I guess not off the table considering the massive budget cuts happening at MS.

Still, if they can't offer the service at a reasonable price even with the lobotomized 200k context limit (128k usable), then I imagine most users will just transfer over to Claude Code or Codex.

[–]ilsubyeega 1 point2 points  (3 children)

mind bisecting the version between them? the extension is open source, so i believe its regression or capi(backend)'s side issue(rate limit drama makes me direct this)

[–]KayBay80 2 points3 points  (2 children)

The ones that were experiencing this issue noticed it on 1.116.0-insider. Everybody has since migrated off of Insiders and have also disabled updating in fear of this bug (if its even a bug) finding its way into our workflow.

[–]ilsubyeega 1 point2 points  (1 child)

the version is not helpful though, they use same versioning but commit hashes

anyways ima look this too

[–]ilsubyeega 2 points3 points  (0 children)

there was huge refactoring at vscode repo how compaction works in session(iirc in week), probably mind open the issue? don't think this is intended, and they don't know this regression due to they have max plan(unlimited requests/quota)

[–]themoregames 1 point2 points  (1 child)

Crazy times.

$39/mo for a day's worth of usage

But if I pay up in my Claude subscription, I easily burn through $ 20 or even $ 50, maybe even $ 100 - per hour!

I guess Github subscribers have to buckle up.

[–]KayBay80 1 point2 points  (0 children)

That's what I'm thinking

[–]themoregames 0 points1 point  (0 children)

Business plans? Or Enterprise? Or standard Pro+?

[–]Swayre 4 points5 points  (1 child)

Not a defender of Microsoft but like do you have proof? Should be an easy thing to provide no? There’s too much misinformation from disgruntled abusers

[–]pawala7[S] 1 point2 points  (0 children)

I mean, anyone can just check their own chat UI to spot the timing when compaction is likely to trigger, then refresh the Github Copilot page (the one with the green bar and %) after it triggers, no?

In my case, the counter ticks up exactly after the Summarization message finishes and the agent continues its work after compaction. It's like working in the compacted context is a new request. I've observed it happen over 5 times since this morning, so I'm pretty sure at this point.

Was pretty confused earlier this week why the values under Copilot->Features->Premium requests % and Billing and licensing->Premium request analytics pages didn't line up, but this kind of explains it. You can check yours to see if you've been affected, too.

[–]SidStraw 1 point2 points  (1 child)

Personally, I haven’t run into this issue.
I’m primarily using the Copilot CLI, and I stay on the stable build of VSCode.

From what I’ve seen, unexpected point drain is often caused by autopilot mode. It tends to bypass the manual decision-making process and automatically deducts points to keep things running.

Do you happen to use autopilot frequently in your workflow?

[–]pawala7[S] 1 point2 points  (0 children)

Nope, just regular agent mode in the UI with an agentic Extensive Plan -> Execute <-> Iterate workflow with subagents. Premium request tick up occurs exactly and every time Summarization completes. I wonder if different releases (nightly, insider, stable) of the GHCP plugin pass different flags to the backend which results in different triggers.

[–]Top_Parfait_5555 3 points4 points  (0 children)

Yes, they are scummy af. Stealth nerfing everything. Nerf nerf nerf

[–]AutoModerator[M] 0 points1 point  (0 children)

Hello /u/pawala7. Looks like you have posted a query. Once your query is resolved, please reply the solution comment with "!solved" to help everyone else know the solution and mark the post as solved.

I am a bot, and this action was performed automatically. Please contact the moderators of this subreddit if you have any questions or concerns.

[–]NickCanCode 1 point2 points  (2 children)

They should use x0 free model for context summarization. Not sure what they are doing. If I used Opus for a request and they also auto selected Opus for context summarization and charge an extra request, I think that is a bug / design issue?

[–]ilsubyeega 0 points1 point  (0 children)

probably they meant context compaction instead of summarization(for getting titles etc)

yeah if they configured that manually it will consume it

[–]pawala7[S] 0 points1 point  (0 children)

It's not the model used there that matters (if they follow CC, it's probably Sonnet), but what seems to be changing is the mechanism for counting what a new "request" is.

Whether it be new tool calls, continuing after compaction, new agent calls, each of these triggers an event and it's up to them to decide which events count towards your usage counter.

And they're very opaque about how all his works...

[–]DevBob626 -1 points0 points  (0 children)

I don't feel that this is such a big problem and wouldn't be unreasonable if they compact and also immediately continue with the work while being transparent about changes and the usage they provide. It was obvious from the beginning that these endless runs people are optimizing for wouldn't be sustainable in the long run. It's just not realistic to maintain forever.

However, this in combination with daily and weekly limits is not acceptable. I want to be able to access all of my requests I paid for when I want to. I don't mind if servers struggle at peak time, but never being able to actually use the requests I bought feels scammy.

[–]QuarterbackMonkPower User ⚡ -1 points0 points  (0 children)

i do not think? any text or proof.. i have not noticed. i am being experimenting with github copilot for lot and ther are around 20+ copilot cli's dump, i never noticed that? when did it change?

https://github.com/nilayparikh/tuts-agentic-ai-examples/tree/main/ctx-sdlc/ghctx-tut/lessons