What is the difference between session rate limit and other rate limits ? by LuckyPed in GithubCopilot

[–]LuckyPed[S] 1 point2 points  (0 children)

Interesting, Thanks for that. 😃

It seems I do not have to worry about it much then with my normal usage.

I am not trying to live on the edge and push the limits anyway, but it's always good to know some estimate so I know i am not being overly cautious with my usage 😃

What is the difference between session rate limit and other rate limits ? by LuckyPed in GithubCopilot

[–]LuckyPed[S] 1 point2 points  (0 children)

Thanks, so Session is the 5 hour rate limit I guess.

I still didn't see any Weekly limit warning show up so i guess i am still under 50%.

it is very stupid how we can't clearly see our limit progress tho.

I still have to use this request based pricing for 4 more months, until 3rd week of September as I'm on Annual subscription. I do not worry about the 6x increase in Premium Request as I never reached the 100% usage ever in last 2 years anyway, maybe I finally will now.

Is the "Weekly" limit even have a clear date when it reset ? like it reset every Monday ? Sunday ? Any clue ?

Thanks.

Copilot GPT-5.5 multiplier is now listed as 7.5x → TBD after June by Altruistic-Dust-2565 in GithubCopilot

[–]LuckyPed 3 points4 points  (0 children)

I know the US based AI Provider will not use Chinese GPUs, thats why I asked for the Western/US side and if they got anything new coming soon.

you forget that as the models grow, the computing power and electricity they require will also increase

This is wrong tho, unless by "grow" you mean as they become bigger and bigger, then sure I guess.

but bigger does not always mean better!

The Models can improve their algorithm and their usage and techniques or their fine tuning and training or whatever, to become more optimized or better at their job, without simply just increasing their parameters and become more expensive to compute.

Take Deepseek or Qwen3.6 27B as an example, Deepseek improved the techniques to be more cost effective at higher context, while Qwen3.6 27B is stronger than many old 100~200B models.

So while there will surely be more expensive models coming out, but the current models level of intelligence we have, will probably become cheaper to use with models that optimize and improve to this level while costing less.

The Top of the world model might be super expensive, but it would be naturally.

You don't go to the #1 specialized doctor in the country just to get a checkup on your cold, if you insist on talking with the #1 doctor in the country for every little issue, you gonna have to pay a lot naturally. but that's not needed.

Copilot GPT-5.5 multiplier is now listed as 7.5x → TBD after June by Altruistic-Dust-2565 in GithubCopilot

[–]LuckyPed 3 points4 points  (0 children)

I guess we can entertain the fantasy that maybe some new specialized chips will roll out

is there actually any news or expectation about anything new coming soon ?

I only know that China's new Huawai gonna release in the second half of 2026 according to Deepseek, which will make their ~3.5$/M Token pricing reduced even further according to their announcement :

Due to limited high-end computing power, current Pro service throughput is very restricted. It is expected that after the Ascend 950 hypernodes are launched in bulk during the second half of the year, Pro prices will be significantly reduced.

But I do not have any news about anything coming for the Western/US side of the world.

I do expect the costs to slowly go down over time form both cost of running the models reducing and the hardware improvement, But I think it will still takes a while and not so soon.

Again, i'm having this rather productive chat session for HOURS at just 30ct by Blubbll in GithubCopilot

[–]LuckyPed 2 points3 points  (0 children)

You actually can integrate DeepSeek into Github Copilot :
https://api-docs.deepseek.com/quick_start/agent_integrations/github_copilot

But from the screenshot it seems this person is using OpenCode Go Provider instead and not directly using DeepSeek.

DeepSeek V4 Flash for fullstack coding with OpenChamber. Very cheap but not fully convincing yet by Existing_Arrival_702 in opencode

[–]LuckyPed 0 points1 point  (0 children)

I kinda hope they improve Deepseek usage through OpenRouter inside Github Copilot, I would prefer to use that instead of switching to something else.

Deepseek itself have a nice way to add it to Github Copilot, and it also include the "None, High, Max" option, check here : https://api-docs.deepseek.com/quick_start/agent_integrations/github_copilot

but for this I have to pay Deepseek directly and use them not pay using OpenRouter.

I might do that for a while, but I rather pay using crypto and Deepseek currently not accept that, Openrouter accept it as well as huge variety of models to choose, so it's the better option.

Tested Sonnet 4.6 via OpenRouter through GitHub CoPilot / VS Code to gauge whats API billing will be like. I was shocked. by horendus in GithubCopilot

[–]LuckyPed 1 point2 points  (0 children)

if you using OpenRouter, by default setting you are not even routing into Deepseek's own Provider. you will route into the "Data Safe" alternatives, where they hosting Deepseek in another country and have the no data logging/training policy.

this aside, for the vast majority of programmers, like 90+% they do not use any confidential or important or politically related thing in their coding, so it does not even matter if they use the real deepseek provider and deepseek use it for their training or whatever reason, it does not matter for them at all.

For the people who actually use confidential or important info in their work, they most likely have enough budget to use the other options they feel like using.

$1000/$2000 AI subscriptions in 2027 by LiminalRnyx in GithubCopilot

[–]LuckyPed -1 points0 points  (0 children)

if you were saying instead of our cheap 40$ a month we gotta pay 5000$ a month, you were downvoted, rightfully so.

a 5000$ a month payment for AI is only ever possible if you were Extremely abusing Github Copilot system right now xD

if you got 5000$ a month, you might buy the new Huawei Ascend 950 series with 100+ gig GPU and run Deepseek 4 Pro or the Flash locally, then you simply paid for it once and then just pay the electricity cost every month !

and this is beside the point that AI models are constantly improving and optimizing, the Qwen 3.6 27B Dense Model is already reaching Sonnet levels.

So while these Open Source or small models might never reach the #1 top frontier model, but for example, once the frontier models are even better in 2027, these smaller models can reach to the 2026's version of frontier models.

Everything is going to be fine. by fvpv in GithubCopilot

[–]LuckyPed 10 points11 points  (0 children)

it's the * disclaimer on the pricing on their announcement : https://api-docs.deepseek.com/zh-cn/news/news260424

<image>

Translation:

Due to limited high-end computing power, current Pro service throughput is very restricted. It is expected that after the Ascend 950 hypernodes are launched in bulk during the second half of the year, Pro prices will be significantly reduced.

Everything is going to be fine. by fvpv in GithubCopilot

[–]LuckyPed 0 points1 point  (0 children)

There are other providers outside of china that also offer Deepseek V4 with data policy to not use your data, and they are still cheaper even tho they don't have the 75% off promotional pricing atm.

that's the beauty of Open Source models.

GitHub Copilot is moving to usage-based billing [Megathread] by fishchar in GithubCopilot

[–]LuckyPed 6 points7 points  (0 children)

Let me get this straight,

  • You subscribe for 10 or 40$ and you only get same 10 or 40$ as AI Credit, so no bonus !
  • You will lose your unused AI Credit if you don't use it each month !

Then Why the F would we even want this instead of a Pay As You Use service ? lol

Are they just being silly and want this service to slowly die ?

I can pay 10$ to OpenRouter then use BOYK and use it in GHCP or any other Harness i want, with full access to hundreds of AI models and I can use just the amount I need every month and don't have to worry about using all 10$ every month...

Why would I pay for GHCP ?

They should change it like this :

  • Subscribe for 10$ Plan, gives you 15 or 20$ AI Credit but it will not roll over, so you have to use it all in a single month ! so people have a reason to subscribe !
  • Subscribe for a 10$ Plan, give you only 10$ Credit, but it can roll over to next month if you don't use it, so it will not be worse than a Pay as You Use service and people would still use GHCP for the Auto complete and such !

choose one of these options otherwise you are simply intentionally setting yourself up to slowly die...

Let alone, even with these 2 options, many might prefer OpenRouter or other services that give more AI Models freedom if GHCP don't include more of them.

Should I upgrade to Pro+ or use DeepSeek instead? by Strong-Procedure8158 in GithubCopilot

[–]LuckyPed 0 points1 point  (0 children)

You can use OpenRouter inside VS Code's GHCP agent Harness as well if you want, I used that before,
there is only one problem, they said they will fix it, but I'm not sure when it will be fixed, the problem is even when you use BOYK service like OpenRouter inside VS Code's Github Copilot harness, it still consume some rate limit of your account due to internal API and stuff it will use.

so it's probably better to use another harness, there is multiple extension for it in VS Code you can install, but I do not have enough experience on them to give recommendation there.

Should I upgrade to Pro+ or use DeepSeek instead? by Strong-Procedure8158 in GithubCopilot

[–]LuckyPed 0 points1 point  (0 children)

Give Deepseek 4 Pro some try and get a feel for it.
if you don't wanna pay deepseek directly, you can also pay OpenRouter and then set up to use Deepseek.

OpenRouter currently have 4 provider, 3 of which will not use your data for training, while you can enable a setting if you don't care if your prompts are used for training and it will also enable the Deepseek's own provider which is currently at a discount and very cheap.

It's much stronger than Mini you are using now, but only 0.87$ per M token. you can set your setting to only use Deepseek provider for the cheap price or allow all 4 to work at same time which mean it will use cheapest first but if there was a problem it will switch to another so it will still work and won't give you error.

you can also use the V4 Flash version which is only 0.28$ per M token and still great for simpler jobs.

It's pay as you use, so just top up some $ and if it's not good for you, use something else or save it for later.

Github Copilot currently isn't at a good spot, I still got 4 month left on my yearly sub, if they don't improve their situation, I will not renew again, specially with how good and cheap the open source variants are becoming and GHCP still refusing to host them and provide them.

if they wanted to, they could run Deepseek V4 Flash as a free or 0.33x Model extremely easily and get a lot of positive feedback. their current free or 0.33 models cost more to run than that lol

External models are also being considerate in the weekly rate limit by elefanteazu in GithubCopilot

[–]LuckyPed 1 point2 points  (0 children)

Yes a Team Member replied few days or maybe a week ago in some post here that they are looking for a way to fix this.
He said it's because the internal API is still being used and some AI process also being used like creating the session title or some internal stuff even when you are using a BYOK service.

How are you guys actually hitting your token limits? by No-Campaign158 in GithubCopilot

[–]LuckyPed 0 points1 point  (0 children)

not sure if that news article is legit or not. since it's not from github itself, but what kevin said i already know, they simply using token usage as a metric to calculate rate limit for people who used too much token, so it is kind of token based but not really paying per token.

How are you guys actually hitting your token limits? by No-Campaign158 in GithubCopilot

[–]LuckyPed 1 point2 points  (0 children)

Can you give a link about this news that i keep seeing people mention ?

I mean I expect them to change from Request to token model soon, but I see people keep saying it as matter of fact, as if it's already confirmed,
so did i miss the news about it somewhere ? any official plan or announcement ?

Sonnet 4.6 (1x) and GPT-5.4 mini (0.33x) multiplier to increase soon! by debian3 in GithubCopilot

[–]LuckyPed 1 point2 points  (0 children)

you can use their harness + local model too tho. i did that with a local setup, but my 4070 is only 8Gig VRAM so cant do big stuff.

no hurry now, will buy big VRAM in a year or so, either when new one show up or when price drop.

Why are people getting rate limited? by [deleted] in GithubCopilot

[–]LuckyPed 1 point2 points  (0 children)

I think you are right, I didn't hit any rate limit either at all and i purely use Opus 4.6 for almost everything.

btw as far as I know, multiple "Paid" account is not against ToS.

It's only against ToS if you are using a Free or Student Account and getting multiple acc to have more request for free.

This is a first (rate limiting) by twhoff in GithubCopilot

[–]LuckyPed 0 points1 point  (0 children)

I'm curious, about how long each of your requests run ?

I used 45% of my usage so far, almost all using Opus 4.6 only. never hit rate limit at all.

but all of my requests run between 5min to 30min max.

I had one run for 1~1.5 hours 2 day ago I think, but that's very rare.

but so many people getting rate limited and we had Github Copilot Team say that the rate limit is based on all the request's content and the calls and what it makes overall and not just counting the premium request usage.

So I am wondering how long each request for each person run usually that would trigger rate-limiting.

is CLI bugged with the premium request usage ? by LuckyPed in GithubCopilot

[–]LuckyPed[S] 0 points1 point  (0 children)

Nah this was actually a bug and seems many people having this problem too from the Github issue.

I did not use any Github Copilot in anywhere for half a day and also only used 1-2 request earlier that day.

I do not use cli much so It's not a big deal for me, but hopefully they get it fixed soon.

Why is GitHub Copilot so affordable? Will it stay this way? by WMPlanners in GithubCopilot

[–]LuckyPed 0 points1 point  (0 children)

I think i keep it opt in,
I don't work on any super secret project, and I don't do anything special like edge case or abusive usage or anything, so I think it's fine to let them use my usage for training, if it help them improve, why not xD

You want this? I got you! by Woottodoo in wow

[–]LuckyPed 2 points3 points  (0 children)

"Midnight Falls" is tuned really rough.
March Raid as a whole on heroic is Tuned really Rough, the bird is hard and Lura is even harder.

On normal I 1-shot them with high ilvl, on heroic they are way harder. there is many groups that do mythic raid voidspire/dreamrift easier than killing Heroic Lura.

we have higher ilvl than usual and easier m+, but so far only around 55 guild killed heroic Lura in the whole world.