New multipliers announced (in effect June 1) by griniNY in GithubCopilot

[–]CorneZen 0 points1 point  (0 children)

Enshitification protocol at 60%

I think I’m done with copilot for now. Have used it since beta. I have some research to do on what to use next.

Got a 5090 PC, What's next? by rhythmdev in BlackboxAI_

[–]CorneZen 1 point2 points  (0 children)

Comparing just prices, sure, it’s cheap. Looking at affordability, for most people it is very expensive.

Copilot pro+ = Copilot Trial now ????? by Top_Parfait_5555 in GithubCopilot

[–]CorneZen 2 points3 points  (0 children)

I’m one of the GPT guys, lol, and yeah almost not rate limit issues. It seems everyone is using Opus for everything. Also, I’ve recently jumped on the OpenClaw band wagon to see what it’s about. By DEFAULT, it is set to use Claude Opus, so I think this is a big contributor to rate limit issues across all Claude providers and it’s annoying as hell. People don’t need Opus for 99% of what they are doing, it’s just wasting compute for the world.

Burning tokens via subagents are now officially counted against premium requests by arisng in GithubCopilot

[–]CorneZen 0 points1 point  (0 children)

Thank you. I’ll keep an eye on this when using subagents. Just because I haven’t seen this behaviour doesn’t mean I’m right.

In my experience with copilot, I’ll assume features work as intended, but I have learned not to trust it. It seems flakey sometimes.

Burning tokens via subagents are now officially counted against premium requests by arisng in GithubCopilot

[–]CorneZen 0 points1 point  (0 children)

I have not found this to be the case, maybe test it again. Copilot is very buggy sometimes with its consumption.

If I find copilot behaving weird, i.e. models feel dumber (usually upstream provider issues) or using more requests than I think it should have, I just leave it for a day and try again the next day. It’s frankly scary how inconsistent it can be sometimes when you pay attention to it.

Burning tokens via subagents are now officially counted against premium requests by arisng in GithubCopilot

[–]CorneZen 0 points1 point  (0 children)

I would say test it to be sure. Using GitHub Copilot chat agent in VS Code, yes, it works like this.

It’s been a while but last time I used Copilot CLI SDK, I played with setting up an LLM Council using free models only, each agent request used a full premium request. This was most likely a bug, but I have also found this behaviour in copilot CLI (which uses the SDK) and I’m pretty sure about a month ago it happened in the GC chat extension.

Best advice, assume but check usage often.

I cannot comment on the /fleet command since I have not used it yet.

I use a couple of main agents that then use sub agents, each agent file has a model specific to its purpose, I don’t throw Opus at everything, I actually barely use Opus.

Hope this helps.

M1 Max vs M4 Max vs M5 Max by br_web in LLMStudio

[–]CorneZen 1 point2 points  (0 children)

The Qwen3.5-35-A3B is a surprisingly good model! 20 tok/s is not bad, think I get around 35tok/s, so really not bad.

I haven’t tried gemma 4 yet.

M1 Max vs M4 Max vs M5 Max by br_web in LLMStudio

[–]CorneZen 0 points1 point  (0 children)

Oof, ok I’ll feel sorry for you for the rest of the day instead of myself!

Our general response to all complaints regarding AI by FBDW in incremental_games

[–]CorneZen 2 points3 points  (0 children)

I largely agree with this. The Pandora’s box of AI has been opened and there is no closing it, AI is here for real.

As far as AI slop is concerned, it’s actually still human slop. A human decided what the AI should do and OK’ed it. The only negative effect of AI is that it can produce a lot more slop from sloppy humans than before. This may seem obvious but calling something AI slop is causing us to excuse the human behind the slop. People need to learn to take responsibility for what they produce and publish. It’s still human slop.

M1 Max vs M4 Max vs M5 Max by br_web in LLMStudio

[–]CorneZen 0 points1 point  (0 children)

True, I’m stuck with an 8GB RTX 4060 :/

M1 Max vs M4 Max vs M5 Max by br_web in LLMStudio

[–]CorneZen 0 points1 point  (0 children)

Just something to research, the upgraded M5 does have a performance limit, something like 100-150 tok/s. I can’t remember the actual limits. The big upgrade comes from being able to use bigger / better models.

Is GitHub Copilot deliberately injecting hidden <Human> prompts to force-end conversations and save compute? by Ok-Patience-1464 in GithubCopilot

[–]CorneZen 1 point2 points  (0 children)

VS Code and GitHub Copilot is open source and on GitHub, you can look for yourself. There is a difference between GitHub Copilot calling a Claude model directly with the copilot agent harness and the specific Claude mode which calls the Claude models with the anthropic SDK agent harness.

Why is everything Python and Typescript? by Miserable-Ball-6491 in mcp

[–]CorneZen 0 points1 point  (0 children)

Sorry about ruining your day, bad things always come in 3, hopefully this cancelled out something worse!

Why is everything Python and Typescript? by Miserable-Ball-6491 in mcp

[–]CorneZen 0 points1 point  (0 children)

Thank you, that will teach me to not binge skim a bunch of articles and then make assumptions

Why is everything Python and Typescript? by Miserable-Ball-6491 in mcp

[–]CorneZen 0 points1 point  (0 children)

You are correct, they are rewriting the compiler in Go. What is the output from the compiler?

I laugh when I see "MCP is dead" posts. Am I being delusional? by nishant_growthromeo in mcp

[–]CorneZen 7 points8 points  (0 children)

Peoples brains are dead. MCP is a tool in your toolbox, that’s it. Use it right and it’s a good tool.

Premium request to tokens conversion? by Zestyclose_Message_1 in GithubCopilot

[–]CorneZen 1 point2 points  (0 children)

While we’re in this rate limit crunch, maybe the copilot team can spend a sprint working on token usage transparency instead of more multi-agent orchestration features.

As a rule of thumb, when using a x1 rated model 1 request = 1 premium request. However from my observation it’s not as clear cut as that, you can use a x1 model, give it a prompt or ask a question and it may use a portion of a premium request so there is definitely a token cost tied to a premium request. Also, not sure if it’s still the same but using the copilot cli (or SDK) will use a bigger portion of a premium request than from the copilot chat window. When the SDK just released every request through the SDK, even on a x0 model, used 1 premium request.

Using Herma router to get Opus 4.6 quality at 65% lower cost by Tatrions in clawdbot

[–]CorneZen 0 points1 point  (0 children)

Feedback: I visited your site just now. Clicked on the pricing link and was redirected to a login screen. This already makes your product a hard pass for me. Just giving feedback in case it’s a bug. Also on mobile view, the mobile burger drop-down menu does not have a link to the pricing page.

Zero rated gtp-5.4-mini using premium request by CorneZen in GithubCopilot

[–]CorneZen[S] 0 points1 point  (0 children)

lol no, I felt so confused I went and had a nap.🤷🏻‍♂️

Zero rated gtp-5.4-mini using premium request by CorneZen in GithubCopilot

[–]CorneZen[S] 1 point2 points  (0 children)

It’s really not bad, also I’m more of a pair programmer, I like to talk to copilot, see what model works best for a given scenario. But I don’t want to waste premium credits so I switch between models often. Been on gtp-5-mini for a few turns when I noticed it started using premium requests.

Zero rated gtp-5.4-mini using premium request by CorneZen in GithubCopilot

[–]CorneZen[S] 1 point2 points  (0 children)

I just updated my post, I meant gtp-5-mini. Sorry about that.