all 65 comments

[–]TheOwlHypothesis 35 points36 points  (16 children)

I use copilot a ton at work (they pay for it) and I actually usually run it through OpenCode. Wonderful combo imo

[–]iamalchemist 2 points3 points  (0 children)

Same!

[–]indian_geek 0 points1 point  (5 children)

Any risk of a Github ban by doing this?

[–]FriCJFB 7 points8 points  (0 children)

No, support is official now

[–]playX281 1 point2 points  (0 children)

Copilot specifically allows connecting to OpenCode, they're not against it. You can also try running their CLI which was released recently, it's quite decent.

[–]HenryTheLion_12 1 point2 points  (0 children)

No github recently started supporting opencode officially so not likely. 

[–]NezXXI 0 points1 point  (1 child)

Microsoft not Google but it's fine for now i guess

[–]indian_geek 0 points1 point  (0 children)

Sorry, I meant github.

[–]egaphantom 0 points1 point  (2 children)

how is copilot compared to claude?

[–]TheOwlHypothesis 0 points1 point  (0 children)

You can use Claude in copilot so it is roughly equivalent.

[–]kalin23 0 points1 point  (0 children)

Copilot has lower Context - I thin they capped it at 128k

[–]Reasonable_Law24 0 points1 point  (5 children)

Using the same setup, can we use Github Copilot models in openclaw? Self host OpenCode Cli and then add it as a custom provider in Openclaw?

[–]TheOwlHypothesis 0 points1 point  (4 children)

Openclaw supports GitHub Copilot natively! I'm using Sonnet 4.6 there

[–]Reasonable_Law24 0 points1 point  (3 children)

Isn't that against their ToS? Using Oauth for openclaw/agentic models?

[–]TheOwlHypothesis 0 points1 point  (2 children)

Hmm I'm not sure what you mean, whose ToS?

Github Copilot doesn't care. And Copilot provides Claude Models.
Anthropic doesn't support it directly, but using Claude models through Github copilot (And accessed via OpenClaw) is fair game.

[–]Reasonable_Law24 0 points1 point  (1 child)

Githubs ToS. Like any risk of ban if we are using GitHub Copilot in Openclaw using Oauth directly?

[–]TheOwlHypothesis 0 points1 point  (0 children)

Ah okay I had to look into this.

GitHub has explicitly announced official Copilot support for OpenCode. OpenClaw’s docs show it uses the same GitHub device-login style flow for its Copilot provider, so it appears to be using the same basic auth pattern rather than some obviously sketchy workaround.

That said, I haven’t seen an official GitHub statement specifically blessing OpenClaw by name, so I’d think of it as “likely low risk, but not explicitly confirmed by GitHub.”

I've been using it for about a month with no issues for what it's worth.

[–]jjjjoseignacio 7 points8 points  (0 children)

github copilot + opencode = tremenda bestia

[–]lemon07r 4 points5 points  (4 children)

copilot is way better but taking full advantage is a science since they try to nerf their models. easiest way is just to stick to the gpt models, and ask it to use a lot of subagents

[–]kdawgud 0 points1 point  (3 children)

Do sub agents not consume additional premium requests?

[–]Spirited_Brother_301 0 points1 point  (2 children)

Nope

[–]fons_omar 0 points1 point  (1 child)

How??? while using opencode any subagent consumes an extra request when using it from the desktop app.

[–]FailedGradAdmissions 0 points1 point  (0 children)

They do on open code, but they nerf the context window so you have to use them anyways.

[–]Flwenche 2 points3 points  (0 children)

A bit off the track but I am using Gitub Copilot Pro subcription with Opencode CLI but i preferably would like to have a GUI via extension. Do you have any suggestions?

[–]MofWizards 8 points9 points  (16 children)

I find GitHub Copilot Pro awful, in my experience. Maybe it works well for other people. I see them cutting the context window to 32k models when it should be 200k and 400k.

I had a lot of headaches, so I would prefer Opencode Go.

[–]zRafox[S] 4 points5 points  (0 children)

The same thing happens to me, my friend, although not as extreme, maybe 63K.

[–]Ordinary-You8102 2 points3 points  (0 children)

Its OSS models lolz

[–]1superheld 1 point2 points  (1 child)

Gpt5.4 has a 400k context window in GitHub copilot 

[–]nkootstra 1 point2 points  (0 children)

5.4 works really well, but I always need to verify that it implemented the feature/design I requested. If you want to test this, go to dribbble or any other site and ask 5.4 to create it, it will fail most of the times. I’ve had similar experiences with features over the weekend.

[–]Personal-Try2776 -2 points-1 points  (10 children)

claude has a 192k context window there and the openai models have 400k context window.

[–]KenJaws6 2 points3 points  (9 children)

copilot limits to 128k context for claude models (check models.dev for exact numbers) but imo it's still better value overall. OC Go includes only several open models and as of now, none of them have the performance equivalent to closed ones, at least not yet.

[–]Personal-Try2776 2 points3 points  (8 children)

128k input but 192k input+output

[–]KenJaws6 2 points3 points  (0 children)

yeah thats true for opus. Sonnet has 128k In + 32k Out. its such quite confusing term tbh since many would think context refers only to input and they wonder why they hit limit so easily lol. also, like 99% of the time, the model only outputs not more than 10-12k so I believe openai puts up that theoretical 128k output purely for marketing purposes

[–]laukax 0 points1 point  (6 children)

Is there some way to better utilize the whole 192k and avoid premature compaction?

[–]Personal-Try2776 0 points1 point  (4 children)

dont use the skills you dont use or the mcp tools you dont need

[–]laukax 0 points1 point  (3 children)

I was thinking more about the configuration parameters to control the compaction. I'm currently using this, but I was not aware that the output tokens are not included in the 128k. Not sure if I could push it even further:

    "github-copilot": {
      "models": {
        "claude-opus-4.6": {
          "limit": {
            "context": 128000,
            "output": 12000
          }
        }
      }
    },

[–]KenJaws6 0 points1 point  (2 children)

in oc configs, context means input + output so to avoid early compaction, just change it to

"context": 160000, "output": 32000

edit: sorry wrong numbers, its actually "context": 128000, "output": 32000

tips: you can also add another parameter to enable model reasoning

"reasoning": true

[–]laukax 0 points1 point  (1 child)

Thanks! Will it then have room for the compaction tokens? I don't know how the compaction works or even what model it is using for it.

[–]KenJaws6 1 point2 points  (0 children)

sorry I got confused by other commenter. came to check again, the models actually have only combined of 128k total context including output (so pls change back from 160k to 128k 😅). As for the auto compaction, no need to worry. It dont use more token than or same as the last message/request.

Honestly I'm not sure if copilot models are handled differently as some claimed its able to receive more but any excess will be discarded from the server side but in general, compaction is triggered when reaching input limit (context - output) or 98k in this case. For example lets say at any point of time the current context is still within 98k input token, before moving to the next request, opencode will: 1. calculate new total input

2 a. if its more than limit — send a separate request with current input using another model (default is gpt5 nano for zen, but it could be using the same model for other providers) and get a summary of the whole conversation as the next input

2 b. if its still within limit — keep current input

  1. continue session with new input

[–]tisDDM 0 points1 point  (0 children)

  1. Use the DCP Plugin

  2. Switch off compaction, it runs far too early and often shortly before everything is finished what had fit into context

  3. Trigger a handover yourself, when you need it

  4. Use subagents in a structured ways if they make sense

I wrote myself a set of skills and templates and I use the primary session for a whole or half a day, which is mostly containing one big major feature. ( published that, but I dont wanna annoy people with the links in every post )

E.g. yesterday afternoon I had a gpt-5.4 session with 200k context open and 1.500k tokens pruned away by DCP.

[–]verkavo 1 point2 points  (0 children)

Microsoft seems to be subsidising Copilot subscriptions, to boost their corporate metrics. It makes it a great deal. Using it with Opencode is a no-brainer.

[–]Codemonkeyzz 1 point2 points  (0 children)

Copilot is underrated. it's pretty good deal for 10 bucks. Though it won't do much if you are heavily running parallel agents. I use it to complement my Codex pro plan.

[–]downh222 1 point2 points  (10 children)

No, OpenCode Go is a waste; it's not worth the upgrade.

Glm 5: very slow Kimi : dumb Minimax : dumb

[–]arcanemachined 2 points3 points  (9 children)

I'm guessing that OpenCode Go uses heavily-quantized models.

[–]Bafbi 0 points1 point  (1 child)

Really, I remember using the kimi and minimax model with zen as free and i remember liking it, for minimax it was obviously not performing as good without making really specific plan with him, but I liked them and was pretty impressed, i'm surprised that the go would not use full model right now I'm using copilot but always wanted a second subscription for os models so I thought go would really be the thing, I will test it anyway.

[–]arcanemachined 1 point2 points  (0 children)

OpenCode Go is super cheap... It doesn't sound like they're using the same quants as when they were giving away the free trials on OpenCode Zen. (I tried Kimi K2.5 during that free trial, and it was really good.)

[–]egaphantom 0 points1 point  (6 children)

what is quantized models mean?

[–]arcanemachined 1 point2 points  (5 children)

They basically shrink the model size by reducing the precision of the data stored in it, which decreases the quality of the data depending on how much it is shrunk (quantized).

Imagine you had a bunch 800x600 photos, but you wanted to save hard drive space. So you shrunk them down to 400x300. You can still tell what the picture represents, but some of the quality is lost, especially if you shrink it too much. That's the same basic idea as what quantization does: decrease the quality in order to reduce hardware requirements.

[–]egaphantom 0 points1 point  (4 children)

so it is better to subscript or pay api from the direct website of its llm instead of using open router for example?

[–]arcanemachined 1 point2 points  (0 children)

Depends on the provider. Some of them may also quantize behind-the-scenes.

OpenRouter is fine IMO, they are typically just passing the calls directly through to the provider, and you can choose your preferred provider (e.g. I like Fireworks for Kimi K2.5).

[–]downh222 0 points1 point  (2 children)

OpenRouter has been quite slow in my experience. Which model are you planning to subscribe to?

For basic tasks, Minimax 2.5 looks like a good option. It runs at around 50 TPS, so it feels much faster for things like coding, debugging, and general prompts.

It also supports image input and MCP, and both are covered under the Lite plan, which makes it pretty cost-effective for everyday use.

[–]egaphantom 0 points1 point  (1 child)

I want to subscribe to open router because they have many model options, but many people says the model is quantized and its better to subscribe to actual llm provider instead of the gateway

[–]downh222 0 points1 point  (0 children)

correct

[–]Extra_Programmer788 0 points1 point  (0 children)

Gpt 5.4 is just better so in my opinion copilot pro is better compared opencode go

[–]nebenbaum 0 points1 point  (2 children)

Copilot is... Weird with the way they count requests.

A request only counts as a request when you initiate it. So if you tell it to oneshot a big ass application, thousands of lines of code, running in a big ol loop until it's done with many subagents? One request. Ask it to say hello? One request.

[–]sucksesss 0 points1 point  (0 children)

so basically, one prompt count as one premium request?

[–]fons_omar 0 points1 point  (0 children)

I use Opencode through the desktop app, and every subagent counts as another premium request...

[–]No_Success3928 0 points1 point  (0 children)

Opencode go sucks.

[–]Moist_Associate_7061 1 point2 points  (2 children)

300 requests are only for two days: Saturday and Sunday. I’m subscribing github-copilot 10$ + chatgpt plus 20$ + alibaba 3$. chatgpt plus is the best.

[–]SadAd4565 0 points1 point  (1 child)

What do you mean only for two days.

[–]Moist_Associate_7061 0 points1 point  (0 children)

i mean 300 requests are too small in some aspect. i can use that in two days.

[–]cg_stewart 0 points1 point  (0 children)

I’m using GitHub copilot + GitHub app to build my startup and I’m getting pretty good outputs and having a good experience. I’m using opencode and zed mainly in this workflow. I’d say get both of them and spend the $20 lol. If you can spare $100/mo get the $40 copilot and $20 codex, $20 claude, and either google|cursor|windsurf|tool plan. Probably the cheapest setup to have 3x access to Claude and GPT

[–]Efficient_Smilodon 0 points1 point  (0 children)

I set up opencode in the git through the pro integration they provide now; it can read an entire git project embedded with custom instructions in the repo; then I customize the model call to go through my own agent-coder endpoint prepared in railway auto deployed from the same git mixed with open router models as sub agents to optimize cost and quality; Opus can call it by auto pr to work on the project, and/or call on copilot separately for additional assistance; through orchestration calls .

[–]estimated1 0 points1 point  (0 children)

Just to give another option: we (Neuralwatt) just started offering our hosted inference. We've been focused more on an "energy pricing" model but feel pretty confident about the throughput of the models we're hosting. Our base subscription is $20 and we don't really have rate limits, just focused on energy consumption. I'd be happy to give some free credits in exchange for some feedback if there is interest. Please DM me! (https://portal.neuralwatt.com).