I really dislike these AI middleman plans. The value-add that Microsoft brings to Github Copilot is near zero compared to directly buying from Anthropic or OpenAI, where 99% of the value is being delivered from. I don't understand why anyone would want to deal with Microsoft as a vendor if they don't have to. The short period of discounted usage was always the obvious rug pull.
> I don't understand why anyone would want to deal with Microsoft as a vendor if they don't have to.
It can bill to our Azure sub and I don't have to go through the internal bureaucracy of purchasing a new product/service from a new vendor.
I would also add that the models they supply through Azure Foundry are covered under my employer's existing customer agreement, by which MS is not allowed to train models on our data (which might include IP of the company or its clients). For organizations worried about that, it's nice & cozy.
Bingo. Github Copilot is mostly for organizations that have an existing Azure bill and would rather see that go up then get a new vendor bill. Professional middlemen.
This is pretty straightforward compared to the giant universe of companies that resell Microsoft services.
The number of intermediaries that some customers, especially governmental agencies, go through to get just an Azure bill can be wild...
"than"
It matters in this case. Learning the difference isn't hard.
What reasonable interpretation of the sentence is there if "then" is applied literally? I can only find validity using "than", and therefore the use of "then" doesn't matter as the author's intent isn't lost. That said, carrying the assumption that it does matter forward, how are you certain "then" isn't the correct interpretation of the author's intent?
This is the kind of argument you should take to your 3rd grade teacher.
The trouble with your plan is that nobody outside of high school debate club is going to spend time with an argument.
It's not my plan.
If you’ve ever had to be part of the frankly batshit insane procurement process that some organizations force you to gauntlet through, it becomes a very obvious and appealing option to do this
Ah, the AWS Marketplace procurement model, where products mostly exist so that you can line item things through Amazon rather than going through a lengthy procurement process
I disagree. I like the standard interface, being able to easily switch models as things invariably change from week to week, and having a relationship with one company. That's why I'm a big fan of openrouter and Cursor. Not too much experience with Copilot, but I think there's a huge value add in AI middlemen.
Because if you’re a vscode user up until a couple days ago you could hammer Opus 4.6 all day every day and pay nowhere close to the Claude Max plan. Many people exploited this and the subsidy is closing.
Yeah this was me. I just got a message that I hit my limit and now I am looking into what it takes to run Qwen on local hardware.
A suggestion: Don't invest in any new hardware to run an LLM locally until you've tried the model for a while through OpenRouter.
The Qwen models are cool, but if you're coming from Opus you will be somewhere between mildly to very disappointed depending on the complexity of your work.
Been having a ton of fun with copilot cli directed to local qwen 3.6. If you’re willing to increase the amount of specificity in your prompts then delegating from a GPT-5.4 or Opus to local qwen has been great so far.
I have to say this was how I used GitHub copilot in vscode. I Used opus 4.6 for most tasks. I am not sure I want to keep my copilot plan now.
Just use claude code directly with a pro plan instead of copilot for roughly the same cost.
On wait, nevermind.
The Anthropic Pro plan cost double and gave you, I don't know, a tenth the usage, depending on how efficiently you used Copilot requests, and no access to a large set of models including GPT and Gemini and free ones.
> Just use claude code directly with a pro plan
Usage limits are/were higher in Copilot. They also charge per prompt, not per token.
Well they charge per prompt, but with usage limits it is a mix of token and prompt. If prompt multiplier is higher, tokens are also multiplied, so limit is reached sooner.
It is basically a token based pricing, but you get alos a limitation of prompts (you can't just randomly ask questions to models, you have to optimize to make them do the most work for e.g. hour(s) without you replying - or ask them to use the question tool).
Yes, I loved my $10 a month person subscription for light coding tasks, it worked great. I'd use claude code max for heavy lifting, but the $10 a month copilot plan kept me off cursor for the IDE centric things.
Me too. Claude isn't the best option when all you do is ask "what's this error message", every 10 minutes or so.
Good, I hope Microsoft lost a lot of money in the deal.
From a friend in GitHub: they've been burning so much money because of Opus.
Opus 4.6 is no longer available and Opus 4.7 chews through monthly limits with reckless abandon. The value-add of GH Copilot is basically gone (at least for individuals on the Pro or Pro+ plans.)
Copilot was there in AI based development first with tab completions.
Now, it may be the right call to immediately give up and shutdown after Opus 4.5, but models and subscriptions are in flux right now, so the right call is not at all obvious to me.
The agentic AI models could be commoditized, some model may excel in one area of SWE, while others are good for another area, local models may be at least good enough for 80%, and cloud usage could fall to 20%, etc. etc.
Staying in the market and providing multi-model and harness options (Claude and Codex usable in Copilot) is good for the market, even if you don't use it.
It was so much cheaper! I subscribed with the monthly plan instead of the yearly one thinking that the deal won’t last. It has last a bit longer than expected.
I found the Copilot harness generally more buggy/disfunctional. After seeing a "long" agent response get dropped (still counts against usage of course) too many times I gave up on the product.
It doesn't matter how competent the actual model is, or how long it's able to operate independently, if the harness can't handle it and drops responses. Made me think are they even using their own harness?
At least Anthropic is obviously dogfooding on Claude Code which keeps it mostly functional.
- [deleted]
Except Copilot doesnt bill you per token like all those companies do, they bill you per prompt, at least Copilot in Visual Studio 2026 which is insane to me, are they just hosting all those models and able to reduce costs of doing so?
No they are taking the massive L. Thats why they paused new sign ups.
Just for context to the insanity, they allow recursive subagents to I believe its 5 levels deep.
You can make a prompt and tell copilot to dig through a code base, have one sub agent per file, and one Recursive subagent per function, to do some complex codebase wide audit. If you use Opus 4.7 to do this it consumes a grand total of 0.5% of a Pro+ plan.
Thats why this paragraph is here:
> it’s now common for a handful of requests to incur costs that exceed the plan price
No, like every other provider they're just losing money and hoping this will some day magically become profitable
one subscription for access to most of the models..
I was accounting for that in the 1% of value. I don't see a ton of value in this for development, you end up just always using the smartest model, with maybe tuning subagents to slightly dumber but much faster model. You really only need one subscription to the provider of the smartest model, with maybe 30 minutes of setup time to switch over if SOTA ever switches back to OpenAI.
Some Opus models were free on Copilot, and in my country you cannot attach a repo to Gemini, that is limited to their premium offerings.
Which Opus models were free on Copilot?
I also just saw:
> Claude Code to be removed from Pro Tier? > https://news.ycombinator.com/item?id=47855565
I have thought about making a product out of something I'm building and trying to make the cost of my product a percentage on top of whatever I could resell Anthropic or OpenAI (or whatever) tokens for. I get this may be unpopular, maybe I should just stick with BYO-key.
It makes enterprise deployments much easier because most orgs already have github enterprise.