It can bill to our Azure sub and I don't have to go through the internal bureaucracy of purchasing a new product/service from a new vendor.
Opus 4.6 had a 3x multiplier in Pro. Now the new Opus 4.7 model has 7.5x in Pro+, which offers 5x more requests, but costs 4x more than Pro. So now Opus is essentially 2x the price it used to be.
It’s likely that Sonnet 4.7 will be the new 3x model in Pro — https://github.blog/news-insights/company-news/changes-to-gi...
This whole thing is a massive asshole move, and probably illegal in all countries with a minimum set of consumer protections.
I’ve been using Anthropic models exclusively for the last month on a large, realistic codebase, and I can count the number of times I needed to use Opus on one hand. Most of the time, Haiku is fine. About 10% of the time I splurge for Sonnet, and honestly, even some of those are unnecessary.
Folks are complaining because they lost unlimited access to a Ferrari, when a bicycle is fine for 95% of trips.
Haiku is most definitely not fine for the code bases that I work on. Sonnet is probably fine for most daily tasks, but Opus is still needed to find that pesky bug you've been chasing, or to thoroughly review your PR.
Yeah, I hear that a lot, but it never comes with proof. Everyone is special.
I’m sure you’d find that Haiku is pretty functional if there were a constraint on your use.
I mean at some point some people learn...
I was doing Opus for nasty stuff or otherwise at most planning and then using Sonnet to execute.
Buuuuut I'm dealing with a lot of nonstandard use cases and/or sloppy codebases.
Also, at work, Haiku isn't an enabled model.
But also, if I or my employer are paying for premium requests, then they should be served appropriately.
As it stands this announcement smells of "We know our pricing was predatory and here is the rug pull."
My other lesser worry isn't that Opus 4.7 has a 7.5x multi, it's that the multiplier is quoted as an 'introductory' rate.
However, I never expected Opus 4.6 to be removed for the cheaper plans. I expected the pricing model to change, but not to lose access to the model. Moving to being token-based makes sense. It makes the cost more closely aligned with user pricing.
It was nice while it lasted. I got Opus 4.5 to do a lot of work from the beach by assigning it to detailed issues. With this news I've cancelled my Pro subscription. That will help a bit with their capacity issues.
Yes I was hearing that a lot the past few weeks. Its pretty clear what happened:
Anthropic demand soared from OpenClaw, but they were already over-sold. Cowork shipped, Hegseth flexed and a lot of people and entire orgs moved from ChatGPT to Claude in the space of a month. They couldn't handle the demand - they quantized their models, dropped effective usage limits and made all kinds of tweaks in Claude Code to reduce token burn.
Lots of customers fled to Codex and they got crunched as well. Some people noticed Copilot was still selling dollars for nickles and mentioned it to other people.
The only question I care about: Is z.ai next? GLM 5.1 is simply where its at right now. It is not the best model, but it is much better than Sonnet at 1/5 the cost.
I get the impression that the intersection of HN posters and Copilot users is quite small in practice; that Claude Code and Codex suck up all the oxygen in this room. But it seems plausible we’ll see similar “true costs greatly exceed our current subscription pricing” from Anthropic and OpenAI someday soon…
I never understood the low visibility.
Expensive ram is annoying. I don't look forward to expensive ai.
I think this is really telling. The cost of AI has really been masked HUGELY to drive adoption. The true cost is likely to be unsustainable for the big complex tasks (agents running for hours+) that companies have been pushing.
I was skeptical, then quietly bullish on AI, but I'm now seeing signs the market is cracking and the availability is going to receded/costs balloon.
From my simple checks - and from Microsoft's own blog - per token pricing isn't going to be realistic for agentic coding either.
Opus 4.7 is available today for 7.5 credits per prompt.
They have also suspended new signups.
After testing all of the major IDEs/tools that integrate with LLMs over the last four weeks, I was happy to settle on Copilot. I, and others, seem to be a lot confident in that decision. Especially since there seems to be no refund path for people who prepaid for a year.
In my 30+ years online, I've never seen an industry change so much in terms of pricing, service levels, etc, as I have the last two months.
I'm really curious where all of this lands, and if AI coding tools will be something that only a small percentage can genuinely afford at a competitive level.
Warning: baseless speculation/theorizing ahead.
This is the consequence of LLM inference being really expensive to run, and LLM inference companies being really attractive to VCs. The VC silly money means their costs are totally decoupled from revenue for a while, but I guess eventually people look at incomings vs outgoings and start asking questions.
Previous big trends like SaaS apps, NFTs, blockchain etc were similarly attractive to VCs (for a period of time at least for the last two, the first one is still pretty attractive to VCs), but nowhere near as expensive to run so the behaviour of the companies running them wasn't quite the same.
It's the sort of messy job that agents excel at. Decisions need to be made on free text data, translations done into multiple languages, ambiguity handled.
I now need to recheck it still works with another model, which involves a lot of manual verification; and potentially move to Claude Code and pay more money I can ill afford right now.
I'm not even clear from the post when this comes in, I'm guessing effective immediately.
This really hammers home for me the point that we should not be renting our tools.
My own dumb fault for trusting them, I will make sure to learn from this.
It's a shame because the VsCode copilot experience is quite good out of the box compared to all of the other harnesses I've used. But with typical lack of transparency, and sudden, harsh changes... What are they thinking?
After the restrictive rate limiting they've already instituted, I'm simply cancelling and continuing by using providers directly.
I've been using the Pro+ with Opus 4.6 very successfully and being charged 3x rate was mostly acceptable.
But removing Opus 4.6 and replacing with Opus 4.7 with a 7x rate is just insane!
Also, I've never ran into the quota limit before (I only use inline suggestions). The limits have definitely shrunk over time.
It was clear (see the linked post from 70 days ago) that the current offering was unsustainable, but I'm a bit taken aback at how sharp the clawback is.
I was actually hoping they would change it to something that more closely tracks their actual costs so that they wouldn't have to rug-pull this badly. In particular what was really bad about it was that sending prompts to agents while they were working (to give them corrections) cost extra so I stopped doing that (after initially OpenCode didn't cause billing for that, until they became official).
There are no good solutions for them.
If OpenAI is indeed overbuilt they will completely eliminate Claude.
given the recent changes that kneecapped the plan for students [1], i feel less bad after seeing this. always had monthly limit on premium requests shown in the extension (which i would watch in dread creep up), the daily/weekly "usage limits" part seem ambiguous at best.
using agentic workloads as the basis for this change does not sit quite right with me. if you look at the newly added debug mode, you may notice the token consumption as well as the subagent/tool calls made behind the scenes. my takeaways:
- it consumes way too much tokens for simple tasks (had one use case where the agent burnt 16+ million tokens just to make 50 line change in a monorepo using plan -> agent approach)
- even when you select a model in the dropdown, the subagents/tools can be called with an entirely different model, often the haiku-4.5. gpt-4o is widely used for creating summaries or titles to display for the plan.
- the new reasoning modes have exacerbated the token burning as the agent tends to loop a whole lot. the prompt vs plan token ratio is quite minuscule, and when combined with your own instruction files and skills, it just goes out of the window.
i think they have given a generous model in the past, but by kneecapping the lower tier, it no longer justifies existence. if they want to raise prices, they can raise the floor. or rather put some work in improving their own orchestration system before putting the blame on the users vibing it out.
Looks like I'm ending my subscription, good (likely too good, no way my account was even remotely within profitable range) access to opus-4.6 was the only reason I used this at all.
I had the exact same issues with the latter - randomly stops working, wipes chat history, just generally seems to be totally broken. But the former works totally fine and still lets you select sonnet/opus. My experience was before this recent 4.6 -> 4.7 change though.
Now it's going to cost me an upgrade to $39 Github Pro+ to keep using Opus, and even then it's with much higher multipliers. I don't fully understand the extent to which this reflects actual costs for Opus versus Microsoft leveraging network effects to discourage the usage of a competitor.
I didn't really want to wander outside of VSCode just yet because I was happy with VSCode/Copilot/Opus-4.5 and I don't want to spend all my time experimenting when stuff is changing so fast. But I guess my hand has been forced.
This was my first thought too but apparently you can just use Claude Code within VSC: https://code.claude.com/docs/en/vs-code
I guess overall probably was a good decision.
But 7.5x as well as quota limits is pretty hard to swallow.
The annoying thing about the quota limits is they make it really awkward to actually fully utilize the 1500 premium requests you are paying for.
Like if you don’t plan working around the daily and weekly quotas you may not actually be able to utilize your full request allocation.
Claude has the same issue. Single session blows through the quota.
That's not how my creative energy works. I have time that I want to solve problems, and I want to solve them. I don't want a cooldown timer applied to solving a problem. Not to mention the anxiety of realizing that while I sleep I could have burned tokens in that time.
I'm incredibly disappointed when I sat down to my hobbyist programming time and realized copilot was suddenly and dramatically changed in a way that is incredibly disheartening.
Meter my token usage DON'T tell me when I can use them! ARGH.
From what I've been gathering, this split in success seems to depend a lot on the types of tasks, the domains / programming languages / frameworks used, and style of prompting.
I couldn't get 5.2 to follow instructions for the life of me, even when repeating multiple times to do / not do something. 5.3-codex was an improvement and 5.4 while _usually_ decent still regularly forgets, goes on unnecessary tangents, or otherwise repeatedly stops just to ask for continuation.
Sure, I'm paying 3x more per request, but I'm also doing 5x fewer requests.
Or well, used to. Still bummed about them dropping 4.6.
As far as I can tell, the distinctive feature of my workflow is that I'm giving it small, contained single-commit-sized tasks and limited context. For instance: "For all controller `output()` functions under `Controller/Edit/` and `Controller/Report/`, ensure that they check `Auth::userCanManage`." Others seem to be taking bigger swings.
It felt like I constantly have to go back and either fix things or I just didn't like the results. Like the forward momentum/progress on my projects overally wasn't there over time. Even with tho its cheaper it just doesn't feel worth it, to the point I start to feel negative emotions.
I'm actually a bit worried that I've somehow become to feel more negative emotions with agentic coding. Quicker to feel frustrated somehow when things aren't working.
The joke is on them, though (maybe) because this also means that there's literally no reason to keep that account active.
This points toward a deeper issue though. We’ll probably see more individual offerings dry up over time. That means you’ll have individuals stuck with hand coding while the hyper productive AI assisted coders will all be at large organizations. If that happens, we’ll enter a phase where computing will once more be available exclusively to the elite few.
Given that they've already silently had session + weekly rate limits for the past couple weeks already at least (I've hit them), I wonder if this change is just making them visible to the user, or if it's actually tightening them too.
If it's the former then I can say they're still significantly more generous than claude pro (on the pro+ plan), so this might be okay. If it's the latter, and the new limits are similar to claude pro then copilot is going to be significantly less useful to me.
The per-request model was pretty insane.
And you can then cancel it. I have no idea what a premium request is and it's all just too complicated to use.
Copilot (before today) had one of the simplest & cheapest pricing on the market.
Pricing per turn/request was/is an idiotic model and I'm glad they are paying for it. It just forces you into a workflow just to work around business model. Heck the best laugh would be to create a plan outside vscode with interactive CC/Codex then copy paste into GH copilot to do a single session burn of few M tokens.
Again ridiculous model.
I subscribed two months ago, frustrated with Claude Code and their tight session limits.
The Copilot offer was unbeatable 100 dollars for a 12 months plan, if I remember correctly.
It was pretty clear they were losing money, but hey, it's Microsoft and they need customers, so a competitive push on pricing is expected.
Let's see what these limits look like and I'll decide whether to cancel my subscription or not.
Still a terrible move from them.
I'm a paying customer and I did not receive ANY communication about this. Was using Opus this afternoon and then it disappeared.
Microsoft really can't stop being Microsoft. I don't dispute the need to charge more for those models, but there is a basic decency to do things and as usual the Big Tech fuckery and complete lack of morals makes them do this in a way that generates total mistrust where it could be just annoyance.
I'll see how Sonnet handles the most difficult problems but I'm foresee a subscription cancelation soon.
They're all operating at a loss, enshittification is coming for us all.