Microsoft's GitHub shifts to metered AI billing amid cost crisis
- Reference: 1777336261
- News link: https://www.theregister.co.uk/2026/04/28/microsofts_github_shifts_to_metered/
- Source link:
The US seafood restaurant's all-you-can-eat shrimp promotion [1]led the company to bankruptcy in 2024 and while Microsoft is nowhere near so financially overextended, the software giant's code hosting biz has decided it no longer wants Copilot to operate at a loss.
GitHub is therefore shifting Copilot from request-based billing to usage-based billing on June 1, 2026.
GitHub absorbed much of the escalating inference cost, but the current premium request model is no longer sustainable
Under request-based billing, GitHub Copilot subscribers will be allowed to submit a set number of premium requests, with certain models priced at a higher request rate but without any consideration for the complexity of the request. So complex prompts that require a lot of "thinking" often cost GitHub more than the company earned in subscription fees.
"Today, a quick chat question and a multi-hour autonomous coding session can cost the user the same amount," explained Mario Rodriguez, chief product officer on the GitHub product team, in a [2]blog post . "GitHub has absorbed much of the escalating inference cost behind that usage, but the current premium request model is no longer sustainable."
[3]
Under usage-based billing, there's a more direct correlation with metered tokens – sets of three or four characters that represent the basic economic unit for selling AI services.
[4]
[5]
It's not quite as simple as $X for X tokens – different models meter tokens at different rates – so GitHub has devised a virtual currency unit called GitHub AI Credits that's worth $0.01.
Copilot customers consume input tokens, output tokens, and cached tokens, each priced based on the model used. Microsoft converts that to a cost measured in AI Credits.
[6]
"Instead of counting premium requests, every Copilot plan will include a monthly allotment of GitHub AI Credits, with the option for paid plans to purchase additional usage," said Rodriguez. "Usage will be calculated based on token consumption, including input, output, and cached tokens, using the listed API rates for each model."
Knowing the outcome of this calculation in advance will be difficult – usage-based billing is non-deterministic, so users can never be sure how much time, and how many tokens, a model will consume to respond to a specific input. Different prompts may involve tools that complicate token consumption calculations.
GitHub at least intends to try to give customers a hint of what's coming. Rodriguez said the company will introduce "a preview bill experience in early May, giving users and admins visibility into projected costs before the June 1 transition."
[7]
AI companies were taken by surprise when OpenClaw attracted widespread attention in February, prompting a surge of experimentation with AI agents running 24/7 on various tasks. And the increasing competency of AI models around this time also encouraged more developers to explore AI coding.
As a result, companies offering subsidized access to AI services through subscription plans faced more demand than they could satisfy with their inferencing infrastructure. The price correction that followed has been rippling across the industry.
[8]Microsoft updates the Windows Update Experience: You can hit pause now
[9]SpaceX dusts off Falcon Heavy for first flight in 18 months
[10]Cursor-Opus agent snuffs out startup's production database
[11]AI reality check: Here's what three companies learned building wallets, homes, and games
GitHub last week signaled its intent to stanch the red ink by [12]suspending the creation of new Copilot, Pro, Pro+, and Student plans.
Before that, [13]Anthropic and [14]Google took steps to limit some uses of its services. OpenAI responded by debuting a more expensive [15]$100 subscription tier in an effort to boost usage of its Codex model, even as the company is [16]contemplating an end to unlimited usage under subscription plans. Cloud providers like [17]AWS and [18]Azure have been dealing with capacity challenges too.
GitHub's subscription rates will remain the same: Copilot Pro is $10/month, Pro+ is $39/month, Business is $19/user/month, and Enterprise is $39/user/month.
At the $0.01 GitHub AI Credit rate, Copilot Pro subscribers get 1,000 AI Credits per month. Copilot Pro+ subscribers get 3,900. Once users exhaust the usage allowed under a plan, they can define an overflow budget – or just stop using AI until the next monthly billing cycle resets their AI Credit balance.
[19]Organizations and enterprises will receive 1,900 and 3,900 API Credits per user per month respectively. However, existing Copilot Business and Copilot Enterprise customers get a higher number of API Credits from June 1 through September 1, 2026, at 3,000 and 7,000 respectively.
Users on annual subscription plans have the option to cancel and receive a pro-rated refund or to be downgraded to Copilot Free upon subscription expiration – those plans will not be renewable. Regardless, those riding their annual subscription plans to the bitter end will see [20]prices skyrocket for premium models. For example, Anthropic's Opus 4.7, subject to a 7.5x multiple under request-based billing, will see its multiplier jump to 27 going forward. OpenAI's GPT-5.4 will see its multiplier rise from 1x to 6x.
The new regime isn't entirely metered. Subscribers who reach their AI Credit limit can continue to access Copilot for code completions and Next Edit Suggestions – these services are unlimited on paid plans.
That's more than you can say about Endless Shrimp, [21]recently revived for a limited time only . ®
Get our [22]Tech Resources
[1] https://www.cnn.com/2024/05/20/business/what-went-wrong-at-red-lobster
[2] https://github.blog/news-insights/company-news/github-copilot-is-moving-to-usage-based-billing/
[3] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=2&c=2afAw45olzAl8M45eCfbr8gAAARY&t=ct%3Dns%26unitnum%3D2%26raptor%3Dcondor%26pos%3Dtop%26test%3D0
[4] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=4&c=44afAw45olzAl8M45eCfbr8gAAARY&t=ct%3Dns%26unitnum%3D4%26raptor%3Dfalcon%26pos%3Dmid%26test%3D0
[5] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=3&c=33afAw45olzAl8M45eCfbr8gAAARY&t=ct%3Dns%26unitnum%3D3%26raptor%3Deagle%26pos%3Dmid%26test%3D0
[6] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=4&c=44afAw45olzAl8M45eCfbr8gAAARY&t=ct%3Dns%26unitnum%3D4%26raptor%3Dfalcon%26pos%3Dmid%26test%3D0
[7] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=3&c=33afAw45olzAl8M45eCfbr8gAAARY&t=ct%3Dns%26unitnum%3D3%26raptor%3Deagle%26pos%3Dmid%26test%3D0
[8] https://www.theregister.com/2026/04/27/microsoft_updates_the_windows_update/
[9] https://www.theregister.com/2026/04/27/spacex_readies_the_first_falcon/
[10] https://www.theregister.com/2026/04/27/cursoropus_agent_snuffs_out_pocketos/
[11] https://www.theregister.com/2026/04/27/ai_use_cases_citi_home_depot_capcom_google_cloud_next/
[12] https://www.theregister.com/2026/04/20/microsofts_github_grounds_copilot_account/
[13] https://www.theregister.com/2026/03/26/anthropic_tweaks_usage_limits/
[14] https://github.com/google-gemini/gemini-cli/discussions/20632
[15] https://x.com/OpenAI/status/2042295688323875316
[16] https://www.businessinsider.com/openai-may-drop-unlimited-chatgpt-plans-exec-says-2026-3
[17] https://www.businessinsider.com/amazon-ai-capacity-crunch-pushed-customers-to-rivals-google-anthropic-2025-11
[18] https://www.theregister.com/2026/04/17/users_complain_of_uk_azure/
[19] https://docs.github.com/en/copilot/concepts/billing/usage-based-billing-for-organizations-and-enterprises
[20] https://docs.github.com/en/copilot/reference/copilot-billing/models-and-pricing#model-multipliers-for-annual-copilot-pro-and-copilot-pro-subscribers
[21] https://www.redlobster.com/news-press/press/2026/04/20/red-lobster-brings-back-endless-shrimp-for-a-limited-time-starting-april-20/
[22] https://whitepapers.theregister.com/
The free ride is nearly over for so-called AI.
Have all you free-loaders learned enough to discover that the over-hyped nonsense is not worth paying for?
Re: The free ride is nearly over for so-called AI.
It is though.
Yep, AI with everything is becoming tiresome and the speed of adoption is deeply concerning, but if it wasn’t genuinely useful, companies wouldn’t be paying for it.
Mr. pusher man
The first few hits are free…
It strikes me as being somewhat suspicious behaviour to all of a sudden claim that the business model was wrong and adjust pricing upward.
Very convenient now that so many companies have adopted copilot work patterns at scale and that bailing out now would be painful.
Well, painful for those who suddenly felt the rush of having a superpower to generate convincing verbosity at the click of a button.
Not so painful for those having to try and wade through it all.
If the price keeps going up, it’s going to be interesting to see how companies adapt.
There could be a lot of egg in faces of those who have leaned on AI who have no actual depth of knowledge when suddenly faced with all their tokens running out with still 3 weeks of a month left. Businesses are surely going to apply stricter caps to keep costs down.
I do hope so as I’m getting tired of having to wade through pages of documentation where a single paragraph would have sufficed.
A bit like my post, except I wrote all this opinionated shit myself.
The Old Shell Game
usage-based billing is non-deterministic
The cover of non-deterministic token consumption is a wonderful umbrella for all sorts of potential billing shenanigans ("revenue enhancement").
Back in the day, we were charged for mainframe computing. CPU seconds, lines printed, cards read, cards punched, terminal connect time, permanent storage used.
But the rates for each of those were published, so unless you did something foolish such as setting a very high CPU time or line output limit on a job which had an infinite loop bug, you weren't surprised when you got your bill.