If anything, these new multipliers are more transparent than anything OpenAI or Anthropic have communicated regarding actual costs and give us a more realistic understanding of what it's costing these providers.
The fact that we were able to get such a substantial amount of usage for $20/$100/$200 a month was never meant to last and to think otherwise was perhaps a bit naive.
This feels like a strategy from the ZIRP era of tech growth where companies burned investor capital and gave away their products and services for free (or subsidized them heavily) in order to prioritize user acquisition initially. Then once they'd gained enough traction and stickiness they'd then implement a monetization strategy to capitalize on said user base.
There's going to be a limit to how much they can raise prices, because someone can always build out a datacenter and fill it up with open source DeepSeek inference and undercut your prices by 10x while still making a very good ROI--and that's a business model right there. Right now I'm sure there's a lot of people who will protest that they couldn't do their jobs with lesser models, but as time goes on that will get less and less. Already right now the consumers who are using AI for writing presentations, cooking recipe generation and ELI5 answers for common things, aren't going to be missing much from a lesser model. That'll actually only start to get cheaper over time.
Also for business needs, as AI inference costs escalate there comes a point where businesses rediscover human intelligence again, and start hiring/training people to do more work to use lesser models--if that is more productive in the end than shelling out large amounts of cash for inference on the latest models. [Although given how much companies waste on AWS, there's a lot of tolerance for overspending in corporations...]
Not sure how it all works out. Currently trillion dollar companies can't make a native app for platforms. Everything is just JS/Electron because economics does not work for them.
And here companies can make GW data center running very expensive GPUs for 1/10th of current prices. Sound little fanciful to me.
Now, I have this high-resolution shiny object that can near instantaneously get any information I want along with _streaming HD video to it_ *anywhere*.
15 years even feels like a stone age. I can't fathom what it has to feel like people in their 60s and 70s.
That's very close to a normal day in 1996. The biggest difference is I read the news on my phone instead of a physical newspaper. The news was not any more interesting or informative because of that. I guess I can also still do the loop reasonably well, but I'm a lot slower than I was in 1996 when I was a cross-country state champion.
My parents are closing in on 70 and I guess I can't speak for them, but I'm at least aware of the daily routines of their lives, too. Walk the dog, do housework, DIY building projects, visit kids and grankids. Seems much the same, too, with the biggest difference being they're now teaching my sister's sons to play baseball rather than me, but shit, one of her sons even looks like exactly the same way I looked when I was 7! The more things change, the more they stay the same.
But - now they are easier - I can read books on an e-ink screen and pretty much instantly find what I want to read next. I get my news on a phone. I used to watch TV/movies broadcast or on tape rentals. Now, I have just about everything I could ever want available - without ADs... those were such a time-waster.
What has changed is that I have access to MORE information than my local (or school) libraries could ever provide - in a variety of more accessible formats. Whatever tools I need to get "work done", I can find a myriad of free and open-source options.
But - the overall days and household family routines are the same - now, instead of reading a paper book while waiting to pickup my kids (or other family members) "back-in-the-day", I can read my device, or connect with my DIY communities online on my phone - or learn something new. I don't have to schedule life around major broadcast events, I can easily do many tasks while I am "out-and-about".
Friction has been reduced.
I am just over 50 myself and I agree with your points. Technology has changed but life is largely very similar to wear it was in the 90s. At least day to day. Attitudes are way worse now.
I always wonder the views of older people. My parents are very technology forward and have been my entire life so it is difficult to gauge how different life is compared to when they were growing up.
It's easy to hear "Oh well I only had 640kb of memory and typed programs out of a magazine I got in the mail!" and see as distinct from having 'unlimited' resources and the internet.
Your insight is good ("The biggest difference is I read the news on my phone instead of a physical newspaper") that life sort of stays the same but the modality changes. People still go to the store like they did in the mid-1800s but now it is by car.
I wonder what our "industrial revolution" will be where the previous generation lived (ie: out in the country on a farm) totally different lives to the current (ie: in the city in a factory). Maybe when space travel and multi-planetary living is normalized?
Since I was there (young, but there), I want to point out that this crosses three eras which all felt quite different:
1978: typed programs in from a magazine or loaded from a cassette (16kB, TRS-80)
1983: loaded programs from a floppy (64kB, Apple ][ and C64 etc)
1988: loaded programs from a hard disk (640kB, IBM PC and Mac).
Exact years vary but these eras were only about 5 years each. Nobody had a floppy in 1978 but almost computer user did by 1983; nobody had a hard drive in 1983 but almost everyone did by 1988.Housework meant no laundry machine, no dishwasher, and possibly no vacuum cleaner. That means hand washing everything, and beating rugs with sticks and brushes to get the dust off of them.
I look at the life my kids live, and it's not so different to my childhood. The toys are similar, their housing is similar. Probably the biggest difference is the availability of content on demand rather than much more fixed TV schedules.
The big difference in the last 30 years hasn't so much been in the kind of middle class life you can live, but the number of people who live that kind of life. In the 90s 40% of people globally were living in extreme poverty. Now its under 10%. The kinds of lives the middle class live in China and Vietnam are closer to those of Europeans today, when even 30 years ago most people in those countries were living much closer to the way your dad grew up.
I wonder if AI will result in a step change of living standards? Perhaps along with robotics we'll finally get to do nothing at all at home? I'm not convinced it'll be quick though. Maybe another 30 years.
And at some point even frontier model costs will hopefully come down (if there is still a meaningful difference between closed and open source models at that point) as all of the compute that's being built out right now comes online.
It has been years now, of cash injections, investors can't keep feeding the beast forever.
That would be, even is, the smart thing to do.
In other words.
The bubble has burst. You're just in denial.
Deepseek API pricing is very low compared to Anthropic/OpenAI API pricing.
For many, the 300% difference in pricing may be difficult to justify, if the quality difference is very small. And there will be many tasks where the most expensive/the best model, is not needed. Currently many people end up using Opus 4.7/GPT 5.5 for many tasks without thinking about it.
I am not shilling China, this is just what is happening right now.
I think the Chinese government works differently than the US government. I think China has been subsidizing their electricity grid for decades and leading the world on sustainable electricity namely solar. While the us has let their infrastructure rot and laughed at government inefficiencies for about half that time. The US has data centers running on gas right now while waging wars blowing up gas infrastructure world wide. It would be comical if it wasn't an environmental disaster. Most of them have no hopes at even getting enough power in well established areas short term.
I realize what I am saying may come off as propaganda because the US holds net negative views on China so here are some links.
https://www.technologyreview.com/2025/07/10/1119941/china-en...
https://www.wired.com/story/data-centers-are-driving-a-us-ga...
I think because openai spent so much money upfront showing how it was possible to do this and laid out a product roadmap China got to get on board much cheaper and easier. I see no reason to not believe any of these companies when they say they didn't squander tons of money to do what they did because I don't know how openai has even spent all the money they have it's actually ridiculous to think about.
https://the-decoder.com/openai-adds-111-billion-to-its-cash-...
Unlike the us chinas focus is on research and sustainable building. China also has really good infrastructure for energy, etc. it is also to their advantage to drop 5 billion instead of 2 trillion and beat the us while turning a profit.
Chinas focus in ai is less flashy and because they are the biggest manufacturing super power in the world right now, it directly feeds their economy. They aren't looking for applications or to replace thought workers with slop bots, they have natural needs for this technology. Us manufacturers can't compete so they have to keep companies from selling their goods there see byd. China sees it as commoditizing their complement, the us is risking its entire economy and it's environment and resources, kind of scary.
If/when it gets to the point where it can replace a skilled worker, the service can be sold for close to the same price as that skilled labour. But the AI can run 24/7, reliably, and scale up/down at a moments notice.
There's not going to be much competition to drive prices down, the barriers to entry are already huge. There'll likely to be one clear winner, becoming a near-monopoly, or maybe we'll get a duopoly at best.
Yes, a lot of people (not me). Why? Well because that was the whole value proposition of these companies, relentlessly pushed by their PR and most of the media- rememmber it was something something Pocket PhDs, massive unemployment etc?
Based on what exactly? So far every time OpenAI, Anthropic or whatever has released a new top performing model, competitors have caught up quickly. Open source models have greatly improved as well.
I expect AI to be just like cloud computing in general - AWS, Azure, GCP being the main providers, with dozens of smaller competitors offering similar services as well.
I think the future of ai will be breakthroughs that let it run on commodity hardware, and the average person will not be paying for it from the cloud unless they want to be surveilled or are stuck on older hardware.
Right now I am running about what was a frontier model 1-2 years ago on a junk machine. Some people are running what was a frontier model 4 months ago on PCs and laptops that cost 5,000. In a year I think the landscape will be even better.
Even if SOTA models in the cloud are a few percentage points better, most work can be routed to local models most of the time. That leaves the cloud providers fighting over the most computationally intensive tasks. In the long term, I think models are going to be local-first.
(Unless providers can figure out a network effect that local models can't replicate).
Why on earth would that happen when everything else is moving into the cloud to tie it to ever-escalating subscription fees and prevent piracy?
Even with gaming, where running high-end 3D games in the cloud seems like madness and inevitably degrades the quality of the experience, they won't stop trying.
Why? There's an inherent efficiency advantage to scale, while the only real advantage for local models (privacy/secrecy) hasn't proven convincing for broader IT either.
The reason local models hasn't caught on is several fold. It's marketing to say your company follows the latest trend, and there's an inherent pressure to keep AI companies afloat so the economy doesn't entirely collapse. The other is, it wasn't until the last month that these models have caught up to frontier models. They just did, and they are more efficient and don't require a team of 500 to deploy.
Maybe in a world where these AI companies behaved with some semblance of ethics and user-friendliness they would be on even ground, but for anyone paying attention local models are obviously the future.
Because of nonexistent regulation. Just wait for it…
The legal situation in for example the EU is crystal clear, only that it will take some time to go though all court instances.
Even with overhead and scaling for peak use and a large profit margin, any company with an ounce of competition will be vastly cheaper than self-hosting. And for models you can run yourself, there will be plenty of competition.
Considering most of the cost of producing a model is the upfront cost rather than the running one, I kinda still do.
The point was never to produce 4 frontier models per company a year.
I see statements like this as strong indicators that the sales people are wrapping up their work and the accountants are taking over. The land rush is switching to an operational efficiency play.
I've been wanting to get off MS more generally and this is good motivation. Will be playing round with OR this week.
Do inference providers have standardized endpoints, or at least endpoints compatible with claude code? Otherwise to pay 5.5% on all your tokens just so it's slightly easier to swap providers (ie. changing a few urls?)
Yep, you can plug deepseek/kimi/minimax into claude code just fine. Or run everything through another harness like opencode instead.
Apple still charges 30%. 5.5 seems pretty reasonable. /shrug I dunno.
Don't you still need to handle tokens with them? Also that's trivial.
> billing
Yes but you'd be paying for billing anyway.
> reliability
They increase reliability?
> middleware
Which you wouldn't need if you paid directly.
I'm not saying they shouldn't get 5.5%, but that list is mostly non-convincing.
> Apple still charges 30%.
3 of the 30 is for billing, with the rest mostly being gatekeeping with a fake justification on top.
Still worth it IMO to be able to switch from Provider A to Provider B if Provider A is having a bad day.
I had copilot mainly so I could write issues and throw agents at it, while I went off and did other things. Has been great for contained spot work.
At this point, I'll go ahead and leave it expire, and then consolidate between Codex and JetBrains AI. Especially since Xcode supports Codex with a first-party integration.
The only model I even used on Copilot was Sonnet and now its got a ridiculous multiplier.
At this point they might as well just charge per Million tokens like every other provider instead of having a subscription.
Pretty sure that's what they will eventually do
Does it effectively bypass regional restrictions for you, so you can use something like the Claude API from unsupported regions such as Hong Kong, or does it still enforce the official providers' geo-restrictions?
You can pay with crypto though, which seems to be convenient for people under sanctions or with limited access, or if you are in low-tax jurisdiction (e.g. HK)
That said I think few people using openrouter are actually being selective about providers.
It took half a day to get my opencode setup, was not friendly. A lot of manually cross referencing model and providers. I was actually mainly optimizing for relatively fast providers. It all is super fragile and I'm sure half out of date; I have no idea if these picks are still fast, no promises they are still the same price (pretty terrifying honestly).
I'm mostly on coding plans so it doesn't super affect me. But man is it a bother to maintain.
Also, the multiplier of 27 for Claude Opus 4.6/4. is way higher than the increase in API price would suggest.
I wonder why that is.
That is not my experience. Each model since at least GPT-4 can fill up an entire context window. In fact, more powerful models can solve tasks faster, so their ratio of multiplier to API price should decrease, not increase.
For example, Claude Sonnet 4.6 has a multiplier of 9 and an API price of $15, which is 0.6 multiplier per dollar.
Claude Opus 4.7 has an API price of $25, so it should have a multiplier of 25 * 0.6 = 15 when extrapolating from Sonnet, but the multiplier is 27.
> Also, they tend to use more thinking tokens.
That might be it. Is there any data on this somewhere?
In not-too-distant future we're going to be running better models on our phones than we can buy access to today in the cloud. Skate where the puck is going: soak the customers until that day comes.
Provide cheap and unlimited access to Grok for programmers (hence the Cursor partnership/purchase for distribution).
-> This would drag massive revenue right before the IPO announcement, like if the company is super growing
-> At a loss, but don't worry, we need these funds to build the biggest datacenter of the universe.
This announcement would create enough momentum to increase valuation, and because of the merge of his companies, would save his X/Twitter investors from a tragedy.
-> Would also be a great service to Cursor investors and so, who are stuck with their VSCode fork
But they can't buy curser before their IPO so thats that?
Perhaps they have to much compute because Musk overpromised and Twittergroq doesn't need that much compute after he nerved the porn stuff?
One theory I think Matt Levine posited, is that SpaceX will go public with dual-class stock that gives Elon control even with a minority ownership stake, and will subsequently buy Tesla, which doesn't have dual class stock, making SpaceX the singular "Elon Musk company", with him having operational control despite being public.
That GPT4-mini change is going to be brutal! Its much better than 5-mini, which was itself much better than earlier free models.
(I know openrouter is not open, but it allows competition and should be easily replaceable if needed)
It's a convenience cost, for sure, but it's not valueless in a fast-moving world. Certainly if you're comfortable with one provider and it's cheaper, do that.
And while i do not spend 200$ privat, in my startup we discussed this and our current mental model is, that instead of hiring someone new, we prefer to have more money for tokens.
This is easier for us and has a bigger benefit. The cost of a new / first employee is very high, a 200$ subscription is not. Upgrading that to lets say 400 or 800$ is still alot easier and if i can run multiply and better agents with that money, lets goooo.
And Gemma 4 and other open models can easily be hosted even for schools.
Another reason to hate that word.
From a different perspective, you were granted an incredible gift from the companies who let you use their product on their dime. Hopefully you made the most of it when you had the opportunity.
Okay then this AI stuff isn’t an example of that even under your definition.
OpenRouter is guaranteed to be about the highest margin operator in the business right now. Everyone wishes they'd be them, skimming 5% off as the middleman without any OpEx.
The 5% fee probably has to factor in Stripe's fees, which would be around 3% to 4% depending on whether it's an international card.
They also show headline prices for the cheapest provider of whatever model, but then need to hit different backends some of which may be more expensive. For now they absorb those costs, but the VCs always come knocking.
Just my opinion though. Totally agreed that they have one of the best positions amongst all AI providers from a financial standpoint.
They do?? I was under the impression I was just playing the price for whatever provider they deemed 'best' for each completion.
Checking now: The way they describe it in their FAQ is that if the price changes, then they will bill you the new price. But I read that as regarding if the primary model provider changes their headline token cost; not in the case of pricing differences for models that have many different backends that host them.
Regardless, I would be more concerned about the streaming costs if the service continues to blow up and they scale aggressively through VC investments. If their 5.5% skim accounted for what they needed, you'd think they could effectively grow organically..
Second, you have no idea what their costs are. It is most likely that they are simply passing on their costs to you. If that was not the setup, users would just go to another service provider who was providing tokens at a cheaper rate. It's not like there is a dearth of competitors in this business.
Now they just increase the price to buy it back