AI subscriptions are a ticking time bomb for enterprise

thestateofbrand.com

348 points by mooreds 9 hours ago


evo_9 - 7 hours ago

Every AI subscription is a ticking time bomb for the frontier provider; within a few years we will be running local models as good as today’s frontier models with almost no cost burden. The floor will fall out of the enterprise market for all the frontier companies.

pvtmert - 2 hours ago

Although I agree with the sentiment in the article, it smells very LLM~y. Especially the sections and punchlines. Such as: `That is not a rounding error. That is a line item that needs its own budget code.`

ben8bit - 5 hours ago

The entire problem with "AI" is that it's easy to do without. The AI companies know it, the users know it - even the most pro AI agent manager knows it. Thought experiment: remove AI from the world right now, all of it - what do you have? Business as usual. This article doesn't do enough to underscore that - dreaded be the day I need to get an actual engineer to review a PR, right?

returnInfinity - 8 hours ago

Brad Gerstner confirmed that tokens aren't being sold at a loss. Whatever the formula, API + Subscription split, the companies are making a profit on net token sale.

They maybe running at loss after all the salaries and stock comp, but tokens are in profit now.

rakel_rakel - an hour ago

The hyperbolic nature of the articles in both AI camps is very exhausting to me.

I'd like to get in front of a whiteboard with someone who knows economics and the token providers businesses well enough to answer my "explain to me like I'm five" questions. But I'll start with these in here:

Is my observation correct that for the token providers this is a margins game, while for the consumers this is a quality of service/product game? If the quality:margin lines will cross at some point on the x-axis, is the race is to reach this point before running out of money? If yes: What historical examples are there where the delta between these two is huge?

I'm guessing LLM's are unique in a sense, since there's really no limit to how good a consumer of the product expects it to get? (Compared to for example email which is much easier to scale in regards to compute.)

Also extreme noob at life question: Why would you want to IPO before having a sustainable business model? What's the upside?

rvcdbn - 8 hours ago

Article is mistaken these subs are not available to businesses. Companies are paying much closer to API prices. The strategy is to get you accustomed to infinite tokens on your personal sub and bet that behavior transfers to work.

fwipsy - 7 hours ago

Disclaimer: didn't finish tfa, so obviously AI even I could tell.

Perhaps OpenRouter can be used as a benchmark for commodity cost to serve AI. I keep hearing it's better value than Claude, which suggests to me that either Anthropic is especially inefficient for some reason, or they're turning a profit on inference. They could be losing money on training, but I suspect that's just part of the cost of staying a leading lab. If any single one goes under due to debt etc. then companies can just switch?

Sharlin - 8 hours ago

I think I'm going to puke if I see one more "It's not X. It's Y." phrase or the word "load-bearing" used metaphorically.

yalogin - 5 hours ago

If it’s replacing developers it makes sense to cost more than 20 or 100 per month. The real issue for these llm companies is that they are yet to show value in other areas. Without that they will be relegated to just coding. That is the rush right now for them. What other workflows can they automate. I guess every paperwork can be automated. Once the other areas are developed they will switch the pricing model

bilater - 4 hours ago

[flagged]

leemoore - 5 hours ago

Enterprise customers aren't running 20 bucks a month for claude pro subscriptions. My company provides developers about 1k worth of usage limits a month and best I can tell they get maybe a 30% savings off of API cost tops. That's not an insane subsidy. Many other jobs titles are only allowed 50 a month and those folks are constantly running out.

Github Copilot has been doing this with business and enterprise seats, but that will be coming to a head very soon. I expect a fast follow after june when they re-align consumer pro and pro+ accounts.

OpenAi seems to be trying to throw tokens at clients to get lock in. So i'd be most worried about the rug pull that will come from open AI post IPO. Anthropic is already acting responsibly in this area and github copilot is attempting to remediate their insane subsidies in the next several months.

Animats - 2 hours ago

So, will the AI companies raise prices? That's the article's main claim. Uber ran at a loss to build market share for over five years after the IPO. So it it not impossible for an overhyped IPO to run in that mode. The AI service industry might do that, too.

Uber raised prices some, but mostly squeezed drivers harder. When Uber started, driving for Uber was a well paid job. It isn't, now. AI companies are mostly capital cost, so they don't have the oppression option.

Hardware price/performance may not improve much near term. Graphics GPU price/performance hasn't improved much in the last decade. DRAM prices have gone up. Fabs are all booked up. NVidia says not to expect better price/performance before 2030.

More efficient, specialized models are a strong possibility. Dumping all of human knowledge into a coding tool may be unnecessary. Although this would work a lot better if the LLM crowd figured out how to get a reliable "I don't know" answer out of a small model, then call on a bigger one for help.

carra - 6 hours ago

Not just AI. Every subscription in general can be a time bomb. You grow more dependent on it, and the provider can disappear or take it away at any moment.

mojosmojo - 39 minutes ago

They have all switched to usage plus cheap seats based costs for enterprise contracts. the seat costs are typically 20-35% of total spend.

gizmodo59 - 8 hours ago

Inference is profitable. Companies lose money because:

1. Training is expensive. Not just compute but getting the data, researchers salaries etc 2. You have to keep producing new models to ensure people use your inference and there seems to be no end to this. So they have to pour more billions to keep the cycle going on 3. People salary and other admin cost are not that high compared to 1 and 2.

fg137 - 8 hours ago

Why does the author assume that enterprises use subscriptions?

Many companies use models deployed on Azure/Bedrock etc are already paying based on usage (often with discounts).

andyfilms1 - 7 hours ago

Replacing your workers with AI:

--You lose control over their "salary"

--You lose control over their "schedule"

--Your company becomes reliant on another party that does not share your interests or values, and can stop working for you on a whim for any reason

But AI is definitely good and trade unions are definitely bad, apparently...

edwinjm - 2 hours ago

Does the writer understand that for every developer who burns all tokens, there are many people who subscribe just to join the AI revolution, but only ask a couple of questions a day?

nunez - 4 hours ago

Precisely why every bigco is spending $$$$$ buying/reusing GPUs to build their own inference serving stack based on open-source models (usually gpt-oss or one of the LLaMa variants; many bigcos in the US cannot run PRC models). That and having more control over data locality.

Those same companies are getting sweetheart deals with the frontier AI labs in the hope that infrastructure costs go down enough in the future to invert profitability, but it's still a risky position for them to be in. (Having their own infrastructure gives the bigcos huge leverage, even if it's only 80% as good as frontier.)

wegwerper - 4 hours ago

Does this article contain any original thought?

It's clearly llm-spew in its mannersims, making me wonder if there were any nuggets of wisdom in its core or if it in entirety is part of some llm-driven blog spam project?

babajabu - 7 hours ago

Even if they are momentarily losing money it’s important to note the value add they are providing.

If you increase the price, the value is still astronomical in comparison.

Companies need to find a way to leverage local models in tandem with frontier models to offset the costs.

It’s all about targeting specific workloads with the appropriate AI. These tools are not sentient beings they are tools that need to be properly configured to match the job at hand.

jeswin - 8 hours ago

Since we can't reliably detect AI generated crap, I think it makes sense to penalize their submission. I say this as a generally pro-AI person.

alxndr - an hour ago

Darkly funny that Pangram 3.3.1 thinks "100% of this text is AI generated"

prash20026 - 4 hours ago

Just as a counter example, Midjourney is completely self funded and profitable. But they are images, LLMs might be more expensive to train but their inference is cheaper.

So the frontier model companies might have crazy valuations and they might never reach that. But that might not mean they are actually unprofitable.

kaydub - 2 hours ago

I'm surprised at how many businesses are using subscriptions instead of paying per token.

AbstractH24 - 4 hours ago

This is true of every vc backed company they rely on

And some parts of most publicly traded ones.

If it’s not a bootstrapped company with a single offering, there’s a highly likely something there doing is at a loss in the name of growth (and even there, the loss might come in the form of deferred compensation)

exabrial - 7 hours ago

Eventually, after the seed funding is spent, you will have to pay the real cost of the coal used to power your queries.

The best course of action is to take advantage of subsidy for awhile, but not integrate is so deeply one can’t retreat. You’ll still have full productivity, just be cognizant of the reality of the situation.

Hopefully the market eventually collapses to where companies are hosting their own inference, and you simply lease a model package to run on your own (or rented ) specialty hardware.

Havoc - 3 hours ago

Bad attempt to estimate company costs using api sales prices numbers.

There will be a repricing for sure as any ends of subsidies does but the world will not end

wan23 - 5 hours ago

I tried out Gemini in Google Sheets the other day. I asked a pretty simple question and the agent ran for like two minutes trying to answer it until I stopped it. I can't imagine these agentic features are cheap to run for what they get you.

paoliniluis - 7 hours ago

The FED will print to infinity as the US gov can’t stop spending, mostly all of that money will keep going to the only industry that’s growing and provides crazy returns for family offices and VC’s right now which is AI. I don’t agree with the authors opinion here as the “time bomb” timer is simply the entire world buying US debt here, which won’t happen in the short/medium term

throwatdem12311 - 4 hours ago

Not my problem I just burn the tokens they give me!

Yhippa - 4 hours ago

Wasn't this the same thing when enterprises started using cloud computing? Did the bomb explode for them?

blondie9x - an hour ago

I think one thing the author overlooked in the solutions/hedging section is using open weight models. Enterprises need to be ready to use their own servers for inference and build pipelines to utilize non proprietary models when possible.

einrealist - 8 hours ago

Those price increases will increase the pressure to use cheaper / free models (commoditization), thus cutting into the revenue projections of the frontier model vendors. Its going to be exciting to see what happens to these huge investments and valuations.

clearstack - 6 hours ago

MSFT, GOOGL, META are spending $60-100B+ annually on AI infra partly to own the cost floor. the moat isnt the model, its the infrastructure.

542458 - 8 hours ago

I’ve said this before on HN, but there are two things that make me optimistic that we won’t see a big rug pull where price-to-capability ratio skyrockets relative to today:

* People keep finding ways of cramming more intelligence into smaller models, meaning that a given hardware spec delivers more model capability over time. I remember not that long ago when cutting edge 70B parameter models could kinda-sorta-sometimes write code that worked. Versus today, when Qwen 27BA3B (1/23 of the active parameters!) is actually *fun* to vibe code with in a good harness. It’s not opus smart, but the point is you don’t need a trillion parameters to do useful things.

* Hardware will continue to improve and supply will catch up to demand, meaning that a dollar will deliver more hardware spec over time. Right now the industry is massively supply constrained, but I don’t see any reason that has to continue forever. Every vendor knows that memory quality and memory bandwidth and the new metrics of note, and I expect to start seeing products that reflect that in a few years.

I hope that one day we’ll look back on the current model of “accessing AI through provider APIs” the same way we now look back on “everyone connecting to the company mainframe.”

siliconc0w - 5 hours ago

Both OpenAI and Claude already charge Enterprise usage rates and they're still buying.

smoghat - 7 hours ago

How do the owners of that site correlate this with their business model, which is to use AI to write articles like this one, so as to get clients in the news?

oldspleen - 7 hours ago

every infra wave starts with land-grab pricing and ends with metered billing, AI is just running the cycle in 18 months instead of 10 years

JumpCrisscross - 7 hours ago

> A knowledge worker running a few hours of Claude daily, uploading documents, drafting reports, analyzing data, can easily burn through several million tokens per week. At API rates, that same workload runs somewhere between $200 and $400 a month per seat. Some power users push well beyond that. But on a Pro subscription, the company is paying $20 per head. Anthropic is not the only one eating this cost.

What? Anthropic's costs aren't the API rate. The article never attempts to estimate that cost, which renders its thesis tautology.

sunaookami - 6 hours ago

Isn't EVERY subscription and SaaS a ticking time bomb for enterprise?

crorella - 4 hours ago

wouldn't move to local models in the future remove part of that risk for companies?

ninjahawk1 - 4 hours ago

It’s a delicate balance currently. Local models are catching up in breaking speeds while OpenAI is publicly stating they want to sell AI like a “utility” aka only through API pricing.

Meanwhile datacenters put out more pollution and use more electricity than all the plane rides Bill Gates took with Epstein combined, for business meetings of course.

wunderlotus - 4 hours ago

This is an (embarrassingly obvious) AI-generated “article” powered by a company whose business model seems to be AISaaS (AI slop as a service).

ghusto - 8 hours ago

TL;DR to save you time:

1. GenAI companies are making a loss in order to gain adoption and later lock-in

2. ???

3. They're going to cash-in soon and start milking you now that business critical systems rely on GenAI

The "???" denotes a complete failure to offer compelling arguments that link 1 and 3.

dboreham - 3 hours ago

My own interest in LLMs increased exponentially when, around 18 months ago, I saw a post somewhere that had a guy who wrote his own inference engine in Rust and demonstrated it running with downloaded open weight models. I tried it out and was quite amazed that even on my laptop (no GPU) I could get an LLM to write Python programs and engage in discussions about Lewis Carroll poetry. It went from "magic thing that needs a data center of unobtanium GPUs to do questionably useful stuff" to "thing that does useful things even on a regular computer".

There's plenty of sand on the planet and clever people (and AI) figuring out how to do more work with less sand and power, so any argument that AI is going to cost so much that it won't be usable, seems just preposterous.

lmeyerov - 5 hours ago

Not really. Claude Code harness with Sonnet 4.5 model showed you don't really need bigger GPU rollouts, and it's only a matter of time for OSS combos to hit that. Overtime, this will only get better, and the set of enterprise tasks smaller deployments can handle will only go up.

48terry - 4 hours ago

Honestly, this isn't too different from any other software or technology nowadays. "What if the service provider pulls the rug on us and jacks up the price exponentially / begins the enshittification" is (and if you aren't doing it, you should be) a factor when procuring and using anything from a third party anymore.

The software world is, by and large, no longer about making products with a focus on the long-term, whether that's about the customer's well being or even the company's own long-term functioning. It's about trapping people, siphoning their money, then running away after setting the building on fire. Founder McBuilder will throw away his entire userbase and tell them "lol idk good luck" about their usage needs if it means he can make an extra dollar.

This is as true for enterprise as it is for consumers. Look at all the lamenting when a liked name gets bought by venture capital or considers an IPO.

phendrenad2 - 5 hours ago

As inflation plays with 10-year highs, fuel prices go up permanently (thanks to the end of middle east oil), and NIMBYs chase datacenters out of their regions, I think it's inevitable that AI is going to go up in price. It's just a question of how much. Companies should have a fallback plan to either switch AI providers, or replace AI with a pool of new hires quickly.

zephyreon - 7 hours ago

Aside from the obvious fact that this is AI slop, the author (prompter?) doesn’t consider the R&D of AI itself. Efficiency gains, more compute, etc.

We all know every frontier AI lab is heavily subsidizing usage, and so do all of the VCs & CEOs funding them.

jauntywundrkind - an hour ago

Good fucking luck DeepSeek. Thoughts & prayers to you with what's about to hit, shit.

dmazin - 7 hours ago

As a few commenters already pointed out, IME enterprises aren't paying for subscriptions. They're paying per token.

But also... is this shit AI written? I'm so tired of this.

tim333 - 5 hours ago

> the gap between what your organization pays for AI today and what it will pay in 18 months is going to be one of the most disruptive line-item increases most companies have ever absorbed

Colour me skeptical on that one. Unless the AI improves a lot so it makes sense to spend more.

lol8675309 - 3 hours ago

LOL AND DUH

PKop - 8 hours ago

> is not a rounding error. It is

Who said it was?

> Pull out the napkin. This matters.

The article wouldn't exist if you didn't think it mattered, just tell us why.

> the question is not whether they got a good deal. The question is

Who said that was the question?

> This Is Not One Company's Problem

Who said it was?

Stop telling us what thing aren't, just speak like a normal human and convey your own thoughts. It's an insult to your audience to throw constant AI slop at them.

> thousands of companies have woven AI subscriptions deep into their operations. Marketing teams draft copy through ChatGPT Plus.

Yea I bet you do..

submeta - 5 hours ago

This is true. At our company they rolled out ChatGPT with Codex. After two months of happily using it, I got a call from the IT OPs telling me I burnt through four hundred million tokens, 200m a month. And created at least a thousand euro bill. That’s after I used all the credit, but I don’t have all details. The guy told me to „watch my usage.“ What does that even mean. He doesn’t use it himself and apparently he doesn’t know how value is created here and how he can monitor and limit usage.

Did OpenAI switch from fixed prices per seat to usage based? This will surprise many companies I reckon.

Personally I use Claude Code, the 200 euro plan. And am a heavy user. A few weeks ago I realized that CC shows the token usage in cli, in the bottom right. Something I never cared about because I thought paying 200 euro a month will give me „unlimited“ access.

But I guess the party is slowly coming to an end? Prices are going to increase slowly? And the flatrates will be removed eventually?

Too bad, it was nice while it lasted.

Lapsa - 6 hours ago

"In 1975, Dr. Joseph Sharp proved that correct modulation of microwave energy can result in wireless and receiverless transmission of audible speech."

Jbunga - 4 hours ago

[flagged]

freakynit - 8 hours ago

[dead]

huqedato - 5 hours ago

[flagged]

simonw - 8 hours ago

[dead]

huflungdung - 8 hours ago

[dead]

jeremie_strand - 6 hours ago

[dead]

lezojeda - 5 hours ago

[dead]

guesswho_ - 6 hours ago

[dead]

jqpabc123 - 8 hours ago

It is "bait and switch" --- done on an industrial scale.

nrawe - 5 hours ago

This mirrors my own thoughts. Additionally, for businesses looking to replace people (particularly developers) with agentic AI, this is arguably worse from an accounting perspective as the cost of using these services will likely be pure OpEx vs capitalised per my understanding of US/UK GAAP accounting.

alaudet - 5 hours ago

I had a conversation with Claude yesterday about this very topic. The AI was pretty candid about the issue and said many of the same things the author said. Now I am not sure if I went in with an unintended bias and it just went into full sycophant mode, I tried to be neutral in my prompts, along the lines of the implications of integrating AI into processes when the true cost is not being charged. But it was obvious that even moderate usage is a loss leader, so heavy users with agentic workloads are in a risky situation and should think long and hard about their business model if costs slowly trickle up in the triple, quadruple etc etc range.

I will continue to use it as an assistant that does the menial stuff quicker than I ever could, but it's just too early to let it do stuff that would hurt if it disappeared. Enjoy it while it lasts.