The AI industry needs token volume to grow 50,000 to 100,000 times larger by 2030 to justify the infrastructure it is currently building. The problem is that AI agents — the tools supposed to deliver that growth — are currently widening the gap instead.
That contradiction is now reaching enterprise balance sheets. On Thursday, a startup called Portal26 launched a product to cap runaway agent token costs, citing Uber Technologies as a company that discovered adoption speed and cost predictability are on a collision course. The product is new. The problem it is solving is not.
To justify the $6.3 trillion in datacenter capital the industry has committed through 2029, AI providers would need to generate roughly $2 trillion in annual AI-driven revenue by 2030, according to Gartner. Current global token processing sits at approximately 100 to 200 quadrillion tokens per year. The volume required to hit that revenue target is closer to 10 sextillion tokens annually. That is the 50,000 to 100,000-fold gap. Not a rounding error. Not a growth curve. A structural math problem — one that the industry's proposed solution is currently making worse.
AI agents can consume between 100 and 10,000 times more tokens per task than the chatbots that preceded them, according to Max Kan, a tokenomics analyst at SemiAnalysis. A chatbot is a handful of turns. An agent that searches the web, writes and tests code, loops with other agents, and runs around the clock is something else entirely. The tool built to solve the volume problem is currently the largest contributor to its own worsening.
The enterprise response is beginning to look like sticker shock. A Goldman Sachs survey of large companies found that AI inference costs are approaching 10 percent of headcount budgets at some organizations, and on track to reach parity within several quarters at current trajectories. The KPMG Q1 AI Pulse survey, drawing on 237 U.S. C-suite leaders at companies with over $1 billion in annual revenue, found that U.S. organizations are now projecting an average of $178 million in AI spending over the next twelve months, nearly double the level from a year ago. The survey also found that 54 percent of organizations are actively deploying AI agents, up from 12 percent in 2024. Adoption is real. So is the bill.
"Even, like, a quarter, two quarters ago, nobody bothered about LLM consumption costs," said Swami Chandrasekaran, head of AI and data labs at KPMG North America. "But this year, those AI consumption costs have been rising to levels that rival human labor costs."
What the math actually requires is not a growth story. A 10x improvement in anything is a banner year. A 10,000x improvement is a different category of problem. The industry needs something closer to a 50,000 to 100,000-fold increase in total token throughput by 2030 to make the revenue targets cohere with the infrastructure commitments. Some of that may come from agent deployment scaling. But the same agents doing the scaling are the ones currently blowing out compute budgets, which is the contradiction at the center of the current moment.
The visible sign of that contradiction arrived in April, when Anthropic moved to restrict third-party access to Claude Code, the coding tool that had become the backbone of a significant slice of the agentic AI ecosystem. Our subscriptions were not built for the usage patterns of these third-party tools, wrote Boris Cherny, head of Claude Code, in a post on X. Microsoft has made similar adjustments to GitHub Copilot pricing. The era of subscriptions priced on chatbot-era usage assumptions is ending.
On the provider side, OpenAI announced $600 billion in spending commitments through 2030 earlier this year, a figure that represented a downward revision from its prior $1.4 trillion target. The direction of travel matters more than the headline number: compute spending is being repriced across the board.
The cost curve will eventually bend. Gartner forecasts that the cost of performing inference on a one-trillion-parameter language model will drop more than 90 percent by 2030. That is real and meaningful. But Gartner's own research notes that demand for advanced model features is growing faster than prices are declining, and that the more capable models driving agentic deployments are the ones where per-token costs remain highest. The cost curve bends in the right direction. The adoption curve is moving faster.
The unconstrained deployment of these more advanced AI tools, you know, like agents, is not a viable strategy for adoption, said Will Sommer, an economist at Gartner, in an interview with The Verge. How much budgets will actually need to grow, how much folks can have to spend is an open question.
Portal26 is betting the answer is: budgets will need to grow, and someone will need to cap them before they spiral. Whether that product solves the structural math or simply makes it visible is the question the industry has not yet answered.
Sources: SiliconANGLE | Marketplace | KPMG | The Verge | HPCwire/Gartner | CapitalAI Daily | SemiAnalysis