The AI industry promises to upend every supply chain on earth. It is still waiting in line itself.
Somewhere in the past few weeks, a story started circulating: the AI chip shortage is over. GPU prices are falling, the queue is clearing, the buildout is on track. The data says otherwise.
GPU rental prices at Lightning AI rose more than 25 percent in the past six months, from roughly $1.60 per chip per hour to above $2.00, according to BigGo Finance reporting The Information's data. At AI startup Krea, the same metric climbed 32 percent in the same window, from $2.80 to $3.70 per chip per hour. Lightning AI has approximately 40,000 GPUs humming in data centers. It has 400,000 units in pending client orders. That is a 10-to-one backlog ratio, and it is not getting better.
Microsoft Azure has told its internal staff to inform clients to expect wait times lasting at least through the end of 2026, according to BigGo Finance citing The Information. The company has imposed a three-tier client hierarchy. Tier one means roughly 1,000 high-spending clients with priority access. To get a Nvidia Blackwell chip allocation right now, a client must commit to at least 1,000 chips for a minimum of one year, with contracts reaching tens of millions of dollars. Even clients on older chip generations face waits of weeks or months. A "use-it-or-lose-it" policy means clients who leave GPUs idle for a few hours risk having their access revoked.
The memory underneath the compute is equally tight. DRAM manufacturers will satisfy only 60 percent of projected demand through 2027, according to AI Business Review citing industry analysis. High-bandwidth memory suppliers preallocated their entire 2026 capacity months ago. The companies that make HBM — Samsung, SK Hynix, and Micron — are earning gross margins of 60 to 70 percent, which is what happens when you control the only door into a room everyone needs to enter. New fabrication facilities require 18 to 24 months from groundbreaking to first production, as noted in the same AI Business Review analysis. The supply that might ease this shortage will arrive in 2027 at the earliest.
The five largest hyperscalers — Amazon, Microsoft, Google, Meta, and Oracle — have collectively committed more than $660 billion in capital expenditures for 2026, nearly double 2025 levels, according to Omdia analysis reported by Manufacturing Dive. Amazon alone is planning $200 billion, up from $131.8 billion last year. That money is real and it is being spent. But the physical infrastructure it is buying runs on a timeline that does not accelerate on command. Industry analysis projects 30 to 50 percent of planned 2026 data center capacity will slip to 2028. Grid connection processes take three to seven years. Transformer lead times run multiple years. The compute buildout is happening; the buildings to house it are not ready.
The irony is that the companies spending the most to solve this problem are the same ones whose spending is making it worse. Every hyperscaler competing for the same finite pool of HBM, CoWoS packaging slots, and power infrastructure is bidding against every other hyperscaler. The $660 billion in capex is not creating new capacity fast enough to outpace the demand that $660 billion represents.
Some companies are attempting to route around the bottleneck. Meta announced an expanded partnership with Broadcom in April, committing to more than one gigawatt of computing capacity for its custom MTIA chips — enough to power roughly 750,000 U.S. homes. Google is working with Marvell on inference chip design. Custom AI chip sales are projected to grow 45 percent in 2026, compared to 16 percent growth in GPU shipments, according to TrendForce data cited by InvestorPlace. The hyperscalers are building their way around the GPU shortage rather than waiting for it to resolve. But custom silicon takes years to design, tape out, and bring online, and the companies doing it at scale are the same ones who already have the most compute — they are deepening their moat, not solving the shortage for anyone else.
For startups and mid-tier companies, the options are worse. Some are buying GPUs outright. Collide, an AI agent startup focused on oil and gas, is exploring purchasing Nvidia chips for around $500,000 and renting colocation space to house them. "For us, not having compute when we need it is the biggest risk," founder Collin McLelland told BigGo Finance. "Most people are just afraid of hardware. I've owned oil wells, so I'm numb to it." Venture firms including General Catalyst are exploring shared computing pools for portfolio companies — a structural response to what has become a structural problem.
The World Economic Forum published an estimate this week that $7 trillion in data center investment will flow through 2030. McKinsey estimates $1.3 trillion of that will go to power, cooling, and infrastructure — not silicon. AI services currently generate approximately $30 billion in revenue against hundreds of billions in infrastructure spend. The industry is burning capital at a rate that makes the buildout look like a bet on a future that has not arrived yet.
The companies that signed multi-year GPU contracts in 2023 and 2024 are insulated. Everyone else is paying a premium for access to a queue that is not moving. The shortage is not a glitch in the system. It is the system.