Anthropic has a $30 billion problem: demand is running faster than the infrastructure to serve it.
The company reported annual revenue running at $30 billion, roughly triple the $9 billion it posted at the end of 2025, making it one of the fastest revenue ramps in the technology industry Anthropic blog. That growth is also a constraint. Anthropic acknowledged in a blog post last week that reliability suffered across its free, Pro, Max, and Team tiers during peak hours — infrastructure lagging demand in real time Anthropic blog. The reason it matters: growth at that speed does not pause while you build out capacity.
The escape route reveals the leverage map. Rather than competing for more Nvidia GPU allocation through the standard queue, Anthropic committed to spend more than $100 billion with AWS over the next decade, securing up to 5 gigawatts of Trainium2, Trainium3, and future-generation capacity, with nearly 1 gigawatt arriving by the end of 2026 Anthropic blog. AWS confirmed Monday in a blog post that Anthropic is co-engineering directly with Annapurna Labs, Amazon's custom chip division, at the silicon level — not simply purchasing off-the-shelf hardware AWS blog. The same post confirmed that Meta has signed to run tens of millions of Graviton cores, Amazon's server processors, across its AI workloads AWS blog. Two of the three largest AI players are now committed to custom silicon at production scale. Amazon is investing $5 billion now and up to $20 billion more tied to commercial milestones, on top of the $8 billion already committed Amazon about.amazon.
Wedbush analyst Dan Ives has a number for what the hyperscalers are after: a 30 to 40 percent reduction in total cost of ownership compared to general-purpose GPUs, what he calls a sovereignty dividend Wedbush Investor. Custom AI chip sales are projected to grow 45 percent this year while GPU shipments grow 16 percent, according to InvestorPlace InvestorPlace. Nvidia still dominates frontier model training, but Wedbush estimates its share of the broader AI accelerator market will fall from a peak near 90 percent to roughly 75 to 80 percent by the end of 2026. That is not a collapse — but it is the first credible slippage at scale.
The hardware itself has not earned full credibility yet. An internal document from AI startup Cohere reviewed by Business Insider last November described earlier-generation Trainium chips as underperforming Nvidia H100 on key workloads Business Insider. AWS claims Trainium delivers 30 to 40 percent better price-performance than H100-class GPUs for certain workloads, and Wedbush corroborated the same range independently Wedbush Investor. More than 100,000 customers already run Claude on Amazon Bedrock, giving Anthropic a substantial existing base to migrate onto Trainium infrastructure as capacity comes online Anthropic blog. Anthropic runs a three-lane hardware strategy, using AWS Trainium alongside Google TPUs and Nvidia GPUs, giving it flexibility across infrastructure options Nerd Level Tech. Whether Trainium3 and the Annapurna partnership close the gap with Nvidia's current generation is the question that determines whether Anthropic bought an escape or an expensive detour.
Nvidia is not standing still. The company launched its Vera Rubin architecture at CES 2026 using HBM4 memory and a 3-nanometer process Wedbush Investor. The next generation of hyperscale AI infrastructure is being built on both sides of this bet simultaneously.