For three years, the AI industry has been building around a central assumption: the GPU is the constraint. More GPU compute, more memory bandwidth, more high-bandwidth memory chips stacked tightly together, and the model runs faster. That assumption has governed hundreds of billions of dollars in infrastructure spending. It might be wrong, at least for the next phase of AI.
The evidence comes from an unlikely place: a Georgia Tech and Intel research paper that has become the quiet reference point for a growing camp inside the semiconductor industry. The paper, titled "A CPU-Centric Perspective on Agentic AI," profiled five representative agentic workloads and found that CPU tool processing, the Python interpretation, database queries, web crawls, and lexical operations that agents run between GPU inference cycles, accounts for up to 90.6% of total end-to-end latency, per the paper. The GPU sits idle while the CPU handles the work the model cannot do itself. In a Semiconductor Engineering interview, Jeff Defilippi, senior director of product management at Arm, described the contrast with traditional queries this way: queries to search for and analyze data from multiple sources will be performed simultaneously by agents and without human intervention, rather than a single request from a live person per the interview. The CPU, not the GPU, is the thing keeping track of all of it.
Arm is not the only company that has reached this conclusion. In March 2026, two product launches arrived in the same month that would seem unrelated to anyone watching the GPU wars: Arm announced its AGI CPU, a 136-core chip built on TSMC's N3 process with a 300-watt thermal envelope, designed specifically for agentic orchestration workloads, Arm reported. Nvidia announced Vera CPU, an 88-core chip built on its Olympus architecture, also targeting the orchestration layer between GPU clusters and external tool chains, per TrendForce. That two companies whose primary businesses are defined by opposite ends of the AI compute stack both moved into the same narrow slice of silicon at the same time is not a coincidence. It is a market signal.
The structural shift is this. In the training era that defined AI infrastructure from 2022 to 2025, the CPU-to-GPU ratio in data centers settled around one CPU for every four to eight GPUs. The GPU was the expensive, limiting factor; the CPU was the cheap glue. Agentic AI inverts that. When an agent is running, the GPU executes the model inference, fast, parallel, predictable. But the agent also needs to query a vector database, run a Python snippet, fetch a web page, call an external API, and route the results back to a parent agent that is coordinating dozens of parallel sub-tasks. Each of those operations runs on a CPU. The aggregate latency of all that CPU work, the Georgia Tech paper found, can dominate total task time even when the GPU portion is relatively small. TrendForce estimates that CPU core density per gigawatt of data center power will need to rise from roughly 30 million cores today to 120 million cores in the agentic era, a fourfold increase that makes the CPU infrastructure investment comparable to the GPU investment it was previously designed to complement, not rival, per TrendForce's analysis.
Arm's position in this shift is unusual. For 35 years, Arm made money by licensing its processor architecture to other chip companies. AWS, Google, Microsoft, and Nvidia all build Arm-based silicon and pay Arm royalties whether it sells directly or not. The AGI CPU is the first time Arm has shipped its own finished silicon, according to Arm's announcement. The dual effect is that Arm now collects both the royalty from every licensee and the margin from its own chip sales, a compounding revenue model that analyst firm Futurum projects could drive Arm to $15 billion in annual revenue by fiscal 2031, up from roughly $4 billion today, against a target data center CPU market estimated at $76.6 billion by 2029 growing at 34.9% annually, Futurum reported. Arm is positioned to capture value whether its licensees win or its own product wins. That is a structural advantage no other chip company has in exactly this moment.
Meta is the lead customer and co-development partner for the AGI CPU, working with Arm to optimize infrastructure for its family of applications and its custom MTIA accelerators, per Arm's announcement. Other launch partners include Cerebras, Cloudflare, OpenAI, Positron, and SAP. The list is notable not for the names but for the breadth: it spans cloud networking, AI inference hardware, enterprise software, and frontier model training. All of them have the same problem.
The skeptical case is worth stating plainly. The 90.6% latency figure comes from profiling five workloads under controlled conditions, Haystack RAG, Toolformer, ChemCrow, Langchain, and SWE-Agent, not from production-scale deployments with hundreds of concurrent agents in a real data center, per the Georgia Tech paper. The CPU bottleneck may look different when the tool calls are hitting cached data rather than cold storage, when the orchestration layer is better optimized, or when the agent framework has been heavily engineered to minimize round-trips. Arm has a product to sell, and the Georgia Tech paper supports its sale. The number is real; the generalization to all agentic workloads is not yet proven.
What is proven is that the chipmakers have reached the same conclusion independently. Nvidia and Arm looked at the same workload profile and built CPUs for it in the same quarter. The GPU-centric model that governed infrastructure planning through the generative AI era is not wrong for training. It may be incomplete for deployment. And the companies that make the chips are the first to know, because they are the ones the hyperscalers call when the architecture assumptions change.