The Quiet Case for Breaking the AI Compute Moat
DeepMind published a training method that works across distributed data centers at internet bandwidth. The question is whether it changes who can afford to build frontier AI.

Training a frontier AI model has always required a single, tightly synchronized machine room. The chips inside need to talk to each other constantly, which means custom network wiring, campus-scale infrastructure, and tolerances measured in microseconds. Google DeepMind published a method this week that says that assumption is negotiable.
The approach, called Decoupled DiLoCo, trains large language models across geographically separated data centers using asynchronous communication between them. The bandwidth required: 2 to 5 gigabits per second, which is ordinary internet connectivity, not custom fiber. DeepMind researchers trained a 12 billion parameter Gemma 4 model across four separate U.S. regions at that bandwidth and achieved the same benchmark performance as conventional tightly synchronized training, according to the DeepMind blog post. The method was more than 20 times faster than conventional distributed approaches, because it avoids the blocking bottleneck where one part of a synchronized system idles while waiting for another to finish its step.
The more striking property is resilience. When DeepMind introduced artificial hardware failures during training runs using chaos engineering, the system continued learning after losing entire learner units and reintegrated them seamlessly when they came back online. No restart required. No checkpoint rewound. This is not a demo artifact: the blog shows it operating on real hardware, not a simulation.
Decoupled DiLoCo also mixed different chip generations in the same training run. TPU v6e and TPU v5p chips, running at different speeds, matched the performance of single-generation clusters. If this holds at scale, it means hardware obsolescence is not a binary event. Older chips do not become worthless when a new generation ships; they become part of a heterogeneous pool that still produces useful work.
Independent validation exists. Prime Intellect, an AI development platform, has been training models through globally distributed reinforcement learning since late 2024. Its INTELLECT-1 model, a 10 billion parameter model trained across distributed hardware, launched in November 2024. A 32 billion parameter version followed in May 2025, and a 100 billion-plus mixture-of-experts model shipped in November 2025, according to the Prime Intellect blog. Prime Intellect built an open-source implementation called OpenDiLoCo in July 2024, making the approach reproducible before DeepMind's own publication. IEEE Spectrum covered the distributed training landscape independently, situating DiLoCo within a broader industry shift toward decentralized compute, in a recent article.
The implication is straightforward: if training can work across ordinary internet connections and recover from hardware failures automatically, the economics of building a frontier AI model change. Idle GPU capacity scattered across regional data centers becomes a candidate for pooling. Stranded compute, hardware sitting underused in research labs or smaller cloud providers, becomes usable rather than written off. The hyperscaler campus, purpose-built and tightly engineered, is not the only viable path.
The caveats are real. The 20 times speedup figure comes from a specific four-region setup with a 12 billion parameter model. One of the two charts in DeepMind's blog post is based on simulated training runs, not physical experiments. Whether the same performance holds at true frontier scale, with models of hundreds of billions or trillions of parameters, is unproven. The mixed-generation result is specific to Google's TPU architecture; it is not clear the same pooling works across Nvidia GPU generations, which constitute most of the external compute market. And capital requirements, data access, and talent remain substantial even if the hardware problem becomes easier.
The power-shift question is therefore not settled. Decoupled DiLoCo is a genuine technical advance on a specific, hard problem. Whether it lowers the barrier to frontier AI training enough to matter, or whether it remains a Google-specific optimization that makes Google's infrastructure more efficient without changing who can compete, is the story worth watching.





