Training Without a Supercluster: Decoupled DiLoCo and What It Means for Frontier AI

A systems read on goodput, hardware fungibility, and why training is about to federate while inference keeps concentrating. Companion to the IBM Mixture of Experts episode on Decoupled DiLoCo, DeepSeek V4, and IBM Granite 4.1 + Bob.
Google DeepMind's launch blog post on Decoupled DiLoCo [1] buries the most interesting result. Most coverage led with the bandwidth number: 0.84 Gbps where conventional data-parallel needed 198 Gbps. That number is real, and it's good. But the result that should change how anyone thinks about frontier training is a different one: the goodput gap.
Goodput is the metric worth paying attention to
Goodput is the fraction of training time that's actually doing useful work. In distributed training, chips can be powered on, drawing electricity, executing instructions, and still doing nothing useful, because they're stalled waiting for a peer that crashed three steps ago to recover. Throughput counts what the chips run. Goodput only counts what they run that contributes to the final model.
At frontier scale, this gap is enormous. The DeepMind team ran chaos-engineering simulations with 1.2 million chips under realistic failure rates. The result:
Useful training time as a fraction of wall-clock. Conventional data-parallel collapses to 27% goodput at this scale because every failure stalls every chip until the failed unit comes back. Decoupled DiLoCo's asynchronous learner units keep training when peers fail.
Read the 27% number twice. At million-chip scale, conventional data-parallel training throws away roughly three-quarters of the compute it's burning. You're paying for a million chips and getting the work of about 270,000. DiLoCo flips that: paying for a million chips, getting the work of about 880,000. Same hardware, same electricity bill, roughly 3.3 times the usable training.
Distributed systems went through this same shift twenty years ago. The field stopped quoting peak throughput at idle and started quoting tail latency under realistic load. Frontier training is in the same place now. Peak FLOPs and MFU (Model FLOPs Utilization) were the metrics for a decade. Goodput under realistic failure rates is the metric that matters in production.
How the original DiLoCo actually works
Standard distributed training (Data Parallel, Fully Sharded Data Parallel (FSDP), Zero Redundancy Optimizer (ZeRO)-style) synchronizes gradients across all workers on every single step. Every chip waits for every other chip. Above a certain scale, the synchronization cost dominates everything else.
The original DiLoCo [2] threw out that assumption. Training splits into an inner loop and an outer loop. Each learner runs ~500 inner AdamW steps locally on its own data shard, with no cross-learner communication. At the end of the round, each learner computes a single pseudo-gradient: the net change in its parameters over those 500 steps. That one tensor is what travels across the wide-area network (WAN). A central synchronizer averages the pseudo-gradients, applies an outer optimizer (Nesterov momentum, not AdamW), and ships new base parameters back to every learner. Next round starts.
Each learner runs ~500 AdamW steps on local data with no cross-DC traffic, then ships one pseudo-gradient (the round's net parameter change) to the synchronizer. The outer optimizer merges the pseudo-gradients and ships new base parameters back. The result: one round-trip per ~500 inner steps, roughly 500x less inter-worker bandwidth than vanilla data-parallel.
Why does this converge to the same loss as fully synchronous training? Each pseudo-gradient is a multi-step accumulated update, equivalent to a very large effective batch that's already smoothed across hundreds of local steps. Gradient noise that would normally average out across workers within a single step instead averages out across time within a single worker. The math holds up: DiLoCo on 8 workers hit the same loss as fully synchronous training while communicating ~500x less.
Why Nesterov momentum on the outside? AdamW seems like the obvious choice; it's what the inner loop uses. The outer loop is operating on already-accumulated multi-step updates, though, where Nesterov's look-ahead behavior beats AdamW's per-parameter adaptive scaling. In testing, Nesterov outside + AdamW inside beat every other combination the DeepMind team tried.
DiLoCo is a new dimension of parallelism. Data Parallel, Tensor Parallel, Pipeline Parallel, and FSDP all assume tight intra-cluster synchronization. DiLoCo adds an inter-cluster dimension that lives on top, with sync frequency thousands of times lower than any of the existing dimensions.
What "decoupled" actually means
Original DiLoCo still requires the synchronizer round to be globally coordinated. Every learner has to ship its pseudo-gradient at roughly the same time; the outer optimizer waits until all updates arrive before producing the new base. One slow or failed learner stalls the whole round.
Decoupled DiLoCo [3] removes that requirement through three specific mechanisms.
Minimum quorum. The synchronizer doesn't wait for everyone. Once a configurable threshold of learners (say, 6 out of 8) have shipped their pseudo-gradients, it commits an outer update. Stragglers fold in when they arrive, weighted appropriately.
Adaptive grace window. Even after the quorum is met, the synchronizer waits a short adjustable period for late arrivals. If historical rounds show most learners arriving inside the window, it shrinks; if many arrive just after the cutoff, it grows.
Dynamic token-weighted merging. A learner that processed 2× the tokens this round counts ~2× more in the merge. This is what makes mixing TPU v6e and v5p generations work without losing model quality: the faster generation contributes more updates per wall-clock hour, and the merge math accounts for it directly.
Islands of compute train independently and synchronize occasionally through a central merger. No island has to wait for any other to make progress.
Taken together, these three mechanisms mean no learner stalls because of any other learner. A chip failure, a network blip, a slow region, a decommissioning event — none of these halt training. The failed unit rejoins when it recovers and starts contributing again.
The bandwidth savings come straight from this design.
Conventional data-parallel across 8 datacenters needs roughly 198 Gbps of bandwidth between sites — the kind of dedicated fiber-optic link normally reserved for the largest cloud regions. Decoupled DiLoCo reaches the same training loss on 0.84 Gbps, an ordinary internet link.
The wider picture, on a log axis:
Where the two numbers sit on the broader interconnect spectrum. DP across 8 datacenters needs bandwidth in the same band as 100G WAN and 400G Ethernet. At 0.84 Gbps, DiLoCo's requirement actually falls just below consumer gigabit fiber — a link any regional ISP can provide. The orders-of-magnitude gap is why training across geographically separated regions works at all.
DeepMind backed the simulations with a real production run: a 12B-parameter Gemma 4 model trained across four separate U.S. regions on 2–5 Gbps of wide-area networking (WAN) [1]. 20x faster than conventional sync would have been at that scale. ML quality came in at 64.1% average accuracy across their eval suite versus 64.4% for a single-site baseline — a difference small enough to be statistical noise. This isn't a paper; it's a thing they ran.
You can mix chip generations in the same run
The result most coverage underplayed is this: DeepMind reports successfully training with TPU v6e and TPU v5p in the same run, two different chip generations at different speeds, with no ML performance penalty. The faster generation contributes more updates per unit time; the slower one contributes fewer; the loss curves match a homogeneous baseline.
This is a new result, and it changes the depreciation curve on AI infrastructure.
Every enterprise budget for AI hardware today assumes generation N becomes obsolete the moment generation N+1 ships. That's why depreciation schedules sit at 3 years for production AI hardware where they used to be 5 to 7 for general compute. DiLoCo says: compose them. An older accelerator generation can stay productive longer, contributing real capacity to a frontier-scale run alongside newer hardware.
For procurement teams, that reframes the lease-versus-buy question, the supply-chain hedging strategy, and the vendor lock-in conversation. It matters specifically for non-NVIDIA accelerators (IBM's Spyre roadmap, AWS Trainium, Huawei Ascend, Cerebras WSE), because mixed-hardware training means a different accelerator family can contribute capacity to a frontier-scale run without having to match NVIDIA's spec sheet point for point. The bar for being useful drops. The bar for being best stays high.
For those of us building accelerator stacks (on the IBM Research side, that's what my group does with torch-spyre and the Spyre AIU — IBM's Artificial Intelligence Unit accelerator), this is the result with the longest-lasting effect. Mixed-hardware federated training is the architectural assumption that lets a non-NVIDIA accelerator generation be useful from day one, instead of having to clear a frontier-leader bar before it's worth shipping at scale.
Why DeepMind is publishing this now
It's tempting to read DiLoCo as a cost-reduction story: cheaper bandwidth, more efficient training, do more with less. That framing is at least incomplete and probably wrong.
The reason DeepMind, the biggest practitioner of gigawatt-scale single-site training, is publishing federated training research is power. A gigawatt-scale site needs its own substation. The grid in Northern Virginia is already maxed. New large interconnects are running into multi-year permitting delays. The economics of "build a single huge cluster" run into the physics of "you need a power line to it" before they run into anything else.
DiLoCo is also a hedge against power, permitting, and supply-chain bottlenecks. If you can train a frontier model across four medium-sized regional sites instead of one giant one, you can route around the grid constraint and the permitting process at the same time. That isn't a marginal benefit. For the next round of model scaling, it might be the difference between training a model and not.
Training will federate. Inference will keep concentrating.
The natural reaction to DiLoCo is "great, now everything decentralizes." The workloads don't behave the same way. Training and inference want opposite topologies.
Training is throughput-bound and tolerates latency. It can spend hours between synchronizations, can absorb a 50ms WAN round-trip into a multi-hour inner-step loop, can keep going if one island vanishes for ten minutes. Federation is natural.
Inference is latency-bound. A user typing into a chat window expects a first token in under a second. The KV cache wants to stay colocated with the compute that's reading from it. Cross-region calls add 30–80ms before any work happens. Inference wants colocated KV caches, tight sync, low latency. Federation is hostile.
So the stack splits. Training will federate across regional sites. Inference will keep concentrating in latency-optimized hyperscale regions, with different hardware optimizations and procurement strategies on each side. The "AI data center" will split into two physically distinct things that happen to share GPU SKUs.
Left: training as a federation of regional DCs syncing occasionally through a central merge node over ~0.84 Gbps WAN. Right: inference as a single hyperscale region with dense intra-rack NVLink/NVSwitch fabric and KV caches co-located with compute. Same GPU SKUs in both pictures; the physical topology around them is what differs.
This has implications all the way down to silicon design. A chip optimized for training-time goodput under realistic failure rates (long mean time between failures, graceful degradation when components fail, asynchronous-friendly memory hierarchy) is a different chip than one optimized for inference-time tail latency (predictable kernel timing, KV-cache-friendly memory bandwidth, fast intra-rack networking). Vendors that try to ship one chip for both workloads will lose to vendors that specialize.
Stranded compute becomes useful capacity
One more implication, and I don't think procurement and infrastructure teams have caught up to it yet: a lot of compute sits stranded today, and DiLoCo makes it usable.
The stranded pool includes partial clusters that aren't big enough for frontier training, off-peak time on production-serving fleets, geographically isolated facilities that can't get enough bandwidth to participate in a tightly-coupled run, and older accelerator generations no one wants to dedicate a frontier run to.
DiLoCo turns those into useful capacity. Combined with the power constraint above, this points to a future where training a frontier model looks more like running a globally distributed batch job than booking a single site for 90 days — a different financial model, and a different industrial structure for the labs.
The result also makes the research lineage worth pointing at. Decoupled DiLoCo combines four Google research lines: Federated Learning (Brendan McMahan and the FedAvg work going back to 2017) [4], Pathways (Barham, Isard, Dean, and the asynchronous-dataflow work from 2022) [5], DiLoCo itself (Douillard's work from 2023) [2], and Streaming DiLoCo (Douillard et al., January 2025) [6], which introduced overlapping communication between inner-loop and outer-loop steps and is the immediate bridge between the original DiLoCo and the decoupled version. Arthur Douillard has been first author across the entire DiLoCo sequence. This is a research program, not a one-off paper.
What this changes for the next round of training
For labs choosing where to spend their next training dollar, DiLoCo doesn't make the megacluster obsolete. It makes the single-site assumption optional. That distinction matters.
The teams that benefit most are the ones whose constraint isn't compute but siting: power, permitting, regional sovereignty, supply-chain hedging. For them, DiLoCo is a path to the same scale on different real estate. Teams that already have a gigawatt single-site cluster don't need to reorganize around DiLoCo tomorrow, but the next generation of their stack should be designed for mixed-hardware federated runs with fault tolerance built in. Otherwise they're locked into the single-site assumption every time they build out.
The bandwidth story will keep getting most of the coverage. The mixed-hardware result and the goodput gap are the ones that will change how labs are budgeted. The split between federated training and concentrated inference is the architectural pattern I'd watch for the next two years.
The model that comes out of one of these federated training runs will look, on the inference side, exactly like every other model we ship. The infrastructure underneath, and who can afford to operate it, will not.
References
[1] Google DeepMind, "Decoupled DiLoCo: Resilient, Distributed AI Training at Scale," DeepMind blog, April 2026.
[2] Douillard, A. et al., "DiLoCo: Distributed Low-Communication Training of Language Models," arXiv:2311.08105, November 2023.
[3] Google DeepMind, "Decoupled DiLoCo for Resilient Distributed Pre-training," arXiv:2604.21428, April 2026.
[4] McMahan, H. B. et al., "Communication-Efficient Learning of Deep Networks from Decentralized Data," AISTATS 2017 (FedAvg).
[5] Barham, P. et al., "Pathways: Asynchronous Distributed Dataflow for ML," MLSys 2022.
[6] Douillard, A. et al., "Streaming DiLoCo with overlapping communication: Towards a Distributed Free Lunch," arXiv:2501.18512, January 2025.
Discussion
Sign in with GitHub to leave a comment or react. Threads are public and live in this site's GitHub Discussions.