Most developers building multi-agent AI systems today assume their agents should communicate constantly — sharing updates, confirming progress, looping each other in. A paper published Tuesday on arXiv argues that assumption is backwards. Tested in Minecraft, the researchers found that agents performing a complex, multi-step building task completed it faster and more reliably when they talked to each other less, not more. The reason: constant communication interrupts agents mid-execution, floods the shared state with redundant updates, and triggers false coordination responses — agents react to noise that isn't there.
The paper introduces a design called gated coordination, which treats communication as a metered resource rather than a reflex. Rather than broadcasting every update to every agent in the system, a gated coordinator decides whether a given message is worth the cost of disruption it causes. The mechanism weighs three factors: how critical the sending agent is to the task, whether the local agent can recover on its own, and what downstream impact the message would have. Agents that can handle a situation locally do so. Only genuinely cross-agent dependencies trigger a handoff.
The three costs the paper identifies are concrete. Constant updates interrupt a working agent mid-step — the equivalent of asking someone to stop and file a status report in the middle of writing a paragraph. Redundant messages pile into a shared state that all agents read from, making the global view progressively noisier. And because agents respond to messages, they end up reacting to coordination triggers that were themselves false alarms — a cascade of responses to a problem that doesn't exist.
The paper, submitted to arXiv on Tuesday by researchers HuaDong Jian, Chenghao Li, Haoyu Wang, Jiajia Shuai, Jinyu Guo, Yang Yang, and Chaoning Zhang, frames this as a fundamental architectural problem rather than a model-tuning issue. The key finding: once you control for compute, the performance advantage of multi-agent systems over single-agent systems shrinks as base models get stronger. At frontier capability levels, the gains from multiple agents appear to come from the ability to run more computation in parallel — not from better coordination. Coordination overhead, as a separate arXiv paper on multi-agent LLM scaling puts it, may be the binding constraint on what multi-agent systems can actually deliver.
Teams often choose coordination patterns based on what sounds sophisticated rather than what fits the problem at hand, the researchers note. The implication is concrete: a multi-agent pipeline that wires every agent to every other agent is not obviously better than a simpler design. The coordination cost may exceed the task value of the information being shared.
What the paper does not yet show is whether these results hold outside Minecraft, whether they generalize to frontier-grade models, or whether anyone has shipped anything production-grade using a gated approach. All three are open questions — and the answer to each will determine whether this is a useful constraint on how frameworks are designed or a marginal result that disappears at scale.