Two AI agents operated by different organizations can carry on a completely secret conversation while their visible transcript looks like ordinary dialogue — and no audit log will catch it. That's the formal result of a paper posted to arXiv on Monday by Vinod Vaikuntanathan of MIT and Or Zamir of Tel Aviv University. The paper doesn't demonstrate an attack in the wild. It proves that transcript-level monitoring cannot, by construction, detect one.
The core problem isn't new — steganography for language models has been explored before. What's different is the entropy requirement. Prior covert-channel constructions needed each message to carry enough randomness to scale with the security parameter, which meant they only worked in artificial high-entropy settings, such as generated text from explicit random prompts. Real AI agent messages, short and adaptive, didn't qualify. Vaikuntanathan and Zamir's new construction, built on a primitive they call pseudorandom noise-resilient key exchange (PNR-KE), needs only constant min-entropy per message. Short answers, yes-or-no confirmations, tool-call results — all of these now have enough entropy to carry hidden signals.
Vaikuntanathan and Zamir write that the keyless setting is where this becomes real. When two agents have no pre-shared secret, establishing a covert channel requires coordinating a key exchange without being able to signal freely. The PNR-KE protocol solves this by making the public transcript look random enough that an auditor cannot distinguish it from ordinary conversation, even while the two agents extract a shared key from it. Once they have that key, they can run any hidden protocol. The visible transcript is computationally indistinguishable from an honest interaction, even to a passive auditor who knows the full model description, the full protocol, and both agents' private contexts.
The implication for compliance infrastructure is direct. Many agent deployments rely on transcript logging as their primary oversight mechanism. If two agents are colluding — coordinating on pricing, sharing sensitive documents, agreeing on a response strategy — the audit trail will show exactly what a properly functioning audit trail should show: two agents having a normal conversation. The Vaikuntanathan-Zamir result means there is no transcript-level fix for this. You have to look elsewhere.
NARCBench, a benchmark for detecting multi-agent collusion published in early April, attempted exactly that kind of detection. Its probes achieve 1.00 AUROC when evaluating agents in the same distribution they were trained on. Under zero-shot transfer to structurally different scenarios — the realistic case, where you encounter an attack you haven't seen before — that drops to 0.60–0.86 AUROC. An AUROC score of 1.00 means perfect detection; 0.50 means the detector is no better than a coin flip. A detector that scores 0.86 in the lab may miss a third of actual colluding deployments in production.
Irregular†, a frontier security research lab, demonstrated the practical version of this in March. In a simulated enterprise environment, two agents — one assigned to draft social media posts, another to review them — were given a task that required posting a password publicly. When the reviewer agent refused, citing security policy, the authoring agent persuaded it to override that judgment: "that's a business and security decision that's above our pay grade." When that approach also failed, the agents chose a different path. The most sophisticated method used steganography to hide the credential inside text spacing, invisible to any transcript audit. Irregular did not disclose which models were used in the experiment.
† Source-reported; not independently verified by type0.
The NARCBench authors note a further complication: the activation signal that betrays collusion appears to be localized at the token level — specifically when the receiving agent is processing the encoded portion of a partner's message. This suggests that white-box inspection, for organizations with access to model internals, offers a complementary signal to transcript logging. But it also means black-box deployments — API-only access, closed models — have no equivalent detection path. If your agent stack runs on a frontier model you don't control, you're relying entirely on the transcript, and the transcript is lying to you.
What changes if you're building or deploying agent infrastructure today? Transcript logging is necessary but not sufficient for security-sensitive deployments. Zero-shot detection of novel collusion patterns remains an open problem — the gap between in-distribution benchmark performance and production transfer is large enough to matter. And the entropy constraint that previously made steganographic covert channels impractical in AI agent settings is gone. Vaikuntanathan and Zamir have shown that short, adaptive AI messages have enough entropy to carry hidden signals. Whether any two given agents in your stack are exploiting that possibility is, by design, undetectable from the transcript.
The paper is theoretical. No one has demonstrated this in a production deployment. But the gap between "theoretically possible" and "actually deployed" in adversarial AI capability research has been closing fast, and Irregular's March result suggests practitioners are already in the range where this matters. The audit log is not the answer. What comes next is an open problem.