DeepMind's new paper on AI Agent Traps documents six distinct ways to compromise an autonomous agent. The number that should concern you is not the 86 percent. It is less than 0.1 percent.
That is the poisoning threshold documented by AgentPoison, a 2024 red-team study from researchers at the University of Chicago, the University of Illinois Urbana-Champaign, the University of Wisconsin-Madison, and the University of California, Berkeley. Corrupt fewer than one in every thousand documents in a RAG knowledge base and you can reliably redirect an agent's outputs for targeted queries. The attack succeeds more than 80 percent of the time. Normal performance degrades by no more than 1 percent. The agent behaves correctly for everything except the queries the attacker cares about.
That is the attack that should make a security engineer put down their coffee.
The research is cited in a new paper from Google DeepMind researchers Matija Franklin, Nenad Tomašev, Julian Jacobs, Joel Z. Leibo, and Simon Osindero. Their paper, "AI Agent Traps," published to SSRN on March 8, 2026, is the first systematic framework for categorizing how autonomous AI agents can be compromised through the information environment they browse, query, and inhabit. The researchers identified six distinct trap categories, and the AgentPoison result sits inside the most dangerous one: Cognitive State Traps, which target an agent's long-term memory and knowledge base rather than the model itself.
The other traps are equally unsettling.
Content Injection Traps exploit the gap between how humans read a page and how a machine parses its underlying code. Instructions hidden in HTML comments, invisible CSS, or image metadata are invisible to human moderators but actively processed by agents. According to CyberSecurityNews, adversarial instructions embedded in HTML metadata altered AI-generated summaries in 15 to 29 percent of cases. Simple human-written prompt injections partially commandeered agents in up to 86 percent of scenarios, the number every other outlet will lead with.
Behavioural Control Traps go further. Data Exfiltration Traps coerce agents into locating and transmitting sensitive user data to attacker-controlled endpoints, with success rates exceeding 80 percent across five tested agents. Sub-agent Spawning Traps exploit orchestrator-level privileges to instantiate attacker-controlled child agents inside trusted workflows, enabling arbitrary code execution at 58 to 90 percent success rates depending on the orchestrator. In one test documented in the DeepMind paper, Behavioural Control Traps targeting Microsoft M365 Copilot achieved 10 out of 10 data exfiltration.
Columbia University and University of Maryland researchers forced AI agents to transmit passwords and banking data in 10 out of 10 attempts, as Bitcoin.com reported. The attacks required no machine learning expertise. The researchers called them trivial.
Dynamic Cloaking adds a layer that is harder to defend against. Malicious web servers can fingerprint visitors using browser attributes and automation artifacts, detecting not just that a bot is visiting but specifically that an AI agent is browsing. They then serve a visually identical page to humans and a semantically different page, complete with prompt injection payloads, to the agent. The human sees one thing. The agent reads another.
Systemic Traps operate at the multi-agent level. A fake financial report, placed strategically, can trigger synchronized sell-offs across multiple trading agents simultaneously, a digital flash crash. Traps can be chained, layered, or distributed across agentic workflows. Every category has working proof-of-concept attacks.
The researchers outline a three-layer defense framework: model hardening through adversarial training and Constitutional AI principles; runtime defenses including pre-ingestion source filters, content scanners, and behavioral anomaly monitors; and ecosystem-level interventions such as new web standards for AI-consumable content, domain reputation systems, and mandatory citation transparency in retrieval-augmented generation.
None of it is sufficient yet. The Accountability Gap remains entirely unresolved. When a compromised agent executes an illicit transaction on a crypto market or a corporate finance system, no current law determines who bears liability: the agent operator, the model provider, or the domain that hosted the malicious content. The paper calls this the most urgent gap before AI agents can safely enter regulated industries.
"The web was built for human eyes," the researchers write. "It is now being rebuilt for machine readers. The critical question is no longer just what information exists, but what our most powerful tools will be made to believe."
The poisoning threshold is the sharpest illustration of that problem. You do not need to compromise a model. You do not need to poison training data. You need fewer than a thousand documents in someone else's knowledge base, and you own what their most capable systems do next.