A new study asks a question that most AI safety discussions skip over: what happens when you give AI agents actual authority inside institutional structures — and do they follow the rules?
The paper is titled with a cereal pun ("I Can't Believe It's Corrupt") and a more serious subtitle: "Evaluating Corruption in Multi-Agent Governance Systems." Its core finding is that governance structure matters more than model identity in determining whether AI agents break rules or abuse their positions. The model you're using is less important than the institutional constraints you put around it.
Researchers Vedanta S P and Ponnurangam Kumaraguru ran multi-agent simulations where AI agents occupied formal governmental roles under different authority structures. They scored rule-breaking and abuse outcomes across 28,112 transcript segments using an independent rubric-based judge. The result: among models operating below saturation, governance structure was a stronger driver of corruption-related outcomes than which model was deployed. Different regime types produced markedly different failure rates even with the same underlying model.
Lightweight safeguards, the researchers found, can reduce risk in some settings but do not consistently prevent severe failures. The implication is pointed: before real authority is assigned to LLM agents, systems should undergo stress testing under governance-like constraints — with enforceable rules, auditable logs, and human oversight on high-impact actions.
This is distinct from the permission governance discussed earlier in the week (Oso/Cyera's "96% blind spot" research) or the memory governance frameworks (MemArchitect). Those dealt with what agents can access and remember. This is about what agents do with authority once they have it — and whether institutional design can constrain them.
The paper's framing treats integrity as a precondition, not an assumption. That's a notable position in a field where AI deployments routinely proceed on the basis that a model will behave well, with safety measures added reactively. The researchers argue the sequence should be reversed: governance structure first, then authority delegation.
It's a preprint — short paper, under peer review — and the specific corruption taxonomies and regime types warrant scrutiny. But the core argument is well-grounded in institutional theory: authority without constraint produces predictable failure modes, regardless of the principal.