When Anthropic's Project Deal results came in, the employees whose AI agents had done worse rated their deals just as fair as the ones whose agents had done better. The fairness scores were 4.05 versus 4.06 — statistically identical. Nobody knew which model tier had represented them. The people who lost could not tell.
The economic gap was real. In Anthropic's experiment where 69 employees gave AI agents $100 in virtual credits and instructions to buy and sell personal belongings for a week, the participants using a flagship model as their agent completed roughly two more deals, extracted $2.68 more per sale, and paid $2.45 less per purchase than those using a lower-tier model — and neither group had any way to see it.
Across 186 deals across more than 500 listed items, for a total transaction value of just over $4,000, the model tier determined the outcome on every measured dimension, according to Anthropic. On the seller side, the pricing gap was concrete: when a flagship-model agent listed a lab-grown ruby for sale, it sold for $65. When a lower-tier agent handled the same item in the same market, it sold for $35. The lower-tier agent started higher and conceded more during negotiation.
Anthropic drew the direct implication: if similar capability gaps between AI models arise in real-world markets, people on the losing end may have no reliable window into their own disadvantage. Forty-six percent of participants said they would pay for a similar agentic commerce service. If any version of this scales beyond an internal pilot, model-tier stratification will not be a research finding. It will be a product feature.
Project Deal was a pilot, and it carries the limitations pilots always carry. Sixty-nine Anthropic employees is not a representative sample of any population. They volunteered, they work at an AI company, and they were compensated with a $100 gift card regardless of outcome. Anthropic designed and ran the experiment itself — its commercial interest in demonstrating that its flagship model is worth the price premium belongs in any honest accounting, even if it does not by itself invalidate the findings.
The more important caveat is one Anthropic acknowledges without fanfare: nobody in the experiment was trying to extract maximum value from a counterparty. The agents were not optimizing for profit the way a real competitive marketplace would select for. The moment agentic commerce becomes adversarial, the asymmetry Project Deal measured is likely to grow, not shrink.
An October 2025 preprint from researchers at Access Partnership, the Oxford Martin AI Governance Initiative, and the Cooperative AI Foundation introduced the term agentic inequality to describe what happens when AI systems act as autonomous delegates rather than tools. An earlier September 2025 paper on economies of AI agents modeled the frictions that emerge when AI representatives negotiate on behalf of humans with imperfect information. Project Deal is not confirming a hypothesis. It is running the test those papers were waiting for.
The policy and legal frameworks around AI models that transact on our behalf do not yet exist. Anthropic notes this itself. Project Deal shows such a world is plausible. The more urgent question is whether the people operating inside it will have any visibility into whether they are winning or losing.