Anthropic built an AI that finds software vulnerabilities — and three weeks after hackers used credentials from a prior breach to access it, none of the eleven enterprise partners have confirmed whether they ever paid for the product.
Mythos launched with $100 million in committed usage credits and a post-credit price of $25 per million input tokens and $125 per million output tokens, according to Anthropic's Glasswing page. Eleven partners signed on: Amazon Web Services, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorganChase, the Linux Foundation, Microsoft, Nvidia, and Palo Alto Networks. Whether any partner has crossed from free evaluation to paid usage is undisclosed by Anthropic. The distinction matters: if the companies most exposed to a model designed to find vulnerabilities in critical infrastructure are still in the free trial phase, the urgency of the Glasswing pitch runs ahead of actual commercial commitment.
On March 24, an attacker published two malicious versions of LiteLLM to the Python Package Index, the open-source repository where developers download software libraries. The packages were quarantined within 40 minutes, according to Proofpoint's analysis published April 20, but the credential exposure to Mercor, an AI hiring platform, had already occurred. On April 7 — the same day Anthropic announced Mythos — a Discord group used knowledge of Anthropic model formats from that breach to locate and access the system, The Verge, TechCrunch, and The Guardian reported. Anthropic confirmed the unauthorized access on April 21, saying it had no evidence the incident extended beyond the third-party vendor environment, TechCrunch reported. The group provided Bloomberg screenshots and a live demonstration, describing their activity as benign exploration.
The irony is precise: the model Anthropic built to catch supply chain weaknesses was itself accessed via a supply chain weakness.
Mythos is designed to find zero-day vulnerabilities, previously unknown security flaws that can be exploited before developers patch them. The model had identified thousands across critical infrastructure before this incident, Anthropic said. It completed a 32-step cyberattack simulation created by the UK's AI Safety Institute in three out of 10 attempts — tasks that would normally take human professionals days, The Guardian reported. AISI rated it capable of multi-step attacks and autonomous vulnerability discovery without human intervention, the first AI model to achieve that rating. UK AI minister Kanishka Narayan said British businesses should be concerned about the model's ability to identify flaws in IT systems, The Guardian reported. The UK government relies on Anthropic for access.
Anthropic declined to comment beyond its published statement.
What comes next is a set of open questions. Whether any Glasswing partner has crossed from free evaluation to paid usage — Anthropic has not said. How regulators respond to a model with demonstrated autonomous vulnerability discovery capabilities — that is unresolved. And what happens the next time a supply chain attack exposes credentials to a system like this one.