OpenAI Has Run Three Bio Bug Bounties. It Has Disclosed Nothing.
OpenAI launched a new bug bounty program Thursday offering researchers $25,000 to find a universal jailbreak capable of defeating all five bio-safety guardrails on GPT-5.5 — the latest iteration of its most capable model. The offer is time-limited, the scope is narrow, and the fine print matters: all findings stay under NDA. So do the results of the last two times OpenAI tried this.
The company ran bio-safety bug bounties for its GPT-5 model in September 2025 and for its Agent platform in July 2025. Neither program produced a public disclosure. No winners were announced. No payouts were confirmed. No findings — patched or otherwise — reached the outside world. OpenAI declined to confirm whether either program received valid submissions or paid out any money.
"We want researchers to find gaps in our safeguards so we can fix them before release," the company wrote in its Thursday announcement. The company did not answer questions about what it did with the results from the prior two programs.
The new program runs April 23 through July 27, 2026, with applications closing June 22. Participants must apply in advance and access GPT-5.5 exclusively through Codex Desktop. The NDA covers all prompts, completions, findings, and communications — participants cannot publicly disclose what they discover, even if OpenAI patches nothing, according to the program's terms.
Security researchers who commented on the program online were skeptical. "With only $25k in payouts and everything locked down under NDA, I cannot imagine many people will participate," one wrote on Hacker News. "Well, other than those submitting mountains of LLM-generated junk." Others noted that $25,000 is cheap for a frontier AI lab with billions in annual revenue, and that serious red teams at national laboratories or academic biosecurity centers are unlikely to sign agreements that prohibit them from publishing their findings.
OpenAI is not the only lab running bio-safety bug bounties. Anthropic and Google DeepMind have run similar programs, though neither has published post-program results publicly either. The pattern is not unique to OpenAI — it reflects a broader industry practice of closed-door adversarial testing that critics say provides no external accountability.
The bio-safety questions themselves are not public. OpenAI describes them as challenges designed to test whether GPT-5.5 will provide harmful information about biological or chemical weapons synthesis. The five questions are the wall researchers are asked to try to defeat. If no one defeats them, OpenAI will not say so. If someone does, OpenAI will not say that either. The $25,000 prize is not a transparency mechanism — it is a finding-purchasing agreement with built-in silence clauses.
This is the accountability structure OpenAI has offered the public for three consecutive frontier model releases: run an expensive private audit, control what the auditors can say, disclose nothing regardless of what is found.
Whether that constitutes safety testing or safety theater depends on what happens inside the room. Without a disclosure, there is no way to know.
OpenAI declined to comment.