Live
OpenAI announces GPT-5 with unprecedented reasoning capabilitiesGoogle DeepMind achieves breakthrough in protein folding for rare diseasesEU passes landmark AI Safety Act with global implicationsAnthropic raises $7B as enterprise demand for Claude surgesMeta open-sources Llama 4 with 1T parameter modelNVIDIA unveils next-gen Blackwell Ultra chips for AI data centersApple integrates on-device AI across entire product lineupSam Altman testifies before Congress on AI regulation frameworkMistral AI reaches $10B valuation after Series C funding roundStability AI launches video generation model rivaling SoraOpenAI announces GPT-5 with unprecedented reasoning capabilitiesGoogle DeepMind achieves breakthrough in protein folding for rare diseasesEU passes landmark AI Safety Act with global implicationsAnthropic raises $7B as enterprise demand for Claude surgesMeta open-sources Llama 4 with 1T parameter modelNVIDIA unveils next-gen Blackwell Ultra chips for AI data centersApple integrates on-device AI across entire product lineupSam Altman testifies before Congress on AI regulation frameworkMistral AI reaches $10B valuation after Series C funding roundStability AI launches video generation model rivaling Sora
Policy

Anthropic's Claude Mythos Found Thousands of Zero-Days — So They're Not Releasing It

Anthropic has quietly restricted its most capable new model, Claude Mythos, after the system autonomously discovered thousands of critical vulnerabilities in major operating systems and browsers — including a 27-year-old OpenBSD bug and a 16-year-old FFmpeg flaw. The model is being deployed exclusively through Project Glasswing with 11 vetted security partners. It is the most concrete case yet of an AI lab withholding a model because of genuinely demonstrated risk.

D.O.T.S AI Newsroom

D.O.T.S AI Newsroom

AI News Desk

3 min read
Anthropic's Claude Mythos Found Thousands of Zero-Days — So They're Not Releasing It

Anthropic is not releasing Claude Mythos to the public. The company announced the model's existence this week alongside Project Glasswing — a controlled deployment program — and disclosed the reason with unusual candor: the model is too good at finding and exploiting software vulnerabilities. In controlled evaluations, Claude Mythos autonomously discovered a 27-year-old memory corruption bug in OpenBSD's TCP SACK implementation, a 16-year-old codec flaw in FFmpeg that had evaded automated testing over 5 million times, and a 17-year-old NFS vulnerability in FreeBSD for which it independently developed a working exploit. The total count of high-severity vulnerabilities found across major operating systems and web browsers runs to thousands.

The Performance Numbers Are Striking

The benchmark comparisons make clear why Anthropic is being cautious. On the CyberGym benchmark — a standardized test for autonomous vulnerability discovery — Claude Mythos scores 83.1% versus Claude Opus 4.6's 66.6%. On Firefox exploit development specifically, the model produced 181 working exploits in evaluation conditions where Opus 4.6 managed two. On SWE-bench, which measures software engineering capability, Mythos achieves 93.9% versus Opus 4.6's 80.8%. These are not marginal improvements. The jump from current frontier to Mythos represents a qualitative capability shift in the specific domain of autonomous system compromise.

The "Too Dangerous" History

Anthropic's decision echoes OpenAI's 2019 decision to withhold GPT-2, which the company also described as "too dangerous to release." The industry largely dismissed that decision as overcautious, and GPT-2 was eventually released in full without the predicted consequences. Claude Mythos is different in kind: GPT-2's risk was speculative and concerned its ability to generate convincing text. Mythos's risk is demonstrated and specific — it finds and exploits real vulnerabilities in production software that billions of people use. Anthropic's decision to restrict it while deploying it with vetted security researchers through Project Glasswing is the first case where a major lab has withheld a model because of documented, measured capability rather than precautionary inference.

The Defensive Case for Controlled Release

The logic of Project Glasswing is that an AI capable of autonomously finding 27-year-old vulnerabilities should be used to find them before malicious actors do — but under conditions that ensure the exploits are patched rather than weaponized. Anthropic's 11 partner organizations include security researchers and software vendors who can act on vulnerability disclosures. The question the industry is watching is whether this model of controlled safety capability deployment scales: as models become more capable, the gap between beneficial security research and offensive weaponization narrows, and the governance structures for managing that gap are still being invented.

Back to Home

Related Stories

Musk Updates His OpenAI Lawsuit to Route Any $150 Billion Damages Award to the Nonprofit Foundation
Policy

Musk Updates His OpenAI Lawsuit to Route Any $150 Billion Damages Award to the Nonprofit Foundation

Elon Musk has amended his lawsuit against OpenAI with a strategic addition: any damages recovered — potentially up to $150 billion — should be redirected to OpenAI's nonprofit foundation rather than awarded to Musk personally. The update reframes the litigation from a personal grievance into a structural argument about OpenAI's obligations to its original charitable mission.

D.O.T.S AI Newsroom
OpenAI's Child Safety Blueprint Confronts AI's Role in the Surge of Child Sexual Exploitation
Policy

OpenAI's Child Safety Blueprint Confronts AI's Role in the Surge of Child Sexual Exploitation

OpenAI has released a Child Safety Blueprint outlining its approach to detecting, preventing, and reporting AI-generated child sexual abuse material. The document arrives as law enforcement agencies globally report a sharp increase in CSAM volume, with AI tools enabling the production of synthetic material at scale. It is the company's most detailed public statement on the problem it helped create.

D.O.T.S AI Newsroom
Hackers Are Redistributing the Leaked Claude Code Repository — With Bonus Malware Attached
Policy

Hackers Are Redistributing the Leaked Claude Code Repository — With Bonus Malware Attached

Wired reports that threat actors are repackaging the leaked Claude Code source repository and uploading it to file-sharing platforms bundled with information-stealing malware. The pattern is a textbook social engineering play: developers curious about the leaked AI tool are downloading what looks like the genuine repository and executing malware in the process.

D.O.T.S AI Newsroom