Anthropic's Month: Two Leaks in One Week Reveal a Pattern of Human-Error Exposure at the World's 'Safest' AI Lab
Within the span of a single week, Anthropic suffered two separate data exposure incidents caused by misconfigured internal systems — the Mythos model card leak and the Claude Code source code disclosure. For a lab whose brand identity is built on safety and security rigor, the back-to-back incidents raise pointed questions about internal operational controls.

D.O.T.S AI Newsroom
AI News Desk
Anthropic built its brand on a proposition that differentiates it from every other major AI lab: that safety is not a constraint on capability, but a precondition for it. The company's "responsible scaling policy," its Constitutional AI approach, its public positioning on frontier model risk — all of it rests on a claim to exceptional institutional rigor. Which makes the events of the last week at Anthropic unusually dissonant.
Two Incidents, One Week
The first incident emerged when Anthropic's internal content management system was misconfigured, making approximately 3,000 internal documents publicly accessible. Among the exposed materials were near-final draft blog posts and technical evaluation summaries for the company's next-generation frontier model, codenamed "Claude Mythos." D.O.T.S AI News was among the outlets that reviewed the materials before Anthropic secured them. The company confirmed to Fortune that it was "actively training and testing" the model — a disclosure it had clearly not intended to make.
Before the week was out, a second incident surfaced. Anthropic accidentally published the source code for Claude Code — the company's AI-powered developer tool — exposing approximately 512,000 lines of code that were not intended for public release. TechCrunch noted the irony: Claude Code had been positioned in part as a demonstration of Anthropic's engineering excellence, a showcase product used by developers at leading technology companies.
The Pattern Behind the Incidents
Both incidents share a structural cause: a human operator misconfigured a system in a way that made non-public information publicly accessible. Neither appears to have been an external attack or a sophisticated compromise. They were operational errors — the kind that occur at every large organization, but that carry heightened significance when they happen at a company whose core identity involves claims of superior operational discipline.
The incidents are not security failures in the traditional sense — no evidence suggests adversarial exploitation. But they do suggest a gap between Anthropic's public-facing safety narrative and its internal process controls. Safety culture, in the AI safety community's own framing, is supposed to be institutional: embedded in systems, not just in intentions.
Why This Matters Beyond PR
The practical consequences of the Mythos leak are real. Anthropic was forced into a premature public acknowledgment of its next major model, disrupting whatever controlled rollout strategy the company had planned. For a model the documents describe as having "far ahead of any other AI" cybersecurity capabilities — and that the company was deliberately rolling out slowly to security-focused evaluators — an accidental public disclosure is precisely the kind of event responsible scaling policies are designed to prevent.
The second leak compounds the reputational issue. Two misconfiguration incidents in a single week isn't a statistical anomaly; it suggests systematic gaps in how Anthropic manages internal system access controls. For the company's government and enterprise customers — for whom trust in Anthropic's operational security is a prerequisite for engagement — that's a message that will require more than a public statement to address.