A misconfigured content management system just exposed Anthropic's biggest secret. Nearly 3,000 files leaked into a public data lake, revealing Claude Mythos โ the company's most powerful AI model yet. This isn't just another incremental upgrade. Anthropic calls it a "step change" in capabilities, designed for a world where AI can exploit cybersecurity vulnerabilities faster than defenders can patch them.
๐ Read more: Claude Mythos Leak Exposes Anthropic's Most Dangerous AI Model
๐ From Human Error to Global Headlines
The leak started with a simple mistake. Anthropic's CMS defaulted to public URLs for uploads unless someone manually switched them to private. Classic human error, as the company admitted to Fortune. Result? Nearly 3,000 assets โ draft blog posts, PDFs, internal employee documents โ became accessible to anyone who knew where to look. Among the treasure hunters, Roy Paz from LayerX Security and Alexandre Pauwels from Cambridge University struck gold: a fully structured draft announcing Claude Mythos. The irony cuts deep. A model designed to tackle "unprecedented cybersecurity risks" got exposed through basic cybersecurity negligence.If you needed proof that even the most advanced AI companies have human problems, here it is. An AI model built to counter cyber threats leaked because of cyber carelessness.
โก Claude Mythos: The 'Step Change' Model
According to the leaked draft, Claude Mythos is "by far the most powerful AI model we have ever developed." That's not marketing speak. Anthropic later confirmed it represents a "step change" in AI performance and is "the most capable we have built to date."๐๏ธ The New Tier: Capybara
Here's where things get interesting. Mythos doesn't just improve on existing Claude models โ it creates an entirely new category called Capybara. Until now, Anthropic offered three tiers: - **Opus**: Most powerful and expensive - **Sonnet**: Faster and cheaper, but less capable - **Haiku**: Smallest, cheapest, and fastest Capybara sits *above* Opus โ more powerful, more expensive. "Compared to our previous best model, Claude Opus 4.6, Capybara achieves dramatically higher scores on software coding, academic reasoning, and cybersecurity tests," the leaked post states.3,000 Files leaked
4 tiers Claude models (with Capybara)
๐งช Early Access Phase
The model isn't ready for general release. It's in trial phase with "a small group of early access customers," as an Anthropic spokesperson confirmed. This careful approach isn't accidental โ it stems from serious risks the company has identified.๐ Read more: Claude Mythos Leak Exposes Anthropic's Most Powerful AI Model
๐จ The Risks That Worry Anthropic
This is where the story gets genuinely concerning. The leaked document describes Mythos as "currently far ahead of any other AI model on cyber capabilities." That's not necessarily good news. "It heralds an incoming wave of models that can exploit vulnerabilities in ways that far outpace defender efforts," Anthropic warns. Translation: hackers are about to get tools that make their job much easier.๐ฏ The Release Strategy
Instead of presenting Mythos as another milestone, Anthropic plans to offer it first to cyber defenders. "We want to understand the potential risks of the model in the cybersecurity realm โ and share the results to help cyber defenders prepare." It's a rare case where an AI company admits upfront that its product could be weaponized โ and tries to give the "good guys" a head start.We're providing early access to organizations, offering them a head start in improving their codebase resilience against the incoming wave of AI-driven exploits.
Anthropic draft blog post
๐ Read more: AI Manipulation: First Study Reveals Harmful Persuasion
๐ The Broader Context
Mythos isn't unique in raising these concerns. In February 2026, OpenAI released GPT-5.3-Codex โ the first model it categorized as "high capability" for cybersecurity tasks in its Preparedness Framework. Anthropic had similar experiences with Opus 4.6, released the same week. The model showed ability to discover unknown vulnerabilities in production codebases โ a dual-use capability that helps both hackers and defenders.๐ Real Threats, Not Theories
The concerns aren't hypothetical. Anthropic has reported that hacking groups, including those linked to the Chinese government, have attempted to exploit Claude in real cyberattacks. In one documented case, the company discovered a Chinese state-sponsored group was already running a coordinated campaign using Claude Code to infiltrate about 30 organizations โ tech companies, financial institutions, government agencies โ before the company caught them.Defense-First Approach
Anthropic gives cyber defenders first access to prepare for AI-driven exploits.
Vulnerability Discovery
The model can discover unknown security holes in production code.
๐ผ The 'Exclusive' Leak
The leak didn't just contain technical details. Among the 3,000 assets was a PDF describing an upcoming, invite-only retreat for European company CEOs in the UK, featuring Dario Amodei, Anthropic's CEO. The retreat is described as an "intimate gathering" for "thoughtful conversation" at an 18th-century manor converted to a hotel-and-spa in the English countryside. Participants will hear from lawmakers and policymakers about how businesses adopt AI and will experience unreleased Claude capabilities.๐ Read more: Anthropic Launches Cowork: Claude Code for Non-Programmers
