Anthropic is quickly earning a reputation for being a responsible AI standard bearer, especially after CEO Dario Amodei’s moral clash with the US Department of Defense in Feb 2026. But the fact that its upcoming model, dubbed Claude Mythos, leaked online ahead of schedule thanks to a CMS error is ironic to say the least.
According to the leaked draft of Anthropic’s blog post, Claude Mythos is so powerful and “far ahead of any other AI model in cyber capabilities” that it can lead to an overwhelming “wave of models that can exploit vulnerabilities in ways that far outpace the efforts of defenders.”
That’s Anthropic suggesting they’re cooking up something so powerful in their next AI product it can essentially wreak havoc online in terms of cybersecurity, something that the world’s best cybersecurity companies won’t be able to handle – with or without AI.
That’s what scares Anthropic about Mythos’ capabilities, and why (according to the leaked documents) they were planning to engage top business leaders behind closed doors and brief and prepare them on next steps.
How do we know all this ahead of schedule? Because someone at Anthropic accidentally misconfigured these posts on the website’s CMS, making thousands of documents publicly viewable. One wonders if it isn’t a euphemistic rephrasing of a cyber breach. Irony-laced jokes about Claude Mythos notwithstanding, Anthropic claims the Mythos model is currently being trialed by “early access customers” only, deliberately and carefully rolling out much slower than previous Claude models.
Also read: Anthropic accidentally reveals Claude Mythos, its most powerful AI model yet
That underlying concern Anthropic’s exhibiting isn’t just limited to them. AI has the potential to disrupt the cybersecurity industry rapidly, for better and worse. There’s data to back it up.
According to Crowdstrike, AI-enabled attacks jumped 89% year-over-year in 2025, noting that the speed and scale of cyber breaches was only accelerating. Business email hack attacks rose 37% in 2025, suggests a Deepstrike report, using AI deepfake tech like voice cloning and synthetic identities. In its 2026 AI security report, Cisco has warned that AI is fast expanding the cyber attack surface with potential new risks like agent misuse and supply-chain attacks.
In fact, cybercriminals are “weaponizing AI to orchestrate attacks of unprecedented sophistication and scale,” according to Kaspersky Lab’s Evolving Threat Landscape Report of December 2025. According to the cybersecurity company, “malicious actors are deploying generative AI to create convincing phishing emails, voice clones, and deepfake videos that bypass traditional detection mechanisms.” These AI-driven attacks will target AI models corporate systems through evasion, poisoning, and model extraction techniques, said Kaspersky.
Claude Mythos, still unreleased and already controversial, seemingly sits at the crossroads of AI-linked cybersecurity. Because, from the looks of it, the same AI model powerful enough to identify zero-day vulnerabilities and fortify codebases is also equally capable of being weaponised in the wrong hands. Anthropic’s decision to initially limit access to defensive cybersecurity use cases is a measured response, but only for the time being. It can’t control what others working the AI crucible will eventually forge and unleash.
Also read: OpenAI flags rising cyber threats as AI models get more powerful