Anthropic is currently developing a new AI model that can easily outperform current offerings, even as an internal data leak unintentionally revealed details about the project. The company has confirmed the model is being tested with a limited group of early access users. As per the information that surfaced from the publicly accessible internal files, the upcoming model may be called Claude Mythos.
In a draft document reviewed by external sources, the model was described as the most powerful AI the company has made so far, with some improvements in reasoning, coding and cybersecurity capabilities. On the other hand, Anthropic acknowledged that exposure of these documents was due to a configuration error in the content management system, which made unpublished materials visible online. The company stated that the files were early drafts intended for internal use and has since restricted access to the data.
The leaked material also pointed to a new classification of AI systems internally referred to as Capybara, which appears to represent a tier above existing models like Opus. Anthropic currently categorises its models into tiers such as Opus, Sonnet and Haiku, but the new system is expected to exceed these in both capability and cost.
Also read: Google releases Lyria 3 Pro AI model with longer music generation: How to access
The company has flagged potential cybersecurity risks associated with the model. The internal description suggests that its advanced features can be misused to find and exploit different software vulnerabilities at scale, which might outpace current defensive systems. As a result, Anthropic is said to be approaching the rollout cautiously, initially offering access to a few organisations to strengthen the cyber defense.
Adding on, the model details that surfaced online also reportedly included references to a private executive event in Europe aimed at engaging business leaders on AI adoption.
This comes amid the increasingly capable models raising concerns around dual-use risk. Recently, OpenAI has faced similar concerns, particularly around models designed to detect vulnerabilities in software systems.