Anthropic is investigating reports that unauthorized users gained access to its unreleased Claude Mythos Preview, a highly advanced system designed for cybersecurity applications. According to a report by Bloomberg, a small group accessed the model through a third-party vendor environment on the same day Anthropic began limited testing with approved organizations.
The company confirmed it is reviewing the incident, stating that it is examining claims of unauthorized access through external infrastructure. The model, part of Anthropic’s Project Glasswing initiative, is being deployed under strict controls to a limited number of partners for defensive cybersecurity purposes.
The reported breach raises concerns given the model’s capabilities. Mythos is designed to identify software vulnerabilities and simulate cyberattacks, functions that are typically restricted due to potential misuse.
High-Stakes Capabilities and Risks
Anthropic has positioned Mythos as a tool for strengthening cybersecurity by helping organizations detect weaknesses before attackers do. However, its capabilities also highlight the dual-use nature of advanced AI systems.
Reports suggest the model can identify complex and previously unknown vulnerabilities, construct multi-step attack scenarios, and generate functional exploit code. Such capabilities could significantly lower the barrier to entry for cyberattacks if misused.
Regulators and industry observers have already expressed concern about systems like Mythos, which blur the line between defensive and offensive cybersecurity tools. Anthropic has limited access to the model and avoided a full public release, citing safety considerations.
Security Controls Under Scrutiny
The incident underscores the challenges of securing highly capable AI systems, particularly when they are deployed through third-party infrastructure. Even controlled rollouts can introduce vulnerabilities if external systems are involved.
Anthropic said it is continuing discussions with government and industry partners about safe deployment of the technology. The company has emphasized that Mythos is intended for defensive use, such as vulnerability research and red-teaming.
The situation also highlights broader questions about how AI developers should manage access to powerful models. As capabilities increase, ensuring that systems are used responsibly becomes more complex, especially when demand for such tools is high.
Growing Pressure Around AI Governance
The reported access comes amid heightened scrutiny of advanced AI models and their potential impact on cybersecurity. Governments and organizations are increasingly focused on balancing innovation with risk mitigation.
Anthropic has already faced regulatory attention and internal debate over how widely to release Mythos. The company has framed the model as a way to stay ahead of attackers, but the incident illustrates the difficulty of maintaining strict control over such systems.
As AI models become more capable of executing complex, real-world tasks, incidents like this may shape future policies on access, oversight, and deployment of high-risk technologies.