In a rapidly evolving AI landscape, even controlled deployments can face unexpected security challenges.
Anthropic is currently investigating reports that a small group of unauthorized users may have gained access to its unreleased Claude Mythos Preview model — a high-capability system designed for restricted testing under strict security conditions.
The incident highlights growing concerns around safeguarding frontier AI systems, especially those capable of advanced reasoning, coding, and cybersecurity-related tasks.
What Happened
According to a report by Bloomberg News (citing internal documentation and a source familiar with the situation), a handful of users in a private online forum were able to access the Mythos model without authorization.
This access reportedly occurred through a third-party vendor environment, raising questions about supply-chain and infrastructure security in AI deployment pipelines.
Interestingly, the unauthorized access began around the same time Anthropic officially announced limited access to Mythos for selected companies under controlled testing conditions.
Anthropic’s Response
Anthropic confirmed the situation is under investigation, stating:
“We’re investigating a report claiming unauthorized access to Claude Mythos Preview through one of our third-party vendor environments.”
The company has not yet disclosed the scale of the breach or whether sensitive outputs or system behaviors were exposed during the access window.
What is Claude Mythos?
Introduced on April 7 as part of Project Glasswing, Mythos is a restricted AI model being tested with select organizations in a controlled environment.
The goal of the initiative is to evaluate how advanced AI systems perform in defensive cybersecurity scenarios, including:
- Identifying software vulnerabilities
- Supporting secure code analysis
- Strengthening cyber defense mechanisms
However, its capabilities have also raised concerns among regulators and researchers due to its potential dual-use nature — meaning it can be leveraged for both defensive and offensive cybersecurity applications.
Why This Matters
This incident underscores a critical issue in modern AI development:
security is no longer just about the model — it’s about the entire ecosystem around it.
Even if an AI model is tightly controlled, risks can emerge from:
- Third-party vendor integrations
- Access management gaps
- Testing environments outside direct company control
For frontier AI systems like Mythos, the stakes are especially high due to their ability to perform advanced technical tasks, including vulnerability discovery and code generation.
The Bigger Picture
As AI systems become more powerful, organizations are increasingly focusing on:
- Secure model deployment pipelines
- Vendor risk management
- Controlled access testing environments
- Stronger audit and monitoring systems
The Mythos incident may accelerate discussions around how next-generation AI models should be safely distributed — especially in early access or preview stages.

