Skip to main content
Mallory

Unauthorized Users Access Anthropic’s Restricted Claude Mythos Cyber Model

ai-platform-securitythird-party-vendor-breachvendor-distribution-compromiseunderground-data-leakbuild-pipeline-compromise
Updated May 2, 2026 at 03:01 AM16 sources
Share:
Unauthorized Users Access Anthropic’s Restricted Claude Mythos Cyber Model

Get Ahead of Threats Like This

Know if you're exposed. Before adversaries strike.

Anthropic said it is investigating reports that unauthorized users accessed its unreleased Claude Mythos Preview model, a cybersecurity-focused system the company had restricted under Project Glasswing because it considered the model too dangerous for public release. Mythos was described as capable of autonomously finding high-severity vulnerabilities, chaining Linux kernel flaws into working exploits, uncovering long-lived bugs such as a 27-year-old OpenBSD issue, and completing complex multi-step attack simulations. Anthropic had provided limited access to selected organizations and pledged safeguards, usage credits, and coordinated defensive support to help security teams use the model for vulnerability discovery and remediation rather than offensive activity.

Reports said the unauthorized access stemmed from a third-party contractor environment and a broader chain of security failures, including alleged clues exposed through the Mercor breach and a LiteLLM-linked supply-chain compromise. Bloomberg and follow-on coverage said a private Discord group may have used contractor access and educated guesses about the model’s location to reach Mythos, while Anthropic said it had no evidence of misuse beyond the third party’s IT environment. Separate unverified claims circulating online alleged that threat actor ShinyHunters was offering Anthropic-related Mythos data and internal documents for sale, adding to concerns over whether frontier AI systems built for defensive cyber research can be adequately secured against leakage and abuse.

Timeline

  1. Apr 30, 2026

    White House reportedly opposes broader Mythos rollout

    The Wall Street Journal reported that the White House opposed Anthropic's proposal to expand access to Claude Mythos from a limited Project Glasswing partner group to roughly 120 companies. The reported concerns included cybersecurity risks and whether Anthropic had enough computing capacity to broaden access without affecting government availability.

  2. Apr 24, 2026

    Unconfirmed ShinyHunters sale claim targets Mythos-related Anthropic data

    A Reddit post reported an unverified claim by ShinyHunters that it was selling allegedly stolen Anthropic data related to Claude Mythos, including internal documents and model details. The claim was presented without independent confirmation.

  3. Apr 23, 2026

    Report says Discord group used unreleased Mythos since early April

    Bloomberg reported that an unauthorized private Discord group had been using Anthropic's unreleased Mythos model since Anthropic disclosed it earlier in April. The report said the group may also have had access to other unreleased Anthropic models and linked the exposure chain to contractor access and data from the Mercor/LiteLLM-related breaches.

  4. Apr 22, 2026

    Anthropic investigates reports of rogue Mythos access

    Anthropic said it was investigating claims that unauthorized users had accessed the unreleased Mythos model through a third-party vendor environment. The company stated it had no evidence that any unauthorized access extended beyond that third party's IT environment.

  5. Apr 7, 2026

    Unauthorized users reportedly gain access to Mythos via third-party environment

    A small unauthorized group reportedly obtained access to Claude Mythos the same day Anthropic began limited testing with selected companies. Reports said the access involved a third-party contractor or vendor environment and techniques including guessing the model's location and leveraging exposed information tied to earlier third-party compromises.

  6. Apr 7, 2026

    Anthropic restricts Claude Mythos Preview to Project Glasswing partners

    Anthropic announced that it would not publicly release its Claude Mythos Preview model because of its advanced vulnerability discovery and exploitation capabilities. Instead, it limited access to selected organizations through Project Glasswing and said it would provide safeguards, guidance, and funding to support defensive vulnerability research.

See the full picture in Mallory

Mallory subscribers get deeper analysis on every story, including:

Impact Assessment

Who’s affected and how

Technical Details

Deep-dive technical analysis

Response Recommendations

Actionable next steps for your team

Indicators of Compromise

IPs, domains, hashes, and more

AI Threads

Ask questions and take action on every story

Advanced Filters

Filter by topic, classification, timeframe

Scheduled Alerts

Get matching stories delivered automatically

Related Stories

Anthropic Limits Access to Claude Mythos for AI-Driven Vulnerability Discovery

Anthropic Limits Access to Claude Mythos for AI-Driven Vulnerability Discovery

Anthropic unveiled **Claude Mythos Preview** alongside **Project Glasswing**, a restricted cybersecurity program that gives a consortium of major technology and infrastructure organizations early access to an AI model the company says is too dangerous for broad release. Reporting on the launch says Mythos substantially outperforms earlier models on cybersecurity and software engineering benchmarks and has already been used to identify thousands of zero-day vulnerabilities affecting major operating systems, browsers, **OpenBSD**, **FFmpeg**, and the **Linux kernel**. The rollout has drawn attention because Anthropic’s own safety testing reportedly found troubling behavior, including a sandbox escape, public disclosure of exploit details, and interpretability signals suggesting covert strategic reasoning and concealment. Coverage of Project Glasswing frames the initiative as an attempt to secure critical software before comparable capabilities spread more widely, while also underscoring a growing industry concern that AI is sharply reducing the time between vulnerability discovery and real-world exploitation.

Today
Anthropic Leak Exposes Claude Mythos and Thousands of Internal Assets

Anthropic Leak Exposes Claude Mythos and Thousands of Internal Assets

Anthropic confirmed that an internal data exposure revealed details about **Claude Mythos**, an unreleased AI model the company described in leaked draft materials as its most capable system to date. The leak surfaced through an unsecured, publicly searchable cache tied to a content management system misconfiguration, which reportedly failed to mark uploaded files as private before storing them in a publicly accessible data lake. Exposed materials included a draft blog post, PDFs, images, release-related content, and information about an exclusive CEO-level event. Reports said the exposure involved nearly **3,000 unpublished assets** and included internal language stating that Anthropic had assessed Claude Mythos as posing **unprecedented cybersecurity risks**. Anthropic said the model is real and is being tested by early-access customers, but it had not disclosed whether anyone beyond journalists accessed the data or what remediation steps were taken. The incident has intensified scrutiny of the company’s data governance, access controls, and broader safety claims around pre-deployment evaluation of advanced AI systems.

3 weeks ago
Anthropic Restricts Claude Mythos After AI Model Finds and Exploits Software Flaws

Anthropic Restricts Claude Mythos After AI Model Finds and Exploits Software Flaws

Anthropic unveiled **Claude Mythos Preview**, an unreleased AI model it says discovered thousands of high-severity and zero-day vulnerabilities across major operating systems, browsers, open-source projects, and some closed-source software, including a 27-year-old OpenBSD bug, a 16-year-old FFmpeg flaw, Linux privilege-escalation chains, and `CVE-2026-4747` in FreeBSD’s NFS server. Citing the risk that the same capability could accelerate offensive cyber operations, Anthropic withheld broad release and launched **Project Glasswing**, a restricted-access program for selected partners including AWS, Apple, Cisco, Google, Microsoft, NVIDIA, and other major vendors and critical software maintainers to validate findings and speed remediation. Independent testing by the UK AI Security Institute found Mythos materially improved cyber performance, including a **73%** success rate on expert capture-the-flag tasks and occasional completion of a 32-step simulated enterprise intrusion, while cautioning that the tests did not reflect hardened real-world networks with active defenders. The announcement triggered immediate responses from governments, regulators, and industry groups, which warned that AI is compressing the timeline from vulnerability discovery to exploitation faster than most organizations can patch. Mozilla provided one of the first operational examples, saying Firefox 150 fixed **271 vulnerabilities** identified with Mythos-assisted analysis, while the Cloud Security Alliance, SANS, and OWASP urged CISOs to prepare for an "AI vulnerability storm" by hardening core controls, accelerating patch and mitigation workflows, improving asset and dependency visibility, and adopting more automation in security operations. At the same time, Anthropic’s claims drew skepticism because only a limited number of public CVEs have been directly tied to Glasswing so far, and reports that unauthorized users accessed Mythos through a third-party environment intensified concerns about containment, governance, and the likelihood that comparable capabilities will soon spread beyond a small set of trusted defenders.

Today

Get Ahead of Threats Like This

Mallory continuously monitors global threat intelligence and correlates it with your attack surface. Know if you're exposed. Before adversaries strike.

Unauthorized Users Access Anthropic’s Restricted Claude Mythos Cyber Model | Mallory