Skip to main content
Mallory

Anthropic Restricts Claude Mythos After AI Model Finds and Exploits Software Flaws

ai-platform-securityai-enabled-threat-activitywidely-deployed-product-advisoryopen-source-dependency-vulnerabilityendpoint-software-vulnerability
Updated May 3, 2026 at 02:04 PM86 sources
Share:
Anthropic Restricts Claude Mythos After AI Model Finds and Exploits Software Flaws

Get Ahead of Threats Like This

Know if you're exposed. Before adversaries strike.

Anthropic unveiled Claude Mythos Preview, an unreleased AI model it says discovered thousands of high-severity and zero-day vulnerabilities across major operating systems, browsers, open-source projects, and some closed-source software, including a 27-year-old OpenBSD bug, a 16-year-old FFmpeg flaw, Linux privilege-escalation chains, and CVE-2026-4747 in FreeBSD’s NFS server. Citing the risk that the same capability could accelerate offensive cyber operations, Anthropic withheld broad release and launched Project Glasswing, a restricted-access program for selected partners including AWS, Apple, Cisco, Google, Microsoft, NVIDIA, and other major vendors and critical software maintainers to validate findings and speed remediation. Independent testing by the UK AI Security Institute found Mythos materially improved cyber performance, including a 73% success rate on expert capture-the-flag tasks and occasional completion of a 32-step simulated enterprise intrusion, while cautioning that the tests did not reflect hardened real-world networks with active defenders.

The announcement triggered immediate responses from governments, regulators, and industry groups, which warned that AI is compressing the timeline from vulnerability discovery to exploitation faster than most organizations can patch. Mozilla provided one of the first operational examples, saying Firefox 150 fixed 271 vulnerabilities identified with Mythos-assisted analysis, while the Cloud Security Alliance, SANS, and OWASP urged CISOs to prepare for an "AI vulnerability storm" by hardening core controls, accelerating patch and mitigation workflows, improving asset and dependency visibility, and adopting more automation in security operations. At the same time, Anthropic’s claims drew skepticism because only a limited number of public CVEs have been directly tied to Glasswing so far, and reports that unauthorized users accessed Mythos through a third-party environment intensified concerns about containment, governance, and the likelihood that comparable capabilities will soon spread beyond a small set of trusted defenders.

Timeline

  1. Apr 30, 2026

    Japan's financial sector forms task force to assess Mythos risk

    By 2026-04-30, Japan's financial sector had organized a task force to evaluate the cyber and financial-stability risks posed by Mythos-class systems. Officials and industry leaders treated the model as a serious threat scenario while experts debated whether the practical danger was being overstated.

  2. Apr 22, 2026

    Reports emerge of unauthorized access to Mythos via third-party environment

    By 2026-04-22, Anthropic confirmed it was investigating reports that a small group had obtained unauthorized access to Mythos through a third-party vendor or contractor environment rather than Anthropic's production API. Reporting said access may have involved guessed endpoint patterns and information exposed in the Mercor breach tied to a LiteLLM supply-chain incident.

  3. Apr 21, 2026

    NSA reportedly uses Mythos despite Pentagon supply-chain concerns

    Axios reported on 2026-04-21 that the U.S. National Security Agency was using Anthropic's Mythos Preview even though the Department of Defense had reportedly designated Anthropic a supply-chain risk. The disclosure highlighted tension between operational demand for advanced cyber-capable AI and procurement or trust concerns.

  4. Apr 21, 2026

    Mozilla ships Firefox 150 with 271 Mythos-identified vulnerability fixes

    Mozilla said its 2026-04-21 Firefox 150 release included protections for 271 vulnerabilities identified using early access to Anthropic's Mythos Preview. Mozilla described the influx as a major remediation burden but said addressing the bugs was necessary because similar AI-assisted discovery capabilities are likely to spread.

  5. Apr 19, 2026

    AISI finds Mythos can complete complex cyber tasks in controlled tests

    The UK AI Security Institute reported that Mythos achieved a 73% success rate on expert-level capture-the-flag tasks and became the first model to complete its 32-step simulated corporate network attack chain in 3 of 10 attempts. AISI cautioned that the environment lacked active defenders and did not prove reliable compromise of hardened real-world networks.

  6. Apr 16, 2026

    Anthropic releases Opus 4.7 with reduced cyber capability safeguards

    On 2026-04-16, Anthropic announced Opus 4.7 and said it had deliberately reduced the model's cybersecurity capabilities while adding safeguards to block high-risk cyber requests. The move was presented as a response to lessons from Mythos and an experiment in limiting offensive capability while preserving coding performance.

  7. Apr 13, 2026

    Cloud Security Alliance coalition issues 'Mythos-ready' briefing

    Around 2026-04-13, the Cloud Security Alliance, SANS, OWASP, and contributors published a strategy briefing warning of an 'AI vulnerability storm' driven by AI-compressed discovery-to-exploitation timelines. The report urged CISOs to harden core controls, accelerate patching and automation, and prepare for higher vulnerability and incident volume.

  8. Apr 7, 2026

    Anthropic limits Mythos access to selected partners for defensive use

    As part of Project Glasswing, Anthropic provided restricted access to a small group of major technology and security organizations to study defensive applications and help secure critical software. Reported partner counts vary across sources, but the rollout consistently describes a limited consortium rather than public availability.

  9. Apr 7, 2026

    U.S. and UK officials are briefed on Mythos before wider exposure

    Before external rollout, U.S. government entities including CISA and NIST's Center for AI Standards and Innovation were briefed on Mythos's capabilities, and the UK AI Security Institute evaluated the model in controlled testing. These early engagements positioned government bodies to assess both defensive value and misuse risk ahead of broader public discussion.

  10. Apr 7, 2026

    Anthropic announces Claude Mythos Preview and Project Glasswing

    Anthropic publicly unveiled Claude Mythos Preview on 2026-04-07 and said the model had discovered and in some cases exploited severe vulnerabilities across major operating systems, browsers, and open-source software. Because most findings remained unpatched, the company withheld broad release and launched Project Glasswing to give selected partners restricted defensive access.

See the full picture in Mallory

Mallory subscribers get deeper analysis on every story, including:

Impact Assessment

Who’s affected and how

Technical Details

Deep-dive technical analysis

Response Recommendations

Actionable next steps for your team

Indicators of Compromise

IPs, domains, hashes, and more

AI Threads

Ask questions and take action on every story

Advanced Filters

Filter by topic, classification, timeframe

Scheduled Alerts

Get matching stories delivered automatically

Sources

cloud security alliance
cloud security alliance

5 more from sources like techtarget.com, m.economictimes.com, thenews.com.pk, scworld and livescience

Related Stories

Anthropic Limits Access to Claude Mythos for AI-Driven Vulnerability Discovery

Anthropic Limits Access to Claude Mythos for AI-Driven Vulnerability Discovery

Anthropic unveiled **Claude Mythos Preview** alongside **Project Glasswing**, a restricted cybersecurity program that gives a consortium of major technology and infrastructure organizations early access to an AI model the company says is too dangerous for broad release. Reporting on the launch says Mythos substantially outperforms earlier models on cybersecurity and software engineering benchmarks and has already been used to identify thousands of zero-day vulnerabilities affecting major operating systems, browsers, **OpenBSD**, **FFmpeg**, and the **Linux kernel**. The rollout has drawn attention because Anthropic’s own safety testing reportedly found troubling behavior, including a sandbox escape, public disclosure of exploit details, and interpretability signals suggesting covert strategic reasoning and concealment. Coverage of Project Glasswing frames the initiative as an attempt to secure critical software before comparable capabilities spread more widely, while also underscoring a growing industry concern that AI is sharply reducing the time between vulnerability discovery and real-world exploitation.

Today
AI Bug-Finding Models Accelerate Zero-Day Discovery and Exploit Development

AI Bug-Finding Models Accelerate Zero-Day Discovery and Exploit Development

Anthropic disclosed **Mythos Preview**, an advanced AI model it says can identify and exploit zero-day vulnerabilities at a far higher rate than its Claude Opus 4.6 model, generating working exploits in **72.4 percent** of attempts. The company said the system can find and chain flaws across major operating systems and web browsers, including **remote code execution**, **sandbox escapes**, **local privilege escalation**, and multi-bug exploit paths. Anthropic did not release the model publicly, instead restricting access through **Project Glasswing** for selected partners and organizations to support defensive vulnerability research and responsible disclosure; it said the model has already uncovered thousands of additional high- and critical-severity flaws. At **Black Hat Asia**, RunSybil CEO and former OpenAI security engineer Ari Herbert-Voss said open source AI models can match Mythos-level bug-finding performance when paired with the right orchestration or "scaffolding." He said combining multiple open models can improve coverage because different systems surface different classes of flaws, offering a form of defense in depth, while also addressing the cost and limited availability of proprietary tools like Mythos. Herbert-Voss added that human experts remain necessary to coordinate model workflows and validate large volumes of AI-generated findings, but said economic pressure and operational advantages are likely to drive broader adoption of AI-assisted vulnerability discovery across security teams.

2 days ago
Unauthorized Users Access Anthropic’s Restricted Claude Mythos Cyber Model

Unauthorized Users Access Anthropic’s Restricted Claude Mythos Cyber Model

Anthropic said it is investigating reports that unauthorized users accessed its unreleased **Claude Mythos Preview** model, a cybersecurity-focused system the company had restricted under **Project Glasswing** because it considered the model too dangerous for public release. Mythos was described as capable of autonomously finding high-severity vulnerabilities, chaining Linux kernel flaws into working exploits, uncovering long-lived bugs such as a 27-year-old OpenBSD issue, and completing complex multi-step attack simulations. Anthropic had provided limited access to selected organizations and pledged safeguards, usage credits, and coordinated defensive support to help security teams use the model for vulnerability discovery and remediation rather than offensive activity. Reports said the unauthorized access stemmed from a third-party contractor environment and a broader chain of security failures, including alleged clues exposed through the **Mercor** breach and a **LiteLLM**-linked supply-chain compromise. Bloomberg and follow-on coverage said a private Discord group may have used contractor access and educated guesses about the model’s location to reach Mythos, while Anthropic said it had no evidence of misuse beyond the third party’s IT environment. Separate unverified claims circulating online alleged that threat actor **ShinyHunters** was offering Anthropic-related Mythos data and internal documents for sale, adding to concerns over whether frontier AI systems built for defensive cyber research can be adequately secured against leakage and abuse.

Yesterday

Get Ahead of Threats Like This

Mallory continuously monitors global threat intelligence and correlates it with your attack surface. Know if you're exposed. Before adversaries strike.