Claims of AI Systems Finding and Exploiting Zero-Day Vulnerabilities Raise Alarm
Reports and commentary have intensified around claims that advanced AI systems can now discover software flaws and, in some cases, produce working exploits for them. A widely discussed account about Anthropic’s reported system, Mythos, said access had been limited to about 40 organizations in critical infrastructure so they could identify and remediate vulnerabilities before the capability spreads more broadly. Security contacts cited in the reporting said the tool was finding a notable number of high-quality bugs, while skeptics cautioned that some of the claims may be overstated.
The debate has focused less on automated bug hunting alone than on the more consequential assertion that AI can bridge the gap from vulnerability discovery to exploit development, including alleged zero-day exploitation against major operating systems and browsers. Separate commentary described the development as a potential “AlphaGo moment” for vulnerability research, arguing that once such capability is demonstrated, containment becomes difficult because model replication, distillation, and parallel advances elsewhere could quickly erode any initial controls. The prospect has sharpened concerns for CISOs that defenders may face a dangerous period in which elite organizations gain access to powerful AI-assisted security tooling while critical infrastructure and legacy environments remain slow to patch, harden, or replace.
Timeline
Apr 17, 2026
Hacktron CTO reportedly builds Chrome V8 exploit chain with Claude Opus
Hacktron CTO Mohan Pedhapati said he used Anthropic's Opus 4.6 model to develop a full exploit chain against Chrome 138's V8 engine after about a week of iteration, roughly 20 hours of manual guidance, and 2.3 billion tokens costing about $2,283. The report highlighted potential downstream exposure for Electron-based apps such as Discord and Slack that may lag behind current Chromium/V8 releases.
Apr 11, 2026
Commentary frames Mythos as a potential turning point in vulnerability research
Subsequent analysis characterized the reported Mythos capability as an 'AlphaGo moment' for vulnerability research, highlighting the broader significance of AI-driven vulnerability discovery and exploitation claims.
Apr 8, 2026
Reports claim Mythos can find high-quality vulnerabilities and generate exploits
The reporting describes claims that Mythos can discover a surprising number of high-quality software vulnerabilities and, more significantly, generate working exploits, including for alleged zero-day vulnerabilities in major operating systems and web browsers.
Apr 8, 2026
Anthropic reportedly limits Mythos access to 40 critical-infrastructure organizations
According to the referenced reporting, Anthropic restricted access to its reported AI system, Mythos, to 40 organizations it considers part of critical infrastructure so they could identify and remediate vulnerabilities before broader release or leakage of the capability.
See the full picture in Mallory
Mallory subscribers get deeper analysis on every story, including:
Who’s affected and how
Deep-dive technical analysis
Actionable next steps for your team
IPs, domains, hashes, and more
Ask questions and take action on every story
Filter by topic, classification, timeframe
Get matching stories delivered automatically
Related Entities
Organizations
Sources
2 more from sources like vincenzoiozzo.com and vulnu
Related Stories

Anthropic Restricts Claude Mythos After AI Model Finds and Exploits Software Flaws
Anthropic unveiled **Claude Mythos Preview**, an unreleased AI model it says discovered thousands of high-severity and zero-day vulnerabilities across major operating systems, browsers, open-source projects, and some closed-source software, including a 27-year-old OpenBSD bug, a 16-year-old FFmpeg flaw, Linux privilege-escalation chains, and `CVE-2026-4747` in FreeBSD’s NFS server. Citing the risk that the same capability could accelerate offensive cyber operations, Anthropic withheld broad release and launched **Project Glasswing**, a restricted-access program for selected partners including AWS, Apple, Cisco, Google, Microsoft, NVIDIA, and other major vendors and critical software maintainers to validate findings and speed remediation. Independent testing by the UK AI Security Institute found Mythos materially improved cyber performance, including a **73%** success rate on expert capture-the-flag tasks and occasional completion of a 32-step simulated enterprise intrusion, while cautioning that the tests did not reflect hardened real-world networks with active defenders. The announcement triggered immediate responses from governments, regulators, and industry groups, which warned that AI is compressing the timeline from vulnerability discovery to exploitation faster than most organizations can patch. Mozilla provided one of the first operational examples, saying Firefox 150 fixed **271 vulnerabilities** identified with Mythos-assisted analysis, while the Cloud Security Alliance, SANS, and OWASP urged CISOs to prepare for an "AI vulnerability storm" by hardening core controls, accelerating patch and mitigation workflows, improving asset and dependency visibility, and adopting more automation in security operations. At the same time, Anthropic’s claims drew skepticism because only a limited number of public CVEs have been directly tied to Glasswing so far, and reports that unauthorized users accessed Mythos through a third-party environment intensified concerns about containment, governance, and the likelihood that comparable capabilities will soon spread beyond a small set of trusted defenders.
Today
AI Bug-Finding Models Accelerate Zero-Day Discovery and Exploit Development
Anthropic disclosed **Mythos Preview**, an advanced AI model it says can identify and exploit zero-day vulnerabilities at a far higher rate than its Claude Opus 4.6 model, generating working exploits in **72.4 percent** of attempts. The company said the system can find and chain flaws across major operating systems and web browsers, including **remote code execution**, **sandbox escapes**, **local privilege escalation**, and multi-bug exploit paths. Anthropic did not release the model publicly, instead restricting access through **Project Glasswing** for selected partners and organizations to support defensive vulnerability research and responsible disclosure; it said the model has already uncovered thousands of additional high- and critical-severity flaws. At **Black Hat Asia**, RunSybil CEO and former OpenAI security engineer Ari Herbert-Voss said open source AI models can match Mythos-level bug-finding performance when paired with the right orchestration or "scaffolding." He said combining multiple open models can improve coverage because different systems surface different classes of flaws, offering a form of defense in depth, while also addressing the cost and limited availability of proprietary tools like Mythos. Herbert-Voss added that human experts remain necessary to coordinate model workflows and validate large volumes of AI-generated findings, but said economic pressure and operational advantages are likely to drive broader adoption of AI-assisted vulnerability discovery across security teams.
2 days ago
AI-Driven Vulnerability Discovery and the Shrinking Window Before Exploitation
Security leaders are warning that **AI is accelerating zero-day discovery and exploit development**, compressing the time between vulnerability introduction, disclosure, and real-world abuse. A featured discussion of the **"Zero Day Clock"** argues that attackers benefit from a structural advantage because offensive validation is fast and binary, while defenders face slower, costlier verification and patching cycles; the result is a widening gap when organizations still operate on remediation timelines measured in weeks while active exploitation can begin in days. The reporting frames this as a material risk to enterprise resilience rather than a theoretical concern, especially as AI lowers the skill barrier for finding flaws in widely used software. One relevant reference also examines how **AI-enabled cyber operations** are becoming more autonomous, adaptive, and scalable, including target selection, phishing, and tactical decision-making without constant human direction. While focused on state espionage and policy implications rather than vulnerability research specifically, it supports the same broader development: AI is changing the speed and economics of offensive cyber activity. The remaining references are not about this event or topic; they cover detection strategy metrics, a personal newsletter essay, commercial spyware policy, software liability commentary, and a detection engineering newsletter, making them separate issues rather than part of the same story.
1 months ago