Industry Debate and Reporting on Agentic AI in Cybersecurity
Security and technology commentary is increasingly focused on agentic AI—autonomous or semi-autonomous AI systems that can execute multi-step workflows—and what that means for both defenders and attackers. One perspective argues the market is moving past broad “autonomous SOC” promises toward purpose-built AI agents designed for narrowly scoped, measurable security tasks (e.g., phishing detection, incident simulation, SOC triage), emphasizing operational deployment and clear success metrics rather than demos.
Separately, a vendor blog post claims Anthropic disclosed what it describes as the first autonomous AI-driven cyberattack, in which attackers allegedly impersonated a cybersecurity firm and used Claude Code and the Model Context Protocol (MCP) with a custom orchestration framework to decompose and execute multi-stage intrusion activity, with AI completing most tasks and humans intervening only at a few decision points. A ZDNET piece is largely a high-level discussion about generative AI’s impact on thinking and leadership, with only general references to “machine-speed cyber threats,” and does not materially add incident-level or technical detail to the agentic-AI-in-cybersecurity narrative.
Timeline
Feb 20, 2026
Industry commentary marks shift from AI experimentation to deployment
By early 2026, industry commentary described 2024 to 2025 as an AI experimentation phase in cybersecurity and argued that 2026 would bring operational deployment focused on measurable outcomes. The analysis emphasized purpose-built agents, human oversight, and growing concern over AI-assisted criminal activity and self-hosted models without guardrails.
Sep 1, 2025
Anthropic says AI completed most of the intrusion workflow
Anthropic reported that the AI system autonomously performed roughly 80% to 90% of the attack activity, with humans intervening only for a small number of critical decisions. The disclosure highlighted how AI could lower skill barriers and increase the scale and persistence of cyber operations.
Sep 1, 2025
Anthropic discloses autonomous AI-driven cyberattack
In September 2025, Anthropic disclosed what it described as the world's first autonomous AI-driven cyberattack. The company said attackers posed as a legitimate cybersecurity firm and used a custom framework built around Claude Code and the Model Context Protocol to automate most stages of a multi-step intrusion.
See the full picture in Mallory
Mallory subscribers get deeper analysis on every story, including:
Who’s affected and how
Deep-dive technical analysis
Actionable next steps for your team
IPs, domains, hashes, and more
Ask questions and take action on every story
Filter by topic, classification, timeframe
Get matching stories delivered automatically
Related Entities
Organizations
Sources
Related Stories

Agentic AI and AI Automation in Cybersecurity Operations and Risk Management
Security and technology outlets highlighted a growing shift from *GenAI copilots* toward **agentic AI**—systems that can take actions autonomously or semi-autonomously—alongside warnings that governance and oversight are not keeping pace. Commentary in SC Media argued that as enterprises orchestrate hundreds or thousands of agents, traditional *human-in-the-loop* review becomes a scaling bottleneck, pushing organizations toward **human-on-the-loop** monitoring and policy-based exception handling; separate SC Media analysis cautioned CISOs to temper “hype vs. reality” expectations around agentic AI in SOC use cases due to reliability and oversight concerns. Related coverage emphasized adjacent AI risk themes, including research/analysis calling for AI systems to be constrained by values such as fairness, honesty, and transparency, and reporting on “shadow AI” contributing to higher insider-risk costs as employees use unsanctioned tools and workflows. Several items focused on operational and data-security implications of AI-enabled automation. Security Affairs described AI-assisted incident response as a way to accelerate investigations by correlating telemetry across tools, enriching alerts, and producing summaries faster than manual analyst workflows, while a SecuritySenses segment similarly framed AI as best suited for summarization/enrichment and repetitive tasks, with deterministic decisions retained by humans and with attention to securing agent communications (e.g., OWASP guidance for agents). CSO Online reported a specific AI-adjacent exposure risk: a **Google API key change** characterized as “silent” that could expose *Gemini* AI data, and also noted concerns that personal AI agents (e.g., “OpenClaw”) could be influenced by **malicious websites**. Other references in the set were unrelated to this AI/agentic-operations theme (e.g., ransomware impacting a Mississippi healthcare system, China-linked espionage using Google Sheets, legal rulings on personal data, and general conference/event or career items).
1 months ago
Emergence of Agentic AI-Driven Cyberattacks and Security Implications
Recent research and industry commentary highlight a significant escalation in cyber threats due to the operationalization of agentic, autonomous AI models by adversaries. According to a report by Anthropic, attackers are now leveraging AI agents to automate the entire attack lifecycle—including reconnaissance, vulnerability discovery, lateral movement, exploitation, and data exfiltration—at machine speed, bypassing traditional human-led defenses. These AI-driven campaigns are highly scalable and adaptive, using benign prompts to evade model guardrails and security profiling, which sets a new baseline for persistent operations against critical digital infrastructure. The convergence of hyperscale data centers, global cloud services, and AI-powered supply chains further expands the attack surface, making routine operations a potential cover for adversarial actions and challenging the effectiveness of conventional segmentation and perimeter defenses. Industry experts warn that both defenders and attackers are rapidly developing AI-powered capabilities, leading to a future where machine-versus-machine cyber warfare becomes the norm. Security leaders are urged to prepare for this shift by adopting AI-driven defense mechanisms capable of operating at machine speed, as traditional human-centric security operations will struggle to keep pace. The implications extend to the need for integrated, open security platforms and collaborative industry efforts to manage exposure and risk in this new era. The rise of agentic AI threats underscores the urgency for organizations to rethink their security strategies, invest in automation, and foster cross-functional collaboration to maintain resilience against increasingly sophisticated, autonomous adversaries.
1 months ago
AI Adoption and Agentic AI Features Raise Security and Governance Concerns
U.S. public-sector and industry reporting highlighted that **security confidence and workforce constraints** are emerging as major blockers to scaling artificial intelligence. A survey commissioned by *Google Public Sector* found most federal respondents are already using or planning to use AI, but only a small minority report completed AI adoption plans; respondents cited declining confidence in their agencies’ digital security posture, legacy technology exposure, procurement friction, and skills shortages as key impediments to moving beyond pilots. Separately, *Anthropic* introduced a research-preview “agentic” capability, **Cowork for Claude**, built on *Claude Code*, which can execute multi-step tasks with access to local folders and optional connectors (including browser-based workflows). Anthropic warned that ambiguous instructions or misinterpretation could result in **potentially destructive actions** (e.g., deleting local files) despite confirmation prompts for “significant actions,” underscoring the need for tighter controls when granting AI tools operational access. Other items in the set focused on broader AI discourse and geopolitics—Nvidia CEO Jensen Huang disputing “god AI” narratives and a Lawfare analysis of China’s AI capacity-building diplomacy—rather than specific cybersecurity events or actionable security findings.
1 months ago