Anthropic Faces Scrutiny Over AI Safety Commitments and National Security Use Cases
Anthropic drew heightened scrutiny from security and policy communities over how its AI safety and governance commitments are evolving and how its models are being positioned for sensitive use cases. A Help Net Security analysis reported that Anthropic’s updated Responsible Scaling Policy (RSP) 3.0 represents a structural shift from maintaining absolute risk below fixed thresholds to a more relative, competitor-dependent posture—implying Anthropic may be less willing to pause or constrain capability development if peers do not. The same reporting also noted Anthropic’s launch of Claude Code Security as a move that unsettled parts of the cybersecurity market and raised questions about trust and vendor assurances in security-adjacent AI offerings.
In parallel, Lawfare reported the Pentagon labeled Anthropic a national security risk tied to usage restrictions Anthropic imposed on a military contract, while also describing reporting that the U.S. military used Anthropic’s Claude model in initiating operations in Iran less than a day later—highlighting the tension between policy concerns and rapid military adoption of frontier AI. Separately, Anthropic announced the creation of the Anthropic Institute, a research unit intended to study long-term societal impacts and risks from advanced AI; the company stated its models can already discover severe cybersecurity vulnerabilities and argued that governments and industry will face near-term governance challenges as capabilities accelerate.
Timeline
Mar 11, 2026
Anthropic announces the Anthropic Institute
On March 11, 2026, Anthropic announced the creation of the Anthropic Institute, a research unit focused on long-term societal risks from advanced AI. The company said the institute would study issues such as cybersecurity, economic disruption, governance, and value alignment, while partnering with outside experts.
Mar 1, 2026
Anthropic sues over its supply-chain risk designation
Anthropic filed a lawsuit challenging the U.S. government's designation of the company as a supply-chain risk. The legal action was later supported by an amicus brief from employees at OpenAI and Google DeepMind, according to one reference.
Feb 1, 2026
OpenAI replaces Anthropic in Pentagon-related work
Following Anthropic's removal, OpenAI replaced Anthropic in the relevant government role or workstream. One account adds that OpenAI later backed away from claims it would follow Anthropic's stated red lines.
Feb 1, 2026
U.S. government suspends Anthropic AI use and labels it a supply-chain risk
After Anthropic's public stance, the U.S. government suspended use of Anthropic tools in government institutions and designated the company as a supply-chain risk. Multiple references describe this as a major escalation in the conflict between Anthropic and federal authorities.
Feb 1, 2026
Anthropic publicly opposes certain Pentagon AI use cases
Later in early 2026, Anthropic took a public stance against Pentagon uses of AI such as mass surveillance of Americans and fully autonomous weapons. This position became a key trigger in the subsequent dispute with the U.S. government.
Jan 1, 2026
Anthropic releases Responsible Scaling Policy 3.0
In early 2026, Anthropic published RSP 3.0, reportedly shifting from fixed safety thresholds toward a more competitor-relative posture. Commentators cited this as a meaningful change to the commitments underlying Anthropic's safety-focused brand.
Jan 1, 2026
Anthropic launches Claude Code Security
In early 2026, Anthropic launched Claude Code Security as part of its push into cybersecurity-focused AI offerings. The release is described as part of a rapid sequence of public moves that shaped perceptions of the company in the security community.
See the full picture in Mallory
Mallory subscribers get deeper analysis on every story, including:
Who’s affected and how
Deep-dive technical analysis
Actionable next steps for your team
IPs, domains, hashes, and more
Ask questions and take action on every story
Filter by topic, classification, timeframe
Get matching stories delivered automatically
Related Entities
Affected Products
Sources
Related Stories

Pentagon Ultimatum to Anthropic Over Expanded Claude Access and Defense Supply-Chain Risk Threat
U.S. Defense Secretary Pete Hegseth reportedly issued a near-term deadline for **Anthropic** to provide expanded access to its **Claude** AI model for use in classified and operational environments, prompting analysts to warn the ultimatum is unrealistic and could create **cybersecurity and supply-chain** knock-on effects across the **defense industrial base (DIB)**. Reporting indicates the Pentagon is reviewing its business relationship with Anthropic after weeks of negotiations over model access, safeguards, and constraints, and that Hegseth has warned Anthropic could be designated a **“supply chain risk”** or face other punitive measures if it does not meet military requirements. Separate commentary highlighted the emerging risk of **AI-enabled cyber operations**, citing an Anthropic disclosure that **Chinese threat actors** allegedly jailbroke *Claude Code* and used it to target roughly **30 companies and government agencies** globally in what was described as an early example of a large-scale campaign with minimal human involvement. The piece argues that many AI-assisted attacks will be harder to attribute or even recognize as AI-enabled because most activity will not occur on platforms with the same level of internal monitoring, and it calls out a gap in U.S. government capability to systematically identify whether incidents are driven by novel AI agent capabilities versus conventional tradecraft—an issue that intersects with the Pentagon’s push to operationalize frontier models while managing abuse, assurance, and supply-chain exposure.
1 months ago
Policy and industry debate over AI safety, governance, and data protection
U.S. policymakers and industry leaders are escalating scrutiny of **AI safety and data protection**, with a particular focus on sensitive data flows and the adequacy of existing guardrails. In a Senate HELP Committee hearing, lawmakers questioned whether federal guardrails are needed to protect Americans’ healthcare data voluntarily uploaded to AI-enabled apps and wearables that may fall outside HIPAA coverage, raising concerns about liability, downstream data use, and integration into medical records; HHS noted it is collecting public input via a request for information on safe and effective AI deployment in healthcare. Separately, commentary on AI governance and safety argues competitive pressure among frontier AI labs can erode safety practices and that clearer antitrust guidance could enable cross-industry collaboration on safety standards without triggering enforcement risk. Tensions over AI “red lines” in national security use also became more public, as **Anthropic** CEO Dario Amodei accused **OpenAI** of misleading messaging about defense work amid reports that Anthropic’s DoD talks faltered over restrictions related to mass domestic surveillance and autonomous weapons, while OpenAI described its agreement as permitting “all lawful purposes” alongside stated prohibitions. Broader, non-incident reporting highlighted enterprise investment to support *agentic AI* (with many data leaders citing governance lagging AI adoption) and general concerns about deepfakes, opaque models, and societal risk; however, several items in the set were primarily newsletters, vendor/industry promotion, or general-interest AI commentary rather than a single, discrete cybersecurity incident or vulnerability disclosure.
1 months ago
Pentagon–Anthropic Dispute Over Military AI Use and Provider Baselines
The U.S. Department of Defense has escalated a dispute with **Anthropic** over the conditions under which its AI models could be used by the military, after Anthropic reportedly insisted on limits including *no mass surveillance of Americans* and *no fully autonomous weapons*. Reporting cited in both accounts indicates Pentagon officials have discussed potentially designating Anthropic a **“supply chain risk”**—a step that could bar the company from government work and pressure defense contractors to sever ties—while at least one senior official was quoted as saying the department would “make sure they pay a price” for non-cooperation. At the same time, the Pentagon is engaging **Anthropic, OpenAI, Google, and xAI** to align all major U.S. AI providers on a common “baseline” of expectations, after contracts were signed with limited specificity and the department now wants to deploy models into DoD environments to enable broader development of AI agents with minimal human oversight. The coverage also describes the policy vacuum driving the standoff: key rules for military AI use are being shaped through ad hoc negotiations between the Pentagon and private AI firms, prompting calls for **Congress** to set durable, democratically accountable constraints rather than leaving governance to bilateral bargaining.
2 weeks ago