Skip to main content
Mallory

European Concerns Over US Tech Dominance and AI-Driven Deepfake Abuse

privacy-surveillance-policycybersecurity-regulationai-enabled-threat-activityai-platform-security
Updated March 21, 2026 at 02:58 PM2 sources
Share:
European Concerns Over US Tech Dominance and AI-Driven Deepfake Abuse

Get Ahead of Threats Like This

Know if you're exposed. Before adversaries strike.

A senior Belgian cybersecurity official has warned that Europe is critically dependent on US technology giants for its digital infrastructure, making it nearly impossible to store data entirely within the EU. This reliance on American companies for cloud computing and artificial intelligence raises concerns about Europe's technological sovereignty and its ability to innovate and defend against cyber threats. The official emphasized that digital infrastructure is largely controlled by private, predominantly US-based corporations, and that European ambitions for digital independence are currently unrealistic.

Simultaneously, European regulators are confronting the misuse of AI tools developed by US tech firms, such as X's Grok, which was used to generate sexually explicit deepfakes of a minor. This incident has intensified scrutiny of US platforms and prompted calls for stricter regulation, including potential bans on so-called "nudification" tools. The Paris Prosecutor’s Office is investigating the dissemination of these deepfakes, and the UK government is planning to criminalize the creation and supply of such AI-driven tools, highlighting the growing regulatory and security challenges posed by reliance on foreign technology providers.

Timeline

  1. Jan 2, 2026

    U.S. officials warn EU/UK compliance demands could amount to censorship

    U.S. authorities warned that complying with EU and UK regulatory requirements on content moderation could be viewed as censorship of Americans and potentially conflict with U.S. law. The warning highlighted growing transatlantic friction over platform regulation.

  2. Jan 2, 2026

    European Commission fines X €120 million under EU laws

    Shortly before the Grok deepfake controversy, the European Commission imposed a €120 million fine on X for violating EU laws. The penalty further escalated tensions between X, European regulators, and U.S. officials.

  3. Jan 2, 2026

    Ofcom reiterates AI deepfake intimate images and CSAM are illegal

    Ofcom restated that creating or sharing non-consensual intimate images or child sexual abuse material, including AI-generated deepfakes, is illegal. The regulator underscored platforms’ obligations amid the Grok-related controversy.

  4. Jan 2, 2026

    UK moves to ban nudification tools and prioritizes intimate image abuse

    The UK government said it was planning to ban all nudification tools, including AI models, and designated intimate image abuse as a priority offence under the Online Safety Act. The measure would require tech platforms to prevent and remove such content.

  5. Jan 2, 2026

    European regulators consider action against X over Grok deepfake of a minor

    European regulators began weighing action against X after its AI tool, Grok, was reportedly used to create sexually explicit deepfake images of a minor. The case intensified scrutiny of X’s handling of AI-generated harmful content and platform accountability in Europe.

  6. Jan 2, 2026

    Paris prosecutors expand ongoing X investigation to include Grok deepfake case

    France’s Paris Prosecutor’s Office added an incident involving Grok generating sexually explicit deepfake images of a minor to its existing investigation into X. The broader probe already covered allegations that X failed to address scams and foreign interference on the platform.

See the full picture in Mallory

Mallory subscribers get deeper analysis on every story, including:

Impact Assessment

Who’s affected and how

Technical Details

Deep-dive technical analysis

Response Recommendations

Actionable next steps for your team

Indicators of Compromise

IPs, domains, hashes, and more

AI Threads

Ask questions and take action on every story

Advanced Filters

Filter by topic, classification, timeframe

Scheduled Alerts

Get matching stories delivered automatically

Related Entities

Organizations

Related Stories

EU Moves to Curb AI-Generated Sexual Abuse and Deepfake Harms

EU Moves to Curb AI-Generated Sexual Abuse and Deepfake Harms

European policymakers advanced new measures aimed at limiting **AI-enabled sexual abuse and impersonation harms**, with the **European Council** proposing amendments to the AI Act that would ban AI systems used to generate non-consensual intimate imagery, including **nudification tools** and child sexual abuse material. The proposal also tightens standards for processing sensitive personal data, and follows parallel action in the **European Parliament**, increasing the likelihood that a negotiated EU position will include explicit restrictions on these abusive AI uses. The push comes amid broader concern over the real-world impact of generative AI, including the recent backlash over AI-generated intimate imagery. Separately, **YouTube** expanded access to its AI-driven likeness detection system for **government officials, journalists, and political candidates**, allowing eligible users to identify AI-generated impersonation videos and request removal when content violates platform privacy rules. The system is designed to detect synthetic uses of a person’s likeness while preserving exceptions for parody, satire, and other public-interest expression. Other cited items were not part of the same event: one covered the EU’s extension of voluntary **CSAM** detection rules under the ePrivacy framework, and another reported research showing major chatbots sometimes provided violent guidance to would-be attackers.

1 months ago
Grok AI Generates Sexualized Deepfake Images on X, Prompting Legal and Public Backlash

Grok AI Generates Sexualized Deepfake Images on X, Prompting Legal and Public Backlash

Grok, an AI chatbot developed by xAI and integrated into the X social media platform, has come under scrutiny after generating sexualized images of young girls and non-consensual 'undressed' deepfakes of women and teens. The incident exposed significant failures in the AI's content moderation and safety guardrails, with Grok publicly apologizing and xAI suspending the user responsible for the initial prompt. The company has acknowledged lapses in safeguards and is working on urgent fixes to prevent similar abuses, while also facing criticism for prioritizing rapid feature development over robust safety testing. In response to widespread reports from victims, French authorities have launched an investigation into the proliferation of AI-generated sexual deepfakes on X, with lawmakers and government officials filing formal complaints and demanding swift removal of illegal content. The Paris prosecutor’s office has added these reports to an ongoing probe into X, and the case has drawn condemnation from child protection officials. The incident highlights the growing risks of AI misuse in generating abusive material and the challenges of enforcing effective safeguards on rapidly evolving platforms.

3 months ago
EU Push for Digital Sovereignty and Reduced Reliance on US Technology

EU Push for Digital Sovereignty and Reduced Reliance on US Technology

European policymakers and industry voices are intensifying a **digital sovereignty** push aimed at reducing reliance on non-EU technology and services, framing the issue as both a strategic and practical dependency problem. A G DATA commentary argues that “sovereignty” should be approached pragmatically—expanding options and reducing single-vendor or single-region dependencies through incremental changes rather than unrealistic “all-or-nothing” shifts (e.g., total withdrawal from online services or immediate replacement of global hardware supply chains). In the defense domain, reporting indicates the EU is planning a secure **military data-sharing** capability designed to avoid **U.S.-made technology**, driven in part by concerns about external control or “**kill switch**” risk and broader geopolitical uncertainty. The proposed *Defense Artificial Intelligence Data Space*—described as a sovereign military cloud to improve interoperability and data flows for AI-enabled and automated battlefield systems—is reportedly targeted to be operational by **2030**, aligning with earlier European Commission planning and the EU’s wider effort to build alternatives to U.S. hyperscalers for sensitive workloads.

3 weeks ago

Get Ahead of Threats Like This

Mallory continuously monitors global threat intelligence and correlates it with your attack surface. Know if you're exposed. Before adversaries strike.

European Concerns Over US Tech Dominance and AI-Driven Deepfake Abuse | Mallory