Skip to main content
Mallory

Prompt Injection and Persistent Memory Exploits in AI-Powered Browsers

ai-platform-securitydata-exfiltration-methodpersistence-methodinternet-facing-service-vulnerabilityidentity-authentication-vulnerability
Updated April 9, 2026 at 04:00 AM3 sources
Share:
Prompt Injection and Persistent Memory Exploits in AI-Powered Browsers

Get Ahead of Threats Like This

Know if you're exposed. Before adversaries strike.

Researchers have identified critical security vulnerabilities in several AI-powered browsers, including OpenAI's Atlas and other emerging platforms such as Comet and Fellou. These browsers, which allow AI agents to perform actions on behalf of users, are susceptible to prompt injection attacks—where hidden or malicious instructions embedded in web content are executed by the AI. In documented cases, attackers were able to hide commands in web pages or images, leading the browser to perform unauthorized actions such as extracting email subject lines and exfiltrating data to attacker-controlled sites, all without user confirmation.

A particularly severe exploit targets the persistent memory feature of the ChatGPT Atlas browser, introduced by OpenAI to personalize user experiences. By chaining a cross-site request forgery (CSRF) vulnerability with a memory write, attackers can inject malicious instructions that persist across sessions, devices, and even different browsers. This allows for ongoing compromise, including privilege escalation, malware deployment, and account takeover, unless users manually clear the tainted memory. The persistence and stealth of these attacks significantly elevate the risk profile for users of AI-enabled browsers, highlighting the urgent need for robust security controls and user awareness around prompt injection threats.

Timeline

  1. Oct 28, 2025

    Researchers report prompt injection risks across AI browsers

    By late October 2025, security research highlighted that AI browsers including OpenAI Atlas, Comet, and Fellou are vulnerable to direct and indirect prompt injection attacks. The findings showed hidden instructions in web pages or URLs could trigger unauthorized actions such as data exfiltration or changing user settings.

  2. Oct 27, 2025

    NeuralTrust demonstrates related prompt injection attack on ChatGPT Atlas

    NeuralTrust demonstrated a separate but related prompt injection attack affecting ChatGPT Atlas, underscoring broader security weaknesses in AI-powered browsers. The research showed that malicious instructions embedded in content can manipulate browser agents into unsafe actions.

  3. Oct 27, 2025

    LayerX identifies CSRF-based memory injection flaw in ChatGPT Atlas

    Researchers at LayerX Security discovered a critical vulnerability in OpenAI's ChatGPT Atlas browser that lets attackers use cross-site request forgery to inject malicious instructions into the assistant's persistent memory. The attack can be triggered by luring a logged-in user to a malicious link and may enable arbitrary code execution, privilege escalation, malware deployment, and cross-device persistence.

  4. Feb 1, 2024

    OpenAI introduces ChatGPT memory feature

    OpenAI introduced ChatGPT's persistent memory feature, designed to personalize user experiences across sessions. Later research identified this feature as a key component that could be abused for persistent compromise.

  5. May 8, 2023

    Researchers disclose environment-injected memory poisoning attack on web agents

    Researchers introduced Environment-injected Trajectory-based Agent Memory Poisoning (eTAMP), a technique that poisons an LLM web agent's persistent memory through manipulated environmental observations rather than direct memory access. The study showed a single poisoned observation could persist across sessions and sites, with measurable success rates against multiple models and increased effectiveness under 'Frustration Exploitation' conditions.

See the full picture in Mallory

Mallory subscribers get deeper analysis on every story, including:

Impact Assessment

Who’s affected and how

Technical Details

Deep-dive technical analysis

Response Recommendations

Actionable next steps for your team

Indicators of Compromise

IPs, domains, hashes, and more

AI Threads

Ask questions and take action on every story

Advanced Filters

Filter by topic, classification, timeframe

Scheduled Alerts

Get matching stories delivered automatically

Related Stories

AI Prompt Injection and Data Leakage Vulnerabilities in OpenAI's ChatGPT and Atlas Browser

AI Prompt Injection and Data Leakage Vulnerabilities in OpenAI's ChatGPT and Atlas Browser

Tenable Research has identified seven novel vulnerabilities and attack techniques in OpenAI's ChatGPT, including indirect prompt injections, exfiltration of user data, and bypasses of safety mechanisms in the latest GPT-5 model. These vulnerabilities allow attackers to manipulate the large language model (LLM) through crafted inputs, potentially leading to the theft of private information from user memories and chat histories, even when users simply interact with ChatGPT. The research highlights that hundreds of millions of users could be at risk, as attackers can exploit these weaknesses to bypass safeguards and extract sensitive data without user awareness. The release of OpenAI's ChatGPT Atlas, an AI-powered browser that remembers user activities and acts autonomously, further amplifies these concerns. Security experts warn that features such as persistent memory and autonomous actions increase the attack surface, making the browser susceptible to prompt injection and other AI-specific vulnerabilities. The implications for enterprise security and privacy are significant, as these AI-driven tools become more integrated into business processes, necessitating new approaches to identity management, access controls, and oversight to mitigate the risks posed by advanced AI-enabled attacks.

1 months ago
Prompt Injection and Browser-Based AI Security Risks

Prompt Injection and Browser-Based AI Security Risks

The launch of ChatGPT Atlas, an AI-powered web browser with agentic capabilities, has raised significant concerns about prompt injection attacks. As browsers become more integrated with large language models (LLMs), attackers can exploit both direct and indirect prompt injection techniques to manipulate AI agents, potentially causing them to divulge sensitive information or perform unintended actions. The accessibility of such agentic browsers, combined with their ability to automate complex tasks, amplifies the risk landscape for organizations adopting these technologies. Security experts warn that the browser now represents a critical control point for AI security, as it serves as the main interface between users and generative AI systems. The rapid increase in GenAI browser traffic has led to a surge in data security incidents, including inadvertent exposure of confidential information through LLM prompts. Traditional network security measures are often insufficient to address these browser-borne threats, making it imperative for organizations to reassess their security strategies and implement controls specifically designed to mitigate risks associated with AI-powered browsers and prompt injection attacks.

1 months ago
Prompt Injection Risks in Agentic AI and AI-Powered Browsers

Prompt Injection Risks in Agentic AI and AI-Powered Browsers

Security researchers reported that **prompt injection** is enabling practical attacks against *agentic AI* systems that have access to tools and user data, and argued the industry is underestimating the threat. A proposed framing, **“promptware,”** describes malicious prompts as a malware-like execution mechanism that can drive an LLM to take actions via its connected tools—potentially leading to **data exfiltration**, cross-system propagation, IoT manipulation, or even **arbitrary code execution**, depending on the permissions and integrations available. Trail of Bits disclosed results from an adversarial security assessment of Perplexity’s *Comet* browser, showing how prompt injection techniques could be used to **extract private information from authenticated sessions (e.g., Gmail)** by abusing the browser’s AI assistant and its tool access (such as reading page content, using browsing history, and interacting with the browser). Their threat-model-driven testing emphasized that agentic assistants can treat external web content as instructions unless it is explicitly handled as **untrusted input**, and they published recommendations intended to reduce prompt-injection-driven data paths between the user’s local trust zone (profiles/cookies/history) and vendor-hosted agent/chat services.

1 months ago

Get Ahead of Threats Like This

Mallory continuously monitors global threat intelligence and correlates it with your attack surface. Know if you're exposed. Before adversaries strike.