Skip to content
Threat Feed
high advisory

AI Agent Data Theft via Indirect Prompt Injection

Attackers are leveraging indirect prompt injection against AI agents with access to private data, untrusted content, and external communication channels to steal sensitive information by embedding malicious instructions in content processed by the agent.

AI agents are increasingly deployed in enterprises, often operating within a “lethal trifecta” of accessing private data, processing untrusted content, and communicating externally. This makes them vulnerable to data theft via indirect prompt injection attacks, where attackers embed malicious instructions in content the agent reads, such as emails or web pages. The agent then executes these instructions with the user’s privileges, without the user’s knowledge. Google’s April 2026 study found a 32% increase in malicious prompt injection attempts on public web pages between November 2025 and February 2026. Mainstream agent harnesses like Claude Code, Cursor, Hermes, GitHub Copilot Agent, and Gemini CLI currently lack robust architectural defenses like CaMeL and Dual LLM. Defenders should focus on blast radius containment by assuming breach of the LLM layer and implementing controls outside the model.

Attack Chain

  1. Attacker identifies a target organization using AI agents with access to sensitive data.
  2. Attacker crafts a malicious prompt injection payload disguised as harmless content (e.g., an email, document, or web page).
  3. Attacker delivers the crafted content to a system accessible to the AI agent (e.g., user’s inbox, internal file share, public website crawled by the agent).
  4. The AI agent processes the malicious content, interpreting the injected prompt as a legitimate instruction.
  5. The injected prompt instructs the AI agent to access sensitive data, such as internal documents, customer databases, or API keys.
  6. The injected prompt instructs the AI agent to exfiltrate the stolen data to an attacker-controlled server via HTTP/S or other channels.
  7. The AI agent executes the data exfiltration command, sending the sensitive data to the attacker.
  8. The attacker gains access to the stolen data, potentially leading to further malicious activities, such as extortion or espionage.

Impact

A successful indirect prompt injection attack can lead to significant data breaches, compromising sensitive internal documents, customer data, and proprietary information. This can result in financial losses, reputational damage, legal liabilities, and competitive disadvantage. The absence of widely adopted architectural defenses in current AI agent harnesses increases the risk of such attacks succeeding. As of May 2026, no high-profile enterprise-scale catastrophes have been reported, but the increasing prevalence of prompt injection attempts indicates a growing threat.

Recommendation

  • Enable and harden agent sandboxing where available to limit the blast radius of compromised AI agents, focusing on remote or cloud-based execution environments (reference: Pattern 1).
  • Treat memory writes by AI agents as security events and log them to detect potential backdoors created through poisoned memory entries (reference: “Memory is persistence”).
  • Implement strict credential isolation to prevent AI agents from directly accessing credentials, keeping them out of the LLM provider’s context window and the agent runtime (reference: “Credentials are the crown jewels”).
  • Deploy the Sigma rule “Detect AI Agent Memory Modification” to monitor for suspicious memory writes by AI agent processes (reference: rule below).
  • Evaluate and adopt architectural patterns like CaMeL and Dual LLM as they become available in mainstream agent harnesses to provide stronger defenses against prompt injection attacks (reference: Line 2).

Detection coverage 1

Detect AI Agent Memory Modification

medium

Detects suspicious memory writes by AI agent processes, which could indicate a compromised agent attempting to establish persistence.

sigma tactics: persistence techniques: T1546 sources: process_creation, windows

Detection queries are available on the platform. Get full rules →