Skip to content
Threat Feed
high advisory

Securing AI Agents with CrowdStrike Falcon AIDR and NVIDIA NeMo Guardrails

CrowdStrike Falcon AIDR now supports NVIDIA NeMo Guardrails to protect AI agents from attacks like prompt injection, data exfiltration, and unauthorized actions, enabling organizations to deploy AI applications more securely.

The increasing adoption of AI agents in enterprise environments presents new security challenges. Attackers are developing techniques to compromise these agents, leading to data breaches, unauthorized transactions, and compliance violations. CrowdStrike Falcon AIDR, with the integration of NVIDIA NeMo Guardrails (version 0.20.0), offers enterprise-grade protection for AI agents. This integration allows organizations to define and enforce guardrails, manage data access, control agent responses, and ensure policy compliance. By blocking prompt injection attacks, redacting sensitive data, defanging malicious content, and moderating unwanted topics, Falcon AIDR enhances the security and control of AI agents in production environments. This combined solution aims to address the risks associated with AI agents operating autonomously across sensitive business processes.

Attack Chain

  1. Initial Access: An attacker crafts a malicious prompt designed to exploit vulnerabilities in the AI agent’s input processing.
  2. Prompt Injection: The attacker injects the malicious prompt into the AI agent’s input stream, bypassing initial input validation checks.
  3. Agent Manipulation: The injected prompt manipulates the agent’s behavior, causing it to deviate from its intended functionality.
  4. Data Access: The compromised agent, under the attacker’s control, accesses sensitive data, such as customer PII, financial records, or internal code repositories.
  5. Unauthorized Actions: The agent executes unauthorized actions, such as initiating fraudulent transactions, modifying system configurations, or disclosing confidential information.
  6. Lateral Movement: The attacker uses the compromised agent to access other systems or data sources within the organization.
  7. Data Exfiltration: The attacker extracts sensitive data from the compromised systems and exfiltrates it to an external location.
  8. Impact: The organization suffers financial losses, reputational damage, and legal repercussions due to the data breach and unauthorized actions.

Impact

A successful attack on an AI agent can lead to significant consequences. This includes exposure of customer data, unauthorized transactions, and violations of compliance requirements. The number of potential victims scales with the agent’s deployment size. Organizations in financial services, healthcare, customer service, and software development are particularly vulnerable. The damage can range from financial losses and reputational damage to legal repercussions and loss of customer trust. The risk grows as more organizations adopt AI and the number of vulnerable AI agents increases.

Recommendation

  • Deploy CrowdStrike Falcon AIDR with NVIDIA NeMo Guardrails (v0.20.0) to protect AI agents from runtime attacks and reduce the agentic blast radius.
  • Create named detection policies tailored to specific security requirements using the Falcon AIDR API.
  • Enable detectors to detect, block, redact, encrypt, or transform content at critical points in AI agent workflows as mentioned in the overview.
  • Implement the Sigma rule “Detect Suspicious Prompt Injection Attempts” to identify and block malicious prompts attempting to manipulate AI agent behavior.
  • Monitor AI agent activity logs for suspicious patterns and anomalies, leveraging the insights from CrowdStrike Falcon AIDR.
  • Deploy the Sigma rule “Detect Sensitive Data Exposure by AI Agents” to identify and prevent the exfiltration of sensitive information by compromised agents.

Detection coverage 2

Detect Suspicious Prompt Injection Attempts

high

Detects suspicious prompts containing characters or patterns indicative of prompt injection attacks targeting AI agents.

sigma tactics: defense_evasion techniques: T1068 sources: webserver, linux

Detect Sensitive Data Exposure by AI Agents

medium

Detects AI agent responses containing sensitive data patterns, indicating potential data leakage.

sigma tactics: exfiltration sources: webserver, linux

Detection queries are kept inside the platform. Get full rules →