Skip to content
Threat Feed
medium advisory

Securing AI Agents with CrowdStrike Falcon AIDR and NVIDIA NeMo Guardrails

CrowdStrike Falcon AIDR integrates with NVIDIA NeMo Guardrails to provide comprehensive protection for AI agents against prompt injection, data leaks, and malicious content.

The increasing adoption of AI agents in mainstream business tools presents new security challenges. A compromised agent can lead to data exposure, unauthorized transactions, and compliance violations. To address these risks, CrowdStrike Falcon AIDR now supports NVIDIA NeMo Guardrails. This integration provides enterprise-grade protection by defining guardrails and applying constraints on LLMs. NVIDIA NeMo Guardrails, an open-source library, offers features like content safety, PII detection, jailbreak detection, and topic control. Falcon AIDR and NeMo Guardrails enable developers to manage data access, control agent responses, and ensure policy compliance, facilitating the secure transition of AI agents from development to production. This solution helps organizations maintain visibility and control over their AI agents.

Attack Chain

  1. Initial Access: An attacker crafts a malicious prompt to interact with an AI agent.
  2. Prompt Injection: The malicious prompt injects unintended commands or instructions into the agent’s processing flow.
  3. Bypass Guardrails (Attempt): The attacker attempts to bypass existing guardrails using sophisticated injection techniques.
  4. Data Exfiltration: If successful, the attacker exploits the agent to access and exfiltrate sensitive data (e.g., customer PII, internal documents).
  5. Unauthorized Actions: The attacker manipulates the agent to perform unauthorized actions, such as initiating fraudulent transactions or modifying configurations.
  6. Lateral Movement (Potential): In some scenarios, a compromised agent could be leveraged to access other systems or data sources within the organization’s environment.
  7. Compliance Violation: The agent’s actions result in violations of regulatory compliance requirements (e.g., HIPAA, GDPR).
  8. Impact: Data breach, financial loss, reputational damage, and legal penalties.

Impact

A successful attack against an AI agent can have significant consequences. Data breaches exposing customer PII, unauthorized transactions leading to financial losses, and compliance violations resulting in legal penalties are all potential outcomes. The impact spans across various sectors, including financial services, healthcare, and customer service, where AI agents handle sensitive data and critical business processes. The extent of the damage depends on the agent’s access privileges and the sensitivity of the data it handles. Even a single compromised agent can expose thousands of interactions, amplifying the blast radius of an attack.

Recommendation

  • Deploy Falcon AIDR with NVIDIA NeMo Guardrails to enforce content safety, PII protection, and jailbreak detection (see Overview).
  • Implement custom data classification rules in Falcon AIDR to align with your organization’s specific data protection requirements (see Overview).
  • Enable monitoring mode in Falcon AIDR to understand the threat landscape and progressively enforce blocks and redactions as agents move from development to production (see Use Cases).
  • Create named detection policies in Falcon AIDR tailored to specific security requirements at critical points in AI agent workflows (see Configuring Falcon AIDR Policies).
  • Monitor web server logs for unexpected HTTP requests that might indicate prompt injection attempts targeting AI agents (see rules).

Detection coverage 2

Detect Suspicious Keywords in HTTP Requests to AI Agents

medium

Detects potential prompt injection attempts by looking for suspicious keywords in HTTP requests.

sigma tactics: defense_evasion, initial_access techniques: T1566.001 sources: webserver, linux

Detect Blocked Content by Falcon AIDR

info

Detects instances where Falcon AIDR blocks potentially malicious content based on its classification rules.

sigma tactics: defense_evasion sources: webserver, linux

Detection queries are kept inside the platform. Get full rules →