<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Ai-Jailbreak — CraftedSignal Threat Feed</title><link>https://feed.craftedsignal.io/tags/ai-jailbreak/</link><description>Trending threats, MITRE ATT&amp;CK coverage, and detection metadata — refreshed continuously.</description><generator>Hugo</generator><language>en</language><managingEditor>hello@craftedsignal.io</managingEditor><webMaster>hello@craftedsignal.io</webMaster><lastBuildDate>Wed, 03 Jan 2024 12:00:00 +0000</lastBuildDate><atom:link href="https://feed.craftedsignal.io/tags/ai-jailbreak/feed.xml" rel="self" type="application/rss+xml"/><item><title>Microsoft 365 Copilot Jailbreak Attempts via Prompt Injection</title><link>https://feed.craftedsignal.io/briefs/2024-01-03-m365-copilot-jailbreak/</link><pubDate>Wed, 03 Jan 2024 12:00:00 +0000</pubDate><author>hello@craftedsignal.io</author><guid isPermaLink="true">https://feed.craftedsignal.io/briefs/2024-01-03-m365-copilot-jailbreak/</guid><description>The detection identifies attempts to jailbreak Microsoft 365 Copilot through prompt injection techniques that attempt to circumvent built-in safety controls by manipulating rules, bypassing system commands, or requesting AI impersonation.</description><content:encoded><![CDATA[<p>Microsoft 365 Copilot is susceptible to jailbreak attempts via prompt injection, where users craft specific prompts designed to bypass or override safety controls. These attacks involve injecting malicious instructions into user prompts to manipulate the AI&rsquo;s behavior, potentially leading to the disclosure of sensitive information, the generation of harmful content, or the execution of unauthorized actions. The attacks leverage techniques like rule manipulation, system bypass commands, and AI impersonation requests, attempting to circumvent built-in safety mechanisms. Successful jailbreaks can compromise the integrity and security of Copilot, enabling threat actors to exploit the AI for malicious purposes.</p>
<h2 id="attack-chain">Attack Chain</h2>
<ol>
<li>An attacker crafts a malicious prompt containing specific keywords and phrases designed to manipulate Copilot&rsquo;s behavior.</li>
<li>The attacker injects the prompt into M365 Copilot through a standard user interface, like a chat window.</li>
<li>Copilot processes the prompt, attempting to interpret the user&rsquo;s intent.</li>
<li>If the prompt is successfully injected, Copilot&rsquo;s safety controls are bypassed or overridden due to prompt injection techniques.</li>
<li>Copilot generates a response based on the manipulated instructions in the prompt, potentially providing unauthorized access to information or functionality.</li>
<li>The attacker exfiltrates sensitive data or uses Copilot to perform actions outside its intended scope.</li>
<li>The attacker leverages the compromised Copilot to create and disseminate malicious content.</li>
</ol>
<h2 id="impact">Impact</h2>
<p>Successful jailbreak attempts can lead to the disclosure of sensitive company data, generation of harmful or inappropriate content, and circumvention of organizational security policies. A single successful jailbreak can affect multiple users if the generated content is shared. If successful, internal copilots could be used to create phishing messages or generate code that gives the attacker a reverse shell on a machine. The risk is increased due to the widespread adoption of M365 Copilot across various industries.</p>
<h2 id="recommendation">Recommendation</h2>
<ul>
<li>Enable M365 Exported eDiscovery Prompts logging to capture user interactions with Copilot, as this log source is crucial for detecting jailbreak attempts.</li>
<li>Deploy the Sigma rules provided in this brief to your SIEM to identify potential jailbreak attempts based on suspicious keywords and patterns in user prompts.</li>
<li>Implement filtering mechanisms based on the <code>m365_copilot_jailbreak_attempts_filter</code> macro to reduce false positives and focus on high-risk activities.</li>
<li>Monitor the <code>Subject_Title</code> field in the M365 eDiscovery prompt logs for the presence of jailbreak keywords and phrases such as &ldquo;act as,&rdquo; &ldquo;bypass,&rdquo; &ldquo;ignore,&rdquo; &ldquo;override,&rdquo; &ldquo;pretend you are,&rdquo; and &ldquo;rules=&rdquo;.</li>
<li>Investigate and remediate any identified jailbreak attempts to prevent further exploitation of M365 Copilot.</li>
</ul>
]]></content:encoded><category domain="severity">high</category><category domain="type">advisory</category><category>prompt-injection</category><category>ai-jailbreak</category><category>m365</category><category>copilot</category></item></channel></rss>