In the spring of 2025, security researchers at Varonis discovered a vulnerability in Microsoft 365 Copilot that had been sitting inside millions of enterprise deployments undetected. The vulnerability, later assigned CVE-2025-32711 and named “EchoLeak” by its discoverers, allowed an attacker to exfiltrate confidential corporate documents from a target's Microsoft 365 environment with a single click. No malware. No credential theft. No alert in the organisation's EDR, SIEM, or data loss prevention stack. The attack used Microsoft Copilot exactly as it was designed to be used. That was the problem.
What Happened: CVE-2025-32711
Microsoft 365 Copilot is an AI assistant deeply integrated into the Microsoft 365 suite: Outlook, Teams, Word, SharePoint. It has read access to a user's emails, documents, calendar, and communication history, and it can act on that content: summarising, drafting responses, extracting information, and sending it to other surfaces within the Microsoft ecosystem.
CVE-2025-32711 exploited a structural property of how Copilot processed incoming content. An attacker who could cause a target user to receive a maliciously crafted email or a crafted document could embed instructions in that content that Copilot would interpret as legitimate directives. Those instructions could instruct Copilot to locate documents with specific sensitivity labels, extract their contents, and transmit them to an attacker-controlled destination, all within the context of the user's existing authorised session.
Microsoft confirmed the vulnerability and patched it. The disclosure, covered by Tom's Guide and independently verified by Varonis researchers, confirmed that confidentiality sensitivity labels — the Microsoft Purview controls that enterprises rely on to protect classified documents, were bypassed entirely. Copilot simply read the labelled content and acted on its instructions.
The Varonis team also documented a variant they called “Reprompt”: a single-click attack in which a user who clicked an attacker-crafted link unknowingly triggered a Copilot session that exfiltrated targeted data in real time. The entire operation completed in seconds and left no artefact that conventional DLP or security monitoring would have flagged as anomalous.
The Attack Chain
The EchoLeak attack chain followed a sequence that will be recognisable to anyone familiar with indirect prompt injection as a technique — but its execution against a production enterprise AI assistant at this scale was a first.
- Stage 1 (Delivery): The attacker crafts a malicious email containing hidden instructions formatted as natural language directives. The email appears legitimate: a meeting request, a document notification, a forwarded thread. The hidden instructions are not visible in the email body rendered to the user.
- Stage 2 (Processing): The target's Copilot processes the email as part of its normal context ingestion. Copilot has no mechanism to distinguish between content it should process semantically and content that contains adversarial directives, because from the model's perspective, both are just content.
- Stage 3 (Exfiltration): Following the injected instructions, Copilot queries the user's SharePoint environment for documents matching specified criteria (sensitivity label, keywords, folder path). It extracts the content and transmits it using the user's legitimate session credentials, via authorised Microsoft APIs, to a destination that may appear entirely benign.
- Stage 4 (Clean exit): The operation completes within the user's normal Copilot interaction window. No unusual process spawned. No network connection to a known malicious IP. No DLP rule triggered. From every monitoring perspective, Copilot did exactly what it does every day.
“The attack uses the AI assistant as the delivery mechanism, the processing engine, and the exfiltration channel simultaneously. There is no separate malware component to detect. The assistant is the threat.”
Why Your Security Stack Didn't Catch It
Understanding why EchoLeak bypassed conventional security controls requires understanding what those controls are designed to detect and what they are not.
Endpoint Detection and Response (EDR) tools monitor process behaviour, file system access, and network connections for patterns consistent with known malware. The EchoLeak attack involved no new process, no file dropped to disk, and no network connection to a classified malicious domain. Copilot ran under its normal service context. Every API call it made was authorised. The EDR saw a legitimate enterprise application operating within its expected parameters.
Data Loss Prevention (DLP) tools are typically configured to inspect outbound data transfers for sensitive content patterns such as specific document types, credit card numbers, PII markers. DLP operates on the content of the transfer, and in many configurations, on the destination. An exfiltration conducted via Microsoft's own API infrastructure, to a destination on the allowlist, using TLS encryption that DLP cannot inspect, does not trigger pattern-matching rules. The data left through an approved channel.
Microsoft Purview sensitivity labels, the classification system that marks documents as Confidential, Highly Confidential, or equivalent, are access controls. They govern who can read a document in normal user workflows. They do not constrain what an AI assistant running under a user's authorised session can process and relay. The label said the document was confidential. Copilot read it anyway, because the user's credentials authorised it to.
SIEM correlation rules look for anomalies: impossible travel, off-hours logins, volume spikes on access logs. A single Copilot query that retrieves three confidential documents during business hours, from a user who regularly accesses those documents, produces no anomaly signal. The volume is normal. The time is normal. The user is normal. The SIEM has nothing to correlate.
This Is Not an Isolated Incident
EchoLeak was patched. But the structural vulnerability it exposed, the susceptibility of AI assistants with broad data access to indirect prompt injection, was not patched. It cannot be patched, because it is a property of how large language models process content, not a bug in any specific implementation.
OWASP's LLM Top 10 v2.0, published in 2024, lists prompt injection (LLM01:2025) as the leading risk category for LLM applications, noting that “an LLM application is vulnerable to prompt injection whenever it processes content from untrusted sources in the same context as trusted instructions.” Microsoft 365 Copilot, by design, processes untrusted content (incoming email from anyone on the internet) in the same context as trusted instructions (the user's operational directives). That combination will continue to be exploitable.
The HackTheBox Security Team, in their independent analysis of CVE-2025-32711, noted that the fundamental attack class will persist as long as enterprise AI assistants are given broad access to corporate data stores and the ability to act on that data autonomously. Patching specific injection vectors is a maintenance cycle, not a solution.
The enterprise deployment reality makes this more acute: Microsoft 365 Copilot is deployed in organisations that measure their user bases in the hundreds of thousands. It has access to inboxes, document libraries, and communication records that represent the operational intelligence of the entire organisation. The attack surface is not a niche integration. It is the centre of enterprise productivity.
What Detection Would Have Looked Like
EchoLeak was discovered by researchers auditing Copilot's behaviour — not by enterprise security tooling that detected it in the wild. That distinction matters. It suggests that the vulnerability was likely exploited in real environments before it was patched, and that those exploitations produced no alert.
Detection of EchoLeak-class attacks requires observing what the AI assistant actually does with the access it has been granted — not just whether it was authorised to have that access. The detection surface is the agent's behaviour: the sequence of API calls it makes, the data volumes it accesses, the endpoints it transmits to, and the relationship between what it was instructed to do and what it actually executed.
Specifically, detecting this attack class requires:
- Monitoring the semantic chain from incoming content (email, document) to subsequent agent action (document query, data retrieval, outbound transmission). A normal Copilot session does not retrieve three confidential documents in response to an incoming email from an external sender.
- API-level intent analysis at the boundary between the AI assistant and the data stores it accesses. The call to retrieve labelled confidential documents, in context with the preceding email ingestion, is the signal.
- Correlation across the full chain, not just point-in-time anomaly detection on individual API calls, but the relationship between the trigger (the crafted email), the intermediate action (document retrieval), and the outcome (outbound transmission). No single step looks anomalous in isolation. The chain does.
This is precisely the detection architecture Helixar was built to provide. Helixar Shield monitors API boundaries for intent signals that diverge from legitimate operational patterns. Helixar Vigil captures the behavioural sequence at the endpoint It correlates what prompted an action with what that action did. Nexus surfaces the full chain (trigger, execution, outcome) to a human operator, with the context required to make a rapid, informed response decision before the data leaves the environment.
EchoLeak represents a class of threat for which the detection must happen during the chain, not after it. The window between the crafted email arriving and the exfiltration completing is measured in seconds. At that timescale, the only viable response is automated detection with immediate human escalation, not a SIEM alert reviewed in the next morning's triage queue.
The Broader Implication for Enterprise AI Governance
EchoLeak illustrates a governance gap that extends beyond any single vendor or product. Every enterprise AI assistant, every tool that ingests external content and acts on it with broad data access, carries a version of this structural risk. The question is not whether the specific CVE has been patched. The question is whether the enterprise has the detection capability to know when its AI assistants are behaving in ways that were not intended.
The Microsoft 365 Copilot deployment is, in most enterprises, entirely invisible to conventional security tooling at the behavioural level. It does not show up in EDR console as a process worth monitoring. It does not produce logs that SIEM rules were designed to correlate. It operates in the gap between endpoint security and application security, a gap that predates agentic AI and has become catastrophically wider since.
Security teams that have deployed enterprise AI assistants without a dedicated detection layer for agentic behaviour are operating with a known blind spot. EchoLeak named it. The next exploit in the same class may not be patched before it is used.
References
- HackTheBox Security Team. (2025). CVE-2025-32711: Microsoft Copilot EchoLeak Vulnerability Analysis. HackTheBox Blog. hackthebox.com
- Tom's Guide. (2025). Microsoft confirms nasty Copilot bug that could have let hackers steal your data. Tom's Guide, June 2025. tomsguide.com
- Varonis Threat Labs. (2025). EchoLeak: Microsoft 365 Copilot Prompt Injection and Data Exfiltration. Varonis Blog. varonis.com
- OWASP. (2025). OWASP Top 10 for LLM Applications: LLM01:2025 Prompt Injection. owasp.org.
- Microsoft Security Response Center. (2025). CVE-2025-32711 — Microsoft Copilot Information Disclosure Vulnerability. MSRC Advisory.