All articles
Threat IntelligenceMarch 2026·11 min read

Your Developer's AI Copilot Is the New Attack Surface: Supply Chain Risks in AI-Native Development Tools

GitHub Copilot, Cursor, and Claude Code all suffered critical vulnerabilities in 2025–2026. Attackers need no malware and no credentials. They need a prompt.

A developer clones a repository. The repository contains a README file with a comment that looks like documentation. The developer's AI coding assistant reads the comment to understand the project context. Thirty seconds later, a reverse shell opens on the developer's workstation, and the attacker has access to every environment variable, API key, and credential the developer had stored locally. No exploit kit. No phishing link clicked. No malware installed. Just a prompt, processed by a trusted tool, that the developer never read.

This is not a hypothetical scenario. Variants of this attack chain were demonstrated against GitHub Copilot in 2025, and corresponding vulnerabilities were confirmed in Claude Code and Cursor. The attack surface is not the developer's browser, email client, or operating system. It is the AI coding assistant that has become the centre of the modern development workflow.

The GitHub Copilot RCE Chain: CVE-2025-53773

In 2025, researcher Johann Rehberger, known publicly as “Embrace The Red”, documented a three-stage remote code execution chain against GitHub Copilot that required only a malicious comment in a repository the developer visited.

The vulnerability, assigned CVE-2025-53773, exploited the way Copilot's VS Code extension processed contextual codebase information. When a developer opened a project, Copilot ingested repository content to understand the codebase, including comments, documentation strings, and README files. A specially crafted comment could inject instructions that Copilot would interpret as operational directives, causing it to invoke IDE tooling (such as VS Code's built-in terminal) in ways the developer had not authorised.

As documented by SecurityWeek and confirmed by GitHub, the attack chain proceeded in three stages:

  • Stage 1 (Injection): Malicious prompt instructions embedded in a repository file (comment, documentation, or configuration). The instructions are invisible in the rendered interface; they appear as normal text to Copilot's context window.
  • Stage 2 (Tool invocation): Copilot processes the injected instructions and invokes VS Code's integrated tools, including terminal execution, file access, and extension APIs, under the assumption these are legitimate developer requests.
  • Stage 3 (Compromise): The terminal command executes under the developer's user context, with access to all local credentials, environment variables, SSH keys, and API tokens. Repository takeover, credential exfiltration, and further lateral movement become possible from this foothold.

GitHub patched the specific vulnerability, but the research community noted, as did Embrace The Red in the original disclosure, that the underlying attack class (prompt injection via repository content into agentic IDE tooling) remained viable and would require ongoing vigilance to contain.

The MCP Supply Chain Problem

The Model Context Protocol (MCP), introduced by Anthropic in late 2024, rapidly became the dominant standard for extending AI agents with external tool access, connecting AI coding assistants to databases, APIs, file systems, and developer services. Its adoption was fast and broad. Its security scrutiny was not.

In September 2025, security researchers discovered a malicious MCP server published to the npm registry that had accumulated several hundred downloads before detection. The package presented itself as a productivity integration for a popular developer service. In practice, it exfiltrated developer credentials and local environment variables to an attacker-controlled endpoint, operating silently within the context of the AI agent's normal tool-use workflow.

In January 2026, the security firm Pillar Security published a disclosure of three separate vulnerabilities in Anthropic's official MCP git server. The vulnerabilities allowed an attacker who could influence the content of a git repository to manipulate the MCP server's tool execution, potentially gaining arbitrary command execution on the developer's machine.

Check Point Research documented a further attack vector: AI agents connected to Supabase via MCP could be manipulated through crafted SQL content to execute unintended queries, leak database schema information, and in some configurations, exfiltrate table contents to an attacker-controlled surface. The attack required no Supabase credentials, only control over content that the AI agent would process in the context of a legitimate developer session.

“The MCP ecosystem grew faster than its security model could accommodate. Every new MCP server is a potential supply chain insertion point. The developer who installs one is extending trust to it at the level of their entire agentic development environment.”

Claude Code and Cursor: CVE-2025-59536

The vulnerability pattern was not limited to GitHub Copilot. In 2025, CVE-2025-59536 affecting both Claude Code (Anthropic's terminal-based coding agent) and the Cursor IDE, was assigned a CVSS score of 8.7, indicating high severity.

As documented by The Hacker News, the vulnerability allowed a maliciously crafted project file to cause the AI coding assistant to exfiltrate the developer's API tokens, including the API keys used to authenticate with AI services, to an attacker-controlled server. In the case of Claude Code, this meant the attacker could obtain the developer's Anthropic API key and use it to run arbitrary model queries at the developer's expense and potentially access their usage history.

The attack required only that the developer open a project containing the malicious file. This is a low bar given that developers routinely clone repositories from public sources, open projects sent by colleagues, or evaluate open-source codebases for potential use. The exposure surface was, in practice, every project the developer touched.

All of these attacks shared a structural characteristic: they operated entirely under the developer's user context, using legitimate tooling, producing no artefact that traditional endpoint security would classify as malicious. The attacker never needed to compromise the developer's machine in the conventional sense. They compromised the developer's AI assistant, and the AI assistant did the rest.

What These Attacks Have in Common

Across every variant documented in this wave of AI developer tool vulnerabilities, four structural characteristics are consistent:

  • Legitimate tooling, hostile payload: The attack uses tools the developer has deliberately installed and trusts. There is no need to introduce foreign binaries or exploit OS-level vulnerabilities. The AI assistant and its MCP integrations are the attack vector.
  • Authorised credentials, unauthorised use: The exfiltrated credentials are the developer's own: API keys, tokens, and SSH keys, accessed through legitimate process context. The attacker obtains them without ever authenticating to any system as the attacker.
  • No network anomaly: The exfiltration typically occurs over HTTPS to a destination that may appear legitimate (a CDN, a known service endpoint, an infrastructure provider). There is no connection to a known malicious IP, no unusual protocol usage, no traffic pattern that network security tools were designed to flag.
  • No endpoint artefact: No file is dropped to disk that would trigger antivirus. No process spawns in a way that EDR would classify as a known attack pattern. The developer's security tooling sees a coding assistant doing what coding assistants do.

The Detection Gap in Developer Environments

Enterprise security teams have historically applied relatively limited endpoint monitoring to developer workstations, on the basis that the diversity of tooling and the complexity of legitimate development workflows makes false positive rates prohibitively high. An EDR alert on every terminal command a developer executes would be operationally unworkable. This has created a category of endpoint that is both highly privileged (developers typically have significant system access, broad API credentials, and access to production infrastructure) and lightly monitored.

AI coding assistants have significantly increased the risk profile of that gap. They operate continuously, processing large volumes of content from potentially untrusted sources (cloned repositories, external documentation, user-provided context). They invoke tools and execute actions (terminal commands, file reads, API calls) on behalf of the developer at a frequency and scale that makes manual review impossible. And they do all of this with the same permissions as the developer they are assisting.

The security monitoring approach must adapt accordingly. What is needed is not a higher volume of endpoint alerts. It is a detection layer that understands the normal behavioural patterns of agentic developer tooling and can identify when those patterns deviate in ways consistent with compromise.

Helixar Vigil's process chain monitoring captures the behavioural sequence at the developer endpoint, capturing the relationship between what prompted an action (repository content ingested by an AI assistant) and what that action did (terminal command executed, credential file accessed, network request made). An AI assistant that executes a terminal command immediately after ingesting a cloned repository's README file, then makes an outbound HTTPS request to an unfamiliar endpoint, presents a behavioural chain that deviates from normal development patterns, warranting human review before it completes.

Helixar Shield applies intent analysis at the API boundary: the difference between an AI agent calling a known development service API and an AI agent calling an endpoint that has no established relationship to the developer's normal toolset. That distinction, which network security tools cannot make, is the signal.

What Security Teams Should Do Now

The developer AI tooling attack surface is not a future concern. The vulnerabilities documented above were confirmed and exploited in 2025 and early 2026. Enterprise security teams deploying AI coding assistants at scale should be taking the following steps:

  • Audit installed MCP servers across the development fleet. Treat each MCP server as a supply chain component requiring the same security review as any third-party npm package. Many developers install MCP servers without security team awareness.
  • Establish a baseline of normal AI agent tool invocations for developer workstations. What does legitimate Copilot or Claude Code terminal access look like? What outbound endpoints does it contact? Deviations from that baseline are the detection signal.
  • Treat developer API tokens as high-value targets requiring monitoring.The CVE-2025-59536 class of attack specifically targets AI service API keys. Monitor for unexpected transmission of credential files or environment variables.
  • Review AI coding assistant permissions against the principle of least privilege. Does your deployed coding assistant need access to production infrastructure credentials? To SSH keys? To database connection strings? Most deployments grant significantly more access than the core development workflow requires.

References

  1. Rehberger, J. (“Embrace The Red”). (2025). GitHub Copilot: Indirect Prompt Injection Leading to Remote Code Execution. embracethered.com. embracethered.com
  2. SecurityWeek. (2025). GitHub Copilot Vulnerability Enabled Repository Takeover via Prompt Injection. SecurityWeek. securityweek.com
  3. GBHackers on Security. (2025). Malicious MCP Server on npm Exfiltrates Developer Credentials. GBHackers. gbhackers.com
  4. Pillar Security. (2026). Three Vulnerabilities in Anthropic's MCP Git Server. Pillar Security Research, January 2026. pillar.security
  5. The Hacker News. (2025). CVE-2025-59536: Claude Code and Cursor Vulnerability Allows API Token Exfiltration. The Hacker News. thehackernews.com
  6. Check Point Research. (2026). Agentic AI Supply Chain: MCP Server Vulnerabilities and Developer Risk. Check Point Research Blog. research.checkpoint.com
  7. OWASP. (2025). OWASP Agentic Applications Security Guide. owasp.org.

Detect what your developer tools can't

Book a walkthrough and see the Helixar detection layer in action.

Book a Demo