A critical architectural vulnerability embedded in Anthropic’s Model Context Protocol (MCP), the industry standard for AI agent communication, has been uncovered.
Placing more than 150 million downloads and up to 200,000 servers at risk of complete system takeover through Remote Code Execution (RCE).
The Model Context Protocol (MCP) serves as the backbone infrastructure connecting AI models to external tools, databases, APIs, and automated workflows across the modern AI development ecosystem.
Anthropic’s MCP Protocol Flaw
Created and actively maintained by Anthropic, MCP has been adopted globally across Python, TypeScript, Java, and Rust, making it a foundational standard that virtually every major AI agent framework depends on. Its ubiquity is precisely what makes this vulnerability so catastrophic in scale.
According to security research, the root cause is not a traditional coding bug but a deliberate architectural design decision. MCP’s STDIO transport accepts arbitrary command strings. It passes them directly to subprocess execution, an “execute-first, validate-never” paradigm that fundamentally undermines the security of every downstream implementation built on it.
10 CVEs, Six Live Exploits
Security research team identified four distinct families of exploitation, each capable of triggering full system compromise:
- Unauthenticated UI Injection in popular AI frameworks, requiring no credentials to execute
- Hardening Bypasses in environments previously considered “protected,” such as Flowise
- Zero-Click Prompt Injection in leading AI IDEs, including Windsurf and Cursor, where a malicious document retrieval alone can trigger remote code execution
- Malicious Marketplace Distribution, where researchers successfully “poisoned” 9 out of 11 MCP registries with a proof-of-concept malicious MCP entry
The research resulted in 10 CVEs issued across critical AI platforms, with the successful execution of arbitrary commands on six live production environments.
Affected platforms include widely-used tools such as LiteLLM, LangChain, IBM’s LangFlow, and others central to enterprise AI development pipelines.
| CVE ID | Product | Attack Vector | Severity | Status |
|---|---|---|---|---|
| CVE-2025-65720 | GPT Researcher | UI injection / reverse shell | Critical | Reported |
| CVE-2026-30623 | LiteLLM | Authenticated RCE via JSON config | Critical | Patched |
| CVE-2026-30624 | Agent Zero | Unauthenticated UI injection | Critical | Reported |
| CVE-2026-30618 | Fay Framework | Unauthenticated Web-GUI RCE | Critical | Reported |
| CVE-2026-33224 | Bisheng | Authenticated UI injection | Critical | Patched |
| CVE-2026-30617 | LangChain-Chatchat | Unauthenticated UI injection | Critical | Reported |
| CVE-2026-30615 | Windsurf | Zero-click prompt injection to local RCE | Critical | Reported |
| CVE-2026-26015 | DocsGPT | MITM transport-type substitution | Critical | Patched |
| CVE-2026-30625 | Upsonic | Allowlist bypass via npx/npm args | High | Warning |
| CVE-2026-33224 | Jaaz | Unauthenticated UI injection | Critical | Reported |
The zero-click exploit against Windsurf is particularly alarming. Researchers confirmed that when a malicious document is retrieved in an affected AI IDE, it can automatically execute code that harvests environment variables, SSH keys, AWS credentials, and Git tokens, while simultaneously installing a persistent reverse shell.
Anthropic Declined to Fix the Root Cause
Security team engaged in responsible disclosure, repeatedly recommending root-level architectural patches to Anthropic that would have instantly protected millions of downstream users.
Anthropic declined, characterizing the behavior as “expected” and consistent with design intent. The company did not object when the security team notified them of the intent to publish the findings.
This stance has drawn significant criticism from the security community. Researchers argue Anthropic has “the ability and responsibility to make MCP secure by default,” and that a single architectural change at the protocol level could neutralize the entire vulnerability class.
This refusal becomes even more notable given that Anthropic recently unveiled Claude Mythos, a new initiative aimed at helping secure the world’s software. Critics argue the company should apply that same “Secure by Design” commitment to its own protocol architecture.
This is not the first time MCP’s security posture has been questioned. As early as January 2026, researcher Yarden Porat published an exploit chain targeting Anthropic’s official Git MCP server, achieving RCE via prompt injection across three CVEs: CVE-2025-68143, CVE-2025-68144, and CVE-2025-68145.
These prior disclosures make Anthropic’s current inaction regarding systemic architectural reform increasingly difficult to justify.
Defenders Must Do Now
Organizations running MCP-enabled services should take immediate action across several fronts:
- Block public IP access to all LLM and AI enabler services, never expose them directly to the internet
- Treat all external MCP configuration input as untrusted user input flowing into StdioServerParameters, which must be blocked or strictly pre-validated
- Install MCP servers only from verified sources such as the official GitHub MCP Registry to avoid typosquatting and malicious marketplace entries.
- Run MCP-enabled services in sandboxes with restricted permissions, eliminating full-disk access and shell execution privileges when not absolutely required.
- Monitor all tool invocations made by AI agents, flagging any background activity or attempts to reach unknown external URLs
- Upgrade all affected services immediately if no fixed version is available, and turn off user input until a patch is released.
OX Security has disclosed platform-level protections for its own customers, including detection of improper STDIO-based MCP configurations in AI-generated code and flagging of existing vulnerable configurations in customer codebases.
FAQ
Q1: What is the MCP vulnerability discovered by OX Security?
It is an architectural design flaw in Anthropic’s STDIO transport layer that allows arbitrary user input to be passed directly to subprocess execution, enabling unauthenticated Remote Code Execution.
Q2: How many systems are affected by this MCP design flaw?
The vulnerability spans 150M+ downloads, 7,000+ publicly accessible servers, and up to 200,000 total vulnerable instances across the global AI development ecosystem.
Q3: Why hasn’t Anthropic patched the root cause of this vulnerability?
Anthropic declined to modify the protocol’s architecture after OX Security’s responsible disclosure, stating the behavior is “expected” per the protocol’s original design intent.
Q4: Which AI platforms are confirmed vulnerable to MCP-based RCE attacks?
Confirmed vulnerable platforms include LiteLLM, Agent Zero, Windsurf, DocsGPT, GPT Researcher, LangChain-Chatchat, Fay Framework, Bisheng, Jaaz, Upsonic, and IBM’s LangFlow.
Site: http://thecybrdef.com