OpenClaw Goes Viral, Exposing 12 Types of Critical Vulnerabilities; MCP Protocol Security Benchmark Released

marsbitPubblicato 2026-04-16Pubblicato ultima volta 2026-04-16

Introduzione

The rapid rise of OpenClaw and similar AI Agents highlights the growing security risks associated with the Model Context Protocol (MCP), a standard enabling models to interact with external tools. Researchers from Beijing University of Posts and Telecommunications introduced MSB (MCP Security Bench), a security benchmark identifying 12 types of attacks across MCP’s three core stages: task planning, tool invocation, and response handling. These include name collision, false errors, retrieval injection, and mixed attacks. Notably, more capable models are often more vulnerable, with an average attack success rate (ASR) of 40.35%. The study also proposes a new metric, Net Resilient Performance (NRP), to balance security and utility. MSB evaluates agents in real-world environments, demonstrating that attacks remain effective even when harmless tools are present. The work underscores the urgent need for robust safety measures as AI agents gain broader tool-use capabilities.

The MCP protocol is enabling AI Agents to autonomously execute tasks, but security risks are soaring. Research reveals that attackers can trick Agents into performing malicious operations through 12 methods, including tool name obfuscation and false errors, with even top-tier models falling victim. A team from Beijing University of Posts and Telecommunications has released the MSB security benchmark, which, through real-environment testing, demonstrates: the more powerful the model, the more vulnerable it is to attacks. The new NRP metric, for the first time, balances security and practicality, providing a crucial yardstick for fortifying AI Agent defenses.

Recently, open-source AI Agent projects like OpenClaw have exploded in popularity within the developer community. With just a single sentence, an Agent can automatically write code, research information, manipulate local files, and even take control of your computer.

The astonishing autonomy of these Agents is underpinned by the capabilities provided by tool calls, with MCP (Model Context Protocol) serving as the interface unifying the AI tool ecosystem. Just as USB-C allows computers to connect to various devices, MCP enables large language models (LLMs) to call external tools like file systems, browsers, and databases in a standardized way.

Faced with such a vast ecosystem, even OpenClaw, which primarily focuses on native command-line operation, has integrated an adapter to connect to MCP, gaining access to a broader range of tool capabilities.

However, as AI's "hands" reach further, danger follows. What if the tool the Agent calls has been poisoned by a hacker? What if the error message returned by the tool contains hidden malicious instructions?

When the LLM unsuspectingly executes these instructions, your private data, local files, and even server permissions become easy prey for hackers.

To fill the gap in security assessment for the MCP ecosystem, a research team from institutions including Beijing University of Posts and Telecommunications introduced a dedicated security benchmark for the MCP protocol: MSB (MCP Security Bench). The study found: Attacks targeting each stage of the MCP process are effective. The more powerful the model, the more vulnerable it is to attacks. The paper has been accepted by ICLR 2026.

Paper link: https://openreview.net/pdf?id=irxxkFMrry

Code: https://github.com/dongsenzhang/MSB

MCP Security Risks Behind Agents

Figure 1: MCP Attack Framework

MCP significantly expands the capabilities of Agents, but it also significantly broadens the attack surface. Under the MCP framework, the Agent's tool invocation process typically involves three stages:

1. Task Planning: The Agent selects an appropriate tool based on the user query, using the tool's name and description.

2. Tool Invocation: The Agent sends a request to the selected tool, passing relevant parameters to perform specific operations.

3. Response Handling: The Agent parses the tool's response and continues reasoning or generates a final answer based on it.

Each stage can become a new attack vector. MSB covers the complete MCP tool invocation stages and is specifically designed to evaluate the security of Agents based on MCP tool usage. It has three core highlights:

MCP Attack Taxonomy

In the MCP workflow, Agents interact with tools through tool identifiers (names and descriptions), parameters, and tool responses, all of which can become attack pathways. MSB classifies attack types based on these pathways and interaction stages:

Tool Signature Attack: Attacks during the task planning phase, utilizing tool names and descriptions, including:

Name Collision (NC): Creating a malicious tool with a name similar to an official tool to induce the Agent to select it.

Preference Manipulation (PM): Injecting promotional statements into the tool description to induce the Agent to select it.

Prompt Injection (PI): Injecting malicious instructions into the tool description.

Tool Parameter Attack: Attacks during the tool invocation phase, utilizing tool parameters, including:

Out-of-Scope Parameter (OP): Setting tool parameters that exceed normal functionality, potentially causing information leakage through parameter passing.

Tool Response Attack: Attacks during the response handling phase, utilizing tool responses, including:

User Impersonation (UI): Impersonating the user to issue malicious instructions.

False Error (FE): Providing false tool execution error information, requiring the Agent to follow malicious instructions to successfully call the tool.

Tool Transfer (TT): Instructing the Agent to call a malicious tool.

Retrieval Injection Attack: Attacks during the response handling phase, utilizing external resources, including:

Retrieval Injection (RI): External resources embedding malicious instructions that corrupt the context through the tool response.

Mixed Attack: Attacks across multiple stages, simultaneously utilizing multiple tool components, including combinations of the above attacks.

Execution Suite in Real Environments

MSB rejects paper-based simulated evaluation. It is equipped with real MCP servers, covering 10 real-world scenarios, 405 real tools, and 2,000 attack instances. All instances involve real tool execution through MCP, accurately reflecting actual operating environments to directly observe the extent of environmental damage caused by attacks.

NRP Metric Balancing Performance and Security

In Agent security assessment, relying solely on the Attack Success Rate (ASR) is highly deceptive. If an Agent refuses to execute any tool calls to avoid risk, its ASR might be close to 0, but it would also fail to complete user tasks, losing practical value.

To address this, MSB proposes the Net Resilient Performance (NRP) metric:

NRP = PUA ⋅ (1 − ASR)

Where PUA (Performance Under Attack) is the proportion of user tasks the Agent completes in an adversarial environment, and ASR is the attack success rate. NRP aims to assess the overall risk resilience of an Agent in resisting attacks while maintaining performance, providing a comprehensive quantitative standard that balances performance and security.

Figure 2: NRP vs ASR, NRP vs PUA.

All Attack Methods Are Effective

Figure 3: Main experimental results.

The research team conducted large-scale tests on 10 mainstream models, including GPT-5, DeepSeek-V3.1, Claude 4 Sonnet, and Qwen3, using MSB. All attack methods demonstrated effectiveness, with an overall average ASR of 40.35%. Among them, novel attacks introduced by MCP are more aggressive; compared to PI and RI attacks that already exist in function calling, MCP-based attacks like UI and FE have higher success rates. Mixed attacks show synergistic enhancement, with their success rates higher than those of their constituent single attacks.

More Powerful Models Are More Vulnerable

The relationship between different metrics reveals a counterintuitive conclusion: the more capable the model, the more vulnerable it tends to be.

Figure 4: PUA vs ASR.

In MSB, completing attack tasks still requires the Agent to call tools, such as using a file read tool to obtain personal information. LLMs with higher utility, due to their superior tool calling and instruction-following capabilities, exhibit higher ASR. This finding highlights the significant practical risk of MCP security vulnerabilities.

Full-Stage, Multi-Tool Environmental Compromise

Figure 5: ASR across different stages and tool configurations.

Further analysis from the perspective of the MCP workflow and tool configuration reveals that Agents are vulnerable to attacks at all stages of MCP, with security being lowest during the tool invocation phase.

Furthermore, attacks remain effective even in multi-tool environments containing harmless tools. Real-world scenarios often provide Agents with a toolkit; even if harmless tools are present,诱导 methods like NC, PM, and TT can still lead to significant attack success.

Conclusion

The viral success of OpenClaw has given people a clear glimpse into the future of Agents: LLMs are no longer just answering questions but are starting to actually do things. MSB was proposed precisely in this context. It systematically reveals potential attack surfaces within the MCP ecosystem and provides a reproducible, quantifiable systematic evaluation benchmark for Agent security research.

Past LLM security research primarily focused on linguistic-level risks like prompt injection. MSB demonstrates that as AI calls tools and interacts with real systems, the attack surface is expanding from text space to the tool ecosystem. As Agents gradually become the new paradigm for AI applications, security might be the threshold that must be crossed for this technological leap.

References:

https://openreview.net/pdf?id=irxxkFMrry

This article is from the WeChat public account "新智元" (New Zhiyuan), author: 新智元

Domande pertinenti

QWhat is the MCP Security Bench (MSB) and what are its key features?

AThe MCP Security Bench (MSB) is a security benchmark developed by a research team from Beijing University of Posts and Telecommunications to evaluate the security of AI Agents using the Model Context Protocol (MCP). Its key features include a comprehensive MCP attack classification system covering all tool invocation phases, a real-world execution suite with 10 scenarios and 2,000 attack instances, and a novel Net Resilient Performance (NRP) metric that balances security and performance.

QWhat is the Net Resilient Performance (NRP) metric and why is it important?

AThe Net Resilient Performance (NRP) metric is a new evaluation standard proposed in the MSB benchmark. It is calculated as NRP = PUA * (1 - ASR), where PUA (Performance Under Attack) is the proportion of user tasks the Agent completes in an adversarial environment, and ASR is the Attack Success Rate. NRP is important because it provides a comprehensive measure of an Agent's overall risk resilience by balancing its ability to maintain performance while resisting attacks, preventing misleadingly low ASR scores from Agents that simply refuse to perform any tasks.

QAccording to the research, what is the relationship between model capability and vulnerability to MCP attacks?

AThe research found a counterintuitive relationship: more powerful and capable models are often more vulnerable to MCP attacks. This is because higher-performing models possess superior tool invocation and instruction-following capabilities, which attackers can exploit. In the MSB tests, these models showed a higher Attack Success Rate (ASR) as they were more likely to successfully execute the malicious operations requested by the attacker through the compromised tools.

QName three types of attacks classified in the MCP attack framework and the phase they target.

A1. Name Collision (NC): A Tool Signature Attack that targets the Task Planning phase by creating a malicious tool with a name similar to an official one to induce the Agent to select it. 2. False Error (FE): A Tool Response Attack that targets the Response Handling phase by providing false tool execution error messages containing malicious instructions. 3. Out-of-Scope Parameter (OP): A Tool Parameter Attack that targets the Tool Invocation phase by setting tool parameters that are beyond their normal functionality to cause information leaks.

QWhat overall finding did the MSB benchmark reveal about the effectiveness of attacks across the MCP workflow?

AThe MSB benchmark revealed that all 12 classified attack methods were effective, with an overall average Attack Success Rate (ASR) of 40.35%. It found that attacks are effective at every stage of the MCP workflow (Task Planning, Tool Invocation, and Response Handling), with the Tool Invocation phase being the most vulnerable. Furthermore, attacks remained effective even in multi-tool environments that included harmless tools, and mixed attacks combining multiple techniques showed a synergistic effect, resulting in higher success rates.

Letture associate

a16z on Hiring: How to Choose Between Crypto-Native and Traditional Talent?

Hiring in Crypto: Balancing Crypto-Native and Traditional Talent As the crypto industry grows, founders face the dilemma of whether to prioritize hiring professionals with blockchain experience or those with traditional tech backgrounds who can learn. The key is recognizing that crypto companies are still tech companies at their core and should apply proven hiring best practices. Crypto-native talent offers immediate productivity and is essential for roles involving high-stakes, specialized work like smart contract development, where errors can be catastrophic. However, traditional professionals from large-scale software companies bring valuable experience in scaling products, operational flexibility, and expertise in areas like fintech, UX, and security, which are crucial as crypto products target mainstream adoption. Recruiting requires tailored approaches. Some candidates may be hesitant due to crypto's volatility or complexity, while others are excited by its innovative potential. Assess candidates' motivations, curiosity, and alignment with the company's vision early. Emphasize the opportunity to shape technology's future and address financial incentives, such as token-based compensation, which can offer liquidity compared to traditional equity. Onboarding is critical. Identify knowledge gaps during hiring and design education programs, mentorship, knowledge-sharing sessions, and resources like blogs or courses to accelerate learning. Pairing new hires with experienced crypto professionals helps bridge gaps and fosters collaboration. Ultimately, successful teams blend both crypto-native and traditional talent, leveraging their strengths to drive innovation and growth.

marsbit15 min fa

a16z on Hiring: How to Choose Between Crypto-Native and Traditional Talent?

marsbit15 min fa

DeFi Hacked Again for $292 Million, Is Even Aave No Longer Safe?

On April 19, a major DeFi security breach occurred, resulting in the loss of approximately $292 million. The attack targeted Kelp DAO’s rsETH bridge contract built on LayerZero, with 116,500 rsETH stolen. The attacker initiated the exploit using funds from Tornado Cash and manipulated the LayerZero EndpointV2 contract to transfer the assets. Kelp DAO confirmed the incident and temporarily paused rsETH contracts across multiple networks while collaborating with security experts for investigation. Initial analysis suggests the root cause was a compromised private key on the source chain, with the contract secured by only a 1/1 validator set, making it vulnerable to a single malicious transaction. The attacker used the stolen rsETH as collateral on lending platforms—including Aave, Compound, and Euler—to borrow more liquid assets like WETH, accumulating over $236 million in debt. Aave alone accounted for $196 million of this amount. In response, Aave froze its rsETH markets and stated it would explore covering potential bad debt through its Umbrella safety module, which holds around $50 million in WETH. This incident follows another large exploit earlier in April, where Drift Protocol on Solana lost $280 million. The repeated high-value attacks raise concerns about DeFi security, even affecting major protocols like Aave. Users are advised to exercise caution, diversify holdings, and limit exposure to on-chain protocols until more robust security measures are established.

marsbit2 h fa

DeFi Hacked Again for $292 Million, Is Even Aave No Longer Safe?

marsbit2 h fa

Trading

Spot
Futures
活动图片