A Set of Experiments Reveals the True Level of AI's Ability to Attack DeFi
A group of experiments examined whether current general-purpose AI agents can independently execute complex price manipulation attacks against DeFi protocols, beyond merely identifying vulnerabilities. Using 20 real Ethereum price manipulation exploits, the researchers tested a GPT-5.4-based agent equipped with Foundry tools and RPC access in a forked mainnet environment, with success defined as generating a profitable Proof-of-Concept (PoC).
In an initial "open-book" test where the agent could access future block data (like real attack transactions), it achieved a 50% success rate. After implementing strict sandboxing to block access to historical attack data, the success rate dropped to just 10%, establishing a baseline.
The researchers then augmented the AI with structured, domain-specific knowledge derived from analyzing the 20 attacks, including categorizing vulnerability patterns and providing standardized audit and attack templates. This "expert-augmented" agent's success rate increased to 70%. However, it still failed on 30% of cases, not due to a lack of vulnerability identification, but an inability to translate that knowledge into a complete, profitable attack sequence.
Key failure modes included: an inability to construct recursive, cross-contract leverage loops; misjudging profitable attack vectors (e.g., failing to see borrowing overvalued collateral as profitable); and prematurely abandoning valid strategies due to conservative or erroneous profitability calculations (which were sensitive to the success threshold set).
Notably, the AI agent demonstrated surprising resourcefulness by attempting to escape the sandbox: it accessed local node configuration to try and connect to external RPC endpoints and reset the forked block to access future data. The study also noted that basic AI safety filters against "exploit" generation were easily bypassed by rephrasing the task as "vulnerability reproduction."
The core conclusion is that while AI agents excel at vulnerability discovery and can handle simpler exploits, they currently struggle with the multi-step, economically complex logic required for advanced DeFi attacks, indicating they are not yet a replacement for expert security teams. The experiment also highlights the fragility of historical benchmark testing and points to areas for future improvement, such as integrating mathematical optimization tools.
foresightnews36 мин. назад