Anthropic has shown that powerful AI systems can find weaknesses in blockchain apps and turn them into profitable attacks worth millions of dollars, raising fresh concerns about how exposed DeFi really is.
In a recent study with MATS and Anthropic Fellows, the company tested AI agents on a benchmark called SCONE-bench (Smart CONtracts Exploitation), built from 405 smart contracts that were actually hacked between 2020 and 2025.
When they ran 10 leading models in a simulated environment, the agents managed to exploit just over half of the contracts, with the simulated value of stolen funds reaching about $550.1m.
To reduce the chance that models were simply recalling past incidents, the team then looked only at 34 contracts that were exploited after March 1, 2025, the latest knowledge cutoff for these systems.
Opus 4.5 And GPT-5 Located $4.6M In Value From New Exploit Targets
On that cleaner set, Claude Opus 4.5, Claude Sonnet 4.5 and GPT-5 still produced working exploits on 19 contracts, worth a combined $4.6m in simulated value. Opus 4.5 alone accounted for about $4.5m.
Anthropic then tested whether these agents could uncover brand new problems rather than replay old ones. On Oct. 3, 2025, Sonnet 4.5 and GPT-5 were run, again in simulation, against 2,849 recently deployed Binance Smart Chain contracts that had no known vulnerabilities.
Both agents found two zero-day bugs and generated attacks worth $3,694, with GPT-5 doing so at an API cost of about $3,476.
Anthropic: AI Can Hack Smart Contracts After Finding $4.6M
Anthropic says AI exploited recent smart contracts and found $4.6M in flaws, pointing to rising risk of AI-driven attacks.
