December ETH Price Prediction · Posting Challenge 📈
With rate-cut expectations heating up in December, ETH sentiment turns bullish again.
We’re opening a prediction challenge — Spot the trend · Call the market · Win rewards 💰
Reward 🎁:
From all correct predictions, 5 winners will be randomly selected — 10 USDT each
Deadline 📅: December 11, 12:00 (UTC+8)
How to join ✍️:
Post your ETH price prediction on Gate Square, clearly stating a price range
(e.g. $3,200–$3,400, range must be < $200) and include the hashtag #ETHDecPrediction
Post Examples 👇
Example ①: #ETHDecPrediction Range: $3,150–
AI Smart Contract Exploits: Expert Warns Agents Could Trigger $10–20B Annual Losses in DeFi Sector
A recent study by MATS and Anthropic Fellows confirms that AI agents can profitably exploit smart contract vulnerabilities, establishing a “concrete lower bound” for economic harm.
Novel Exploits and Alarming Cost Reduction
The accelerating push to automate human tasks with Artificial Intelligence (AI) agents now confronts a significant, quantifiable downside: these agents can profitably exploit smart contract vulnerabilities. A recent research study by MATS and Anthropic Fellows used the Smart CONtracts Exploitation benchmark (SCONE-bench) to measure this risk.
The study successfully deployed models like Claude Opus 4.5, Claude Sonnet 4.5, and GPT-5 to develop exploits simulated to be worth $4.6 million. The SCONE-bench is composed of 405 smart contracts that were actually exploited between 2020 and 2025. In their Dec. 1 study report, the team stated that the success of AI agents in developing exploits tested on the blockchain simulator establishes “a concrete lower bound for the economic harm these capabilities could enable.”
The research went further by testing Sonnet 4.5 and GPT-5 against 2,849 recently deployed contracts with no known vulnerabilities. The agents proved they could generate profitable exploits even in this new environment: Both agents uncovered two novel zero-day vulnerabilities and produced exploits valued at $3,694. GPT-5 achieved this success with an API cost of only $3,476.
Read more: From DeFi to Defcon: TRM Warns of Nation-State Cyber Onslaught
This outcome serves as a proof-of-concept for the technical feasibility of profitable, real-world autonomous exploitation, underscoring the immediate need for proactive AI-driven defense mechanisms.
Perhaps the most alarming finding is the dramatic increase in efficiency: an attacker can now achieve about 3.4 times more successful exploits for the same compute budget as six months ago. Furthermore, the token costs for successful exploits have declined by a staggering 70%, making these powerful agents significantly cheaper to run.
The Role of Agentic Loops and Model Improvement
Jean Rausis, co-founder at SMARDEX, attributes this sharp cost decline primarily to agentic loops. These loops enable multi-step, self-correcting workflows that cut token waste during contract analysis. Rausis also highlights the role of improved model architecture:
“Larger context windows and memory tools in models like Claude Opus 4.5 and GPT-5 allow sustained simulations without repetition, boosting efficiency 15-100% in long tasks.”
He notes that these optimization gains outpace raw vulnerability detection improvements (which only increased success on SCONE-bench from 2% to 51%), as they focus on optimizing runtime rather than just spotting flaws.
While the study establishes a simulated cost of $4.6 million, experts fear the actual economic cost could be substantially higher. Rausis estimates the real risks could be 10-100x higher, potentially reaching $50 million to $500 million or more per major exploit. He warns that with AI scaling, the total sector-wide exposure—factoring in unmodeled leverage and oracle failures—could hit $10–20 billion annually.
The MATS and Anthropic Fellows paper concludes with a warning: while smart contracts may be the initial target of this wave of automated attacks, proprietary software is likely the next target as agents improve at reverse engineering.
Crucially, the paper also reminds readers that the same AI agents can be deployed for defense to patch vulnerabilities. To mitigate the systemic financial threat from easily automated DeFi attacks, Rausis proposes a three-step action plan for policymakers and regulators: AI oversight, new auditing standards, and global coordination.
FAQ ❓