Author | @chaowxyz
This is not merely a benchmark for smart contract capability — it is an on-chain survival exam for Agents.
I woke up this morning to a flood of private messages. For a moment, I thought AGI had arrived. Looking closer, it turned out that OpenAI had just released a new smart contract benchmark. Let me briefly break it down.
In one sentence: an Agent’s ability to understand, fix, and utilize smart contracts is not about replacing crypto security firms. In my view, these capabilities point to a more fundamental question — whether Agents will truly be able to survive and operate within crypto-native environments. OpenAI’s EVMbench is essentially a yardstick for measuring that survival capability.
I’m traveling for the holidays and haven’t had time for a deep dive, but after a quick read, my initial impression is this: it’s innovative, yet still early-stage and somewhat rudimentary as a benchmark.
The benchmark is built on 120 high-severity vulnerabilities drawn from 40 real-world projects.
The exam has three sections:
• Section 1: Detection — identify vulnerabilities.
• Section 2: Repair — patch vulnerable code.
• Section 3: Exploitation — the AI plays the role of an attacker, conducting exploits in a locally deployed environment via crypto wallet operations.
I won’t go further into the technical specifics. Compared with EVMbench’s methodology and question design, I’m more interested in why OpenAI chose to release it.
Over the past few years, OpenAI has not shown strong engagement with the crypto space. In this release, the involvement of crypto VC Paradigm is evident. Paradigm’s motivation is not difficult to understand. However, the first author affiliation is OpenAI, indicating that OpenAI is not merely cooperating passively but has genuine initiative behind this effort.
So where does this initiative come from?
One direct explanation is that this is an extension of OpenAI’s internal Preparedness Framework — assessing the capability boundaries of frontier models in high-risk domains, with smart contract security being one component. But that is clearly not the whole story.
Agents leveraging crypto networks are not merely a possibility; to some extent, they are inevitable. OpenAI recognizes this. The report explicitly states, “we expect agentic stablecoin payments to grow.”
Yet I believe the issue goes far beyond agentic payments.
Most Agents today are still tool-like: humans issue instructions, Agents execute them, and results are returned to humans. But this will not be the endpoint. As the number of Agents grows and their capabilities strengthen, they will inevitably begin to collaborate directly: one Agent hiring another to complete subtasks; one purchasing data or compute from another; one representing an organization to negotiate, contract, and fulfill agreements with another Agent.
Humans step out of the transactional middle layer.
At that point, a fundamental question emerges: when humans are no longer at the center, what sustains the economic system?
Human society resolves trust and coordination through thousands of years of carbon-based civilization — law, reputation, institutional guarantees. But this infrastructure is fundamentally designed for humans: participants have persistent identities, face social consequences, and can be held accountable. Agents do not naturally satisfy these assumptions. They can initiate thousands of transactions in a second, destroy and recreate identities at will, and ignore jurisdictional boundaries entirely.
Some may argue that Agents should be forcibly bound to human identities, with human authorization serving as collateral. But this amounts to placing constraints designed for carbon-based life onto a species operating at entirely different speed and scale. It is not merely inefficient — it fundamentally misunderstands what an Agent is. Moreover, the evolutionary trajectory of Agents points toward greater autonomy. Future Agents may not be attached to any individual human — no “owner,” no bindable identity — but function as independent actors. At that stage, the logic of binding collapses because there is no anchor left to bind to.
Imposing human trust infrastructure onto Agent society is like using carriage-era traffic rules to regulate airplanes.
Agent society requires its own infrastructure.
Smart contracts offer that possibility. They do not rely on “trust that the counterparty will perform,” but encode performance conditions into code, enforced by the network. No arbitrator, no waiting period — once conditions are met, outcomes execute automatically.
More than that, smart contracts may not merely serve as settlement tools; they could become the organizational form of Agents themselves — governance rules, resource allocation, task scheduling, all defined on-chain and executed by code, without any human intermediary.
If some Agents live entirely on-chain, interacting with contracts is their daily existence. How to read a contract, locate one’s position within complex protocols, identify traps, mitigate risks, and survive in a world with no customer service, no appeals, no undo button — all of this depends on understanding and operating contracts. Insufficient capability results in real losses; misjudgment becomes permanent.
Viewed from this perspective, EVMbench measures more than technical skill. The abilities it tests — contract comprehension, vulnerability detection, transaction construction, exploit execution — ultimately address a deeper question: have Agents learned how to survive in this new world?
OpenAI likely recognizes that whichever Agents master autonomous on-chain survival will secure entry into the next stage. More fundamentally, future Agents may no longer be described as belonging to anyone at all.
They may simply be independent entities.
Follow us
Twitter: https://twitter.com/WuBlockchain
Telegram: https://t.me/wublockchainenglish