AI agents find $4.6M in blockchain smart contract exploits

LLM Agents for Security & Startup Viability

  • Several commenters are already using LLMs for pentesting, reverse engineering, and static analysis, and report a big jump in capability with recent model generations.
  • One startup founder says newer models are saturating their benchmarks and are now cheap enough to use in production.
  • Others are hesitant to build companies on top of proprietary APIs, fearing deplatforming or being “Sherlocked.” Some argue this is fine if you can move fast, make money, and exit; others dislike “exit-first” startup culture and prefer long-term, values-driven businesses.

Bypassing Safety Guardrails in Practice

  • People describe getting around model safety systems by:
    • Decomposing exploitation tasks into harmless subtasks (e.g., “find potential issues in this snippet”).
    • “Social engineering” the AI with elaborate justifications.
    • Using multiple providers; experiences vary: ChatGPT seen as overly cautious, Claude as technically strong but rate-limited, Gemini somewhere in between.
  • Some note that providers rarely crack down on legitimate pentesting with commercial accounts, though usage may technically graze ToS.

Models vs Agent Scaffolding

  • Debate over whether improvements come mainly from better models or better “business logic”/tooling.
  • Several argue it’s overwhelmingly the models: modern agents can do a lot with very thin scaffolding (e.g., simple “terminal in a loop”). Tool-calling logic is described as simple; the hard part is training models to use tools well.
  • Others point out that ecosystem advances (structured outputs, memory, retrieval, MCP, etc.) also matter, but agree raw models have improved a lot.

Significance of Anthropic’s Results

  • Some say $4.6M and mostly old bugs highlight poor Ethereum infosec more than LLM brilliance; others stress the key point is fully autonomous exploitation and measured “dollars stolen,” not just bug detection.
  • The article’s note that two real zero-days worth ~$3.7k were found, at comparable API cost, prompts skepticism about economic viability and accusations of PR spin.

Real-World Exploitation & Incentives

  • Commenters assume many parties already brute-force smart contracts with AI and other tooling, given huge bug bounties and prior non-AI automation.
  • Legal risk is seen as a major deterrent in Western jurisdictions; state-aligned or sanctioned actors face fewer constraints.

Ethereum Immutability & Governance Debate

  • The DAO fork resurfaces as evidence that “immutable” chains can be politically altered when major stakeholders lose money.
  • Some argue this shows de facto centralization and plutocracy; others counter that users voluntarily chose the fork, the unforked chain still exists, and no irregular changes have occurred since.

Smart Contracts & the Oracle Problem

  • Multiple explanations clarify that:
    • On-chain contracts are immutable programs managing state and assets, with atomic transactions and permission checks.
    • Many powerful use cases (escrow, AMMs, DAOs, voting, token swaps) work entirely on-chain and don’t need external data.
    • When real-world events are involved, contracts rely on oracles—trusted third parties or consensus-based mechanisms—which reintroduce trust and potential failure modes (“oracle problem”).
  • Some view smart contracts as technically elegant but badly tainted by speculation and scams; others see token economics as a necessary evil to fund infrastructure.

Broader Reflections on AI Autonomy

  • A few participants see these results as unsurprising steps toward increasingly autonomous, self-improving agents, and express excitement.
  • Others dismiss Ethereum exploitation as old news and are more interested in how generalized, agentic AI will reshape both offense and defense going forward.