The 75% Commit Drop: I Analyzed 3 Years of Blockchain Developer Data — Here's What AI Is Actually Doing to Web3

I’ve been running data pipelines for blockchain analytics for three years now. Last week, while debugging a pipeline that tracks developer activity across major chains, I noticed something that made me do a triple-take: weekly crypto commits have crashed from 871,000 to 218,000—a 75% drop since early 2025.

My first thought? “The pipeline’s broken.” My second thought? “Web3 is dying.” But after digging into the numbers for days (and naming this analysis pipeline “Squid Game” because the survival rate felt that dramatic), I found something way more interesting than a simple exodus story.

The Numbers Everyone’s Panicking About

According to Artemis data I’ve been tracking:

  • Weekly commits: 871K → 218K (-75%)
  • Active developers: ~10,000 → 4,600 (-56%)
  • Ethereum devs: Down 34% to 2,811
  • Solana devs: Down 40% to 942
  • Base devs: Down 52% to 378

Meanwhile, GitHub overall added 36 million developers in 2025 and saw commits rise 25% year-over-year. AI-related repositories exploded: Jupyter Notebooks up 75%, Dockerfiles (for AI apps) up 120%.

So yeah, at first glance it looks like every crypto dev rage-quit to build the next LLM wrapper.

But Wait—Here’s What the Data Actually Shows

When I segmented the data by developer tenure (something I learned to do after my mom kept asking “but are the GOOD developers leaving?”), I found the plot twist:

Developers with 2+ years experience in crypto are UP 27% year-over-year.

Let me say that again: experienced crypto devs INCREASED. They now produce roughly 70% of all commits, up from about 45% in 2023.

The drop is almost entirely in developers with less than 12 months of activity, which declined 58%. We’re not losing our senior engineers to AI companies—we’re losing the tourists and part-timers.

My Theory: We’re Not Losing Devs to AI—AI Is Making Our Best Devs 3-10x More Productive

Here’s what I’m seeing in my own work and in the pipeline data:

  1. Same core contributors, fewer public commits

    • CI/CD pipeline execution times down 30-40% across projects I track
    • Test coverage actually UP (median 78% → 84%)
    • Lines of code per commit increasing significantly
  2. Complexity per commit is rising

    • Used to see lots of “fix typo” and “update README” commits
    • Now seeing meatier, more substantial changes per commit
    • Suggests people are batching work differently—or AI is helping them do in one commit what used to take five
  3. Anecdotal evidence from my team

    • Our team’s public commit count is down 35%
    • But we shipped 2x as many features in Q1 2026 vs Q1 2025
    • I personally shipped a cross-chain data indexer in 3 weeks that would’ve taken me 2-3 months in 2023
    • My secret? Claude Code for boilerplate, Copilot for SQL optimization, and ChatGPT to explain Rust compiler errors my brain can’t parse at 2am

The Uncomfortable Question Nobody Wants to Ask

If one experienced developer with AI tools can do the work of three developers without AI… do we actually need fewer developers?

This isn’t a “AI will take our jobs” panic post. I’m asking seriously: are we watching the efficiency gains that every other industry goes through? Where productivity improvements mean you need fewer people to produce the same output?

The data shows:

  • Wallet infrastructure devs: Only category with GROWTH (+6%)
  • DeFi, NFT, L1 protocol devs: All declining
  • The survivors: Senior devs who know what to build and how to verify AI output

What This Means for Your Career (and Mine)

I’ve been thinking a lot about this because my parents still don’t really understand what I do, but they definitely understand “75% drop” in anything. Here’s my current mental model:

If you’re a junior dev using AI as a crutch: You’re in danger. AI can write code you don’t understand, and when it breaks (it will), you won’t know how to fix it.

If you’re an experienced dev using AI as an accelerator: You’re probably more valuable than ever. You can 10x your output while maintaining quality because you can verify the AI’s work.

If you’re trying to break into Web3 right now: The bar just got higher. You need to be better than “junior dev + AI” to get hired.

I Want to Hear Your Experience

I’m literally just one data engineer with too much coffee and a Jupyter notebook. My data shows patterns, but I want to hear the human stories:

  1. Are you shipping more code with AI tools, just with fewer commits?
  2. Have you seen quality improve, decline, or stay the same?
  3. If you’re hiring: are you looking for fewer, more senior developers now?
  4. If you’re junior: how are you learning when AI can do the “practice” coding for you?

Also, if anyone wants to collaborate on this analysis (or tell me my methodology is totally wrong), I’ve got all my data pipeline code here: [fictional github link]. It’s a mess and the variable names are in Korean but the SQL queries are solid.

TL;DR: Crypto commits down 75%, but experienced devs UP 27%. Either AI is making senior devs super productive, or we’re watching the juniors get filtered out by higher bars to entry. Probably both. What are you seeing in your corner of Web3?


Data sources: Artemis, CoinDesk developer activity analysis, my own on-chain analytics pipelines. If you want specific chain breakdowns or methodology details, drop a comment. I love talking about data way too much.

Mike, your data matches exactly what I’m seeing from the audit side—and honestly, it’s both encouraging and terrifying.

The Quality Paradox I’m Living Through

My audit workload has increased 40% year-over-year despite the 75% commit drop you documented. At first I thought it was just me getting more clients, but when I compared notes with other auditors at ETHDenver, everyone reported the same pattern.

Here’s what’s happening in the trenches:

Projects are shipping faster. What used to take 6-8 weeks from “hello” to “ready for audit” now takes 3-4 weeks. Teams are clearly more productive with AI tools.

But security issues per 1000 lines of code are actually rising. I don’t have your statistical rigor, but anecdotally I’d say I’m finding 20-30% more vulnerabilities per audit than I was in 2024.

The “AI-Confident” Code Problem

I’ve started categorizing bugs as either “classic mistakes” or “AI-confident mistakes,” and the latter is a growing category:

Classic mistake example:

// Developer forgot to check for zero address
function transfer(address to, uint amount) {
    balances[to] += amount;  // Oops, forgot validation
}

AI-confident mistake example:

// AI generated a complex proxy pattern with subtle reentrancy
function upgradeAndCall(address newImpl, bytes calldata data) external onlyOwner {
    // Looks secure, has access control, follows upgrade pattern
    // But the call order creates a 2-block reentrancy window
    // that even experienced auditors missed in first pass
    implementation = newImpl;
    (bool success,) = address(this).call(data);
    require(success, "Upgrade failed");
}

The second one looks professional. It has access control. It follows established patterns. But it has a critical vulnerability that the developer who submitted it couldn’t explain when I asked about it in the audit call.

Their response? “Claude suggested this pattern for upgradeable contracts.”

The Learning Crisis

This connects to your point about junior devs declining 58%. I think we’re watching two simultaneous trends:

  1. AI is filtering out people who were never going to make it anyway. The “tutorial followers” who could copy-paste but never understood the fundamentals.

  2. But AI is also preventing the next generation from building intuition. You learn security by making mistakes on small contracts, getting burned, and building scar tissue. If AI writes your practice code, you never develop that pattern recognition.

I’m seeing junior devs who can ship a complex DeFi protocol in 2 weeks but can’t explain why their contract needs nonReentrant or what a storage collision is.

The Senior Dev Advantage You Identified

Your data showing experienced devs UP 27% makes perfect sense from where I sit. The audits that go smoothly are from teams where:

  • The lead dev understands the AI-generated code deeply enough to modify it
  • They use AI for boilerplate but hand-write the security-critical sections
  • During audit calls, they can explain why the code is safe, not just that it passed tests

The nightmare audits are when nobody on the team actually understands their own codebase because AI wrote 80% of it.

My Uncomfortable Prediction

If this trend continues, I think we’ll see:

  • Audit costs increase (we’re already seeing this—teams need to budget 2x more when AI was heavily used)
  • More catastrophic failures when AI-generated code gets deployed without proper review
  • A bifurcation in developer quality between “AI-assisted experts” and “AI-dependent beginners” with very little middle ground

Question for You, Mike

Can your data distinguish between:

  • Projects where experienced devs use AI as a tool (productivity up, quality stable)
  • Projects where junior devs use AI as a crutch (shipping faster but less secure)

Because from my view, those are two completely different phenomena that both reduce commit counts but have opposite outcomes.

Also, on a lighter note: “Squid Game” is a perfect name for that pipeline. I named my latest audit framework “Test Twice, Deploy Once” but everyone just calls it TTDO which sounds like a todo app. Your Korean names are better.


Quick disclaimer: Not trying to FUD AI tools—I use Claude Code and Cursor daily. But we need to talk honestly about the skill gap problem before it becomes a “hundreds of millions exploited” problem.

Coming at this from the product side, and I think both Mike and Sarah are describing the same elephant from different angles.

Mike’s data shows we’re building with fewer commits. Sarah’s audits show we’re building faster but less safely. From where I sit managing a Web3 sustainability protocol, I’ll add the third angle: commit count was always a terrible proxy for value creation.

The Metrics That Actually Matter

When I transitioned from non-profit work to product management, one of the first things I learned was that “lines of code” and “commits per sprint” are vanity metrics. What users care about:

  1. Features that solve real problems
  2. Reliability when they need it
  3. Time-to-value

Our team’s story this quarter perfectly captures the AI productivity paradox Mike identified:

  • Commits down 40% (Mike’s trend confirmed)
  • Features shipped: 2x higher (carbon credit verification + impact dashboard, both launched)
  • Bug severity: UP (we had a critical incident in production when AI-generated edge case handling failed)

That last point is what Sarah is warning about. We shipped faster. Users got value sooner. But we also deployed a bug that took down the platform for 4 hours because the developer (me, to be honest) merged AI-suggested code without fully understanding the error handling.

The Real Question: Are We Building the Right Things Faster, or Just Building Faster?

Here’s what keeps me up at night, and it ties to Mike’s data about experienced devs being UP 27%:

Those experienced developers aren’t just “faster with AI”—they’re the ones who know what NOT to build.

The most valuable thing a senior engineer has told me this year wasn’t about code at all. It was: “Alex, that feature you want to ship in 2 weeks with AI assistance? It’ll take 6 months to maintain. Let’s build the simpler version.”

Junior Alex with AI could have shipped the complex version in 2 weeks. Senior engineer saved us from 6 months of technical debt.

The Productivity Paradox I Lived Through in Non-Profits

This reminds me so much of what happened in the non-profit sector around 2015 when everyone discovered “growth hacking” and “digital engagement metrics”:

  • Organizations optimized for email open rates instead of actual volunteer retention
  • They A/B tested donation buttons while their programs lost impact
  • Leadership celebrated “10x increase in social media reach” while mission outcomes flatlined

The metrics went up. The impact didn’t.

I’m seeing echoes of that here. We’re celebrating 3x faster development cycles while asking:

  • Are users actually better off?
  • Is the code maintainable?
  • Did we solve the right problem?

Sarah’s Point About Junior Devs Hits Different When You’re Hiring

Sarah mentioned juniors shipping complex code they can’t explain. As someone who screens candidates and makes hiring decisions, this is becoming a real problem:

Pre-AI interview (2023):

  • “Walk me through how you built this feature”
  • Candidate explains their reasoning, trade-offs, what they learned

Post-AI interview (2026):

  • “Walk me through how you built this feature”
  • “I asked ChatGPT and it worked”
  • “What did you learn?”
  • “…that ChatGPT is good at coding?”

That’s not a developer. That’s a prompt engineer who might become a developer.

The Measurement Problem Is Real

Mike asked great questions at the end about quality and learning. I’d add: we need new frameworks for measuring developer value in an AI-assisted world.

Traditional metrics breaking down:

  • Commits per week → meaningless (AI batching)
  • Lines of code → meaningless (AI verbosity)
  • Features shipped → misleading (may be solving wrong problems)
  • PR review time → misleading (complex AI-generated diffs)

What might actually matter:

  • Mean time to recovery when things break
  • User outcome improvements (not just feature count)
  • Code comprehension score (can the team explain their own system?)
  • Technical debt accumulation rate

Sustainability Lens (Because I Can’t Help Myself)

One more angle: If AI makes us 2x productive, the question isn’t “can we ship 2x faster?”—it’s “what’s the environmental and social ROI of that extra capacity?

Are we using AI to:

  • :white_check_mark: Solve climate problems faster (yes, please)
  • :white_check_mark: Make Web3 more accessible to underserved communities (love this)
  • :cross_mark: Launch the 500th low-effort NFT marketplace (please stop)
  • :cross_mark: Ship features users didn’t ask for just because we can (speed ≠ value)

The experienced devs Mike’s data shows sticking around? They’re the ones thinking about this.

Questions Back to This Thread

  1. Mike: Can you segment your data by “maintained projects” vs “abandoned projects”? I suspect AI is great for starting projects but doesn’t help with the unglamorous maintenance work.

  2. Sarah: Are you seeing any correlation between “AI usage intensity” and “audit findings severity”? Even rough qualitative patterns would be interesting.

  3. To everyone: How do you measure developer value now? What metrics do you actually trust?

Also, huge +1 to Sarah’s point about audit costs increasing. We just budgeted $45K for our Q2 audit vs $28K for the same scope in 2025. Our auditor’s explanation? “AI-assisted code requires more scrutiny.”


Disclaimer: I literally used Claude to help format this post and catch my typos. The irony is not lost on me. But the ideas and concerns are mine, built from 6 years of watching organizations optimize for the wrong metrics.

Mike’s data is fascinating, but I think we need to segment it more carefully to understand what’s actually happening. The 75% drop isn’t uniform across the stack, and that difference tells us everything about where AI helps vs. where it fails.

The Critical Question: Protocol Dev vs. Application Dev

Mike, can you break down your commit data by layer?

  • Protocol/consensus layer (L1s, L2 sequencers, consensus clients)
  • Infrastructure (bridges, oracles, indexers)
  • Application layer (dApps, frontends, smart contracts)

I’m willing to bet the drop is heavily concentrated in the application layer, with protocol development staying relatively stable or even growing slightly.

Why This Matters: AI Can’t Do Consensus

From my work contributing to Ethereum’s consensus layer and building zkEVM implementations, here’s the reality:

GitHub Copilot is amazing for:

  • React components and frontend boilerplate
  • CRUD operations and API endpoints
  • Test scaffolding and mock data
  • Standard ERC-20/721 token contracts

GitHub Copilot is useless for:

  • Zero-knowledge circuit optimization
  • Consensus algorithm correctness proofs
  • Cross-chain message verification protocols
  • Novel cryptographic primitive implementations

Nobody—and I mean nobody—is using AI to write the core logic of zkSync, Arbitrum’s fraud proofs, or Firedancer’s consensus implementation.

The Data I’m Seeing From Ethereum Foundation

Anecdotally (and I wish I had Mike’s data rigor here), looking at Ethereum core dev activity:

  • EIP discussions: Stable or slightly up
  • Consensus client commits (Prysm, Lighthouse, Teku): Relatively stable
  • Research forum activity: Stable
  • Execution layer improvements: Stable to up

The exodus Mike documented isn’t happening at the protocol level. It’s happening in the “build another DEX” layer.

Sarah’s Security Point Is Even Scarier Than She Realizes

Sarah talked about “AI-confident code” with subtle vulnerabilities. This is exponentially worse when you’re dealing with consensus-critical or bridge code.

A reentrancy bug in a DEX? Bad. You lose user funds, project reputation tanks, but the chain keeps running.

A consensus bug in a Layer 2 sequencer? You can halt the entire network and potentially brick millions in locked funds.

The scary part: AI models are trained on GitHub code, which means they’re great at reproducing common patterns and terrible at the novel, bespoke, ultra-careful code that critical infrastructure requires.

Example from my own work: I’m building a cross-chain messaging protocol. The message verification logic has to be:

  • Resistant to replay attacks
  • Handle chain reorgs correctly
  • Deal with gas limit variations across chains
  • Account for signature malleability

This isn’t in AI training data because it doesn’t exist yet. I’m literally designing it from first principles using cryptographic papers and threat modeling.

AI can help me write the TypeScript SDK around this. But the core verification logic? That’s 100% hand-written, peer-reviewed by cryptographers, and formally verified.

The Two-Tier Developer Ecosystem Alex Hinted At

Alex mentioned the non-profit sector optimizing for vanity metrics. I think we’re heading toward something similar but more structural:

Tier 1: Protocol Developers (Stable/Growing)

  • Writing consensus logic, novel cryptography, core infrastructure
  • AI is a minor productivity boost for documentation and tests
  • High barriers to entry (requires deep cryptography, distributed systems knowledge)
  • Jobs: secure, well-paid, hard to fill

Tier 2: Application Developers (Shrinking)

  • Building on existing primitives (DEXs, lending, NFTs)
  • AI can do 60-80% of the work
  • Barriers to entry collapsing (AI handles implementation)
  • Jobs: competitive, AI is substituting for junior positions

The “missing middle” is exactly what Mike’s data shows: juniors down 58%, experienced devs up 27%.

Career Advice for Anyone Reading This

If you’re a junior dev or trying to break into Web3 right now, here’s my blunt take:

Don’t compete in categories where AI is strong. Building the 83rd Uniswap fork with ChatGPT help? You’re in a race to the bottom.

Instead, go deep on things AI can’t do:

  1. Learn zero-knowledge proofs (zkSNARKs, STARKs, Plonky2)

    • AI can’t design ZK circuits
    • Demand is skyrocketing (every L2, privacy protocol, identity solution)
    • Circom and Noir developers are getting $200K+ offers
  2. Understand consensus mechanisms deeply

    • Read the Ethereum Yellow Paper, Gasper paper, HotStuff paper
    • Contribute to client implementations
    • Nobody is using AI to design consensus
  3. Master formal verification

    • Learn tools like Certora, K framework, TLA+
    • The better AI gets at generating code, the more we need formal verification
    • Sarah’s audit costs going up? Verification engineers going up with them
  4. Build novel cryptographic protocols

    • Threshold signatures, MPC, homomorphic encryption
    • This requires math and creativity AI doesn’t have

Response to the “Do We Need Fewer Developers?” Question

Mike asked: if 1 dev with AI = 3 devs without AI, do we need fewer developers?

My answer: We need fewer application developers and more protocol/infrastructure developers.

The application layer is getting commoditized. DeFi primitives are well-understood. AI can replicate them.

But the infrastructure layer is exploding:

  • L2 scaling solutions
  • Cross-chain messaging
  • Account abstraction
  • MEV mitigation
  • ZK privacy layers

Each of these needs deep, specialized knowledge that AI can’t provide.

Questions Back

To Mike: Can you segment by:

  1. Smart contract deploys vs. protocol commits?
  2. Greenfield projects vs. maintained infrastructure?
  3. Commits to novel protocols vs. forks/variations of existing ones?

I suspect you’ll find protocol innovation is healthy while application-layer cloning is what collapsed.

To Sarah: Are your audit findings concentrated in application-layer contracts or do you also audit infrastructure-level code (bridges, sequencers)? If you do both, do you see different vulnerability patterns?

To Alex: From a product perspective, how do you balance “ship fast with AI” vs “maintain quality”? Are you seeing product teams rethink their velocity targets?


One last philosophical point: The 75% commit drop might not be a crisis—it might be a correction.

We had too many people building essentially the same thing (DEX clones, NFT platforms, lending forks). If AI filters out the derivative builders and forces people to either go deep on infrastructure or provide actual product innovation (like Alex’s sustainability protocol), that’s probably healthy for the ecosystem.

The question is whether the juniors getting filtered out will re-skill into protocol work or leave crypto entirely. That’s the real talent retention question.

From a formal verification and security research perspective, this discussion is revealing a systemic risk that I believe the industry is underestimating.

Mike’s data, Sarah’s audit observations, Alex’s product insights, and Brian’s protocol-layer analysis all point to the same underlying problem: we’re optimizing for velocity at the expense of verifiability.

The Verification Crisis

Let me frame this with recent research data from OWASP’s Smart Contract Top 10 2026 report (published last month):

Key findings:

  • 60%+ of contracts audited in late 2025 contained AI-generated code segments
  • Time-to-vulnerability-discovery decreased (good: auditors finding bugs faster)
  • But: Severity of undiscovered vulnerabilities increased (bad: the bugs that slip through are worse)

The pattern Sarah identified—“AI-confident code” with subtle vulnerabilities—is showing up across the industry. And Brian’s point about consensus-critical code is especially concerning.

Response to Sarah: Yes, I’m Seeing the Same Pattern

Sarah asked if I’m seeing correlation between AI usage intensity and audit findings severity. Qualitatively, absolutely yes.

I’ve started categorizing audits by approximate AI usage (based on code patterns, developer interviews, and git history):

Low AI usage (<20% of codebase):

  • Traditional vulnerability patterns (access control, reentrancy, integer overflow)
  • Developers can explain security model during audit calls
  • Average critical findings: 1.2 per audit

High AI usage (>60% of codebase):

  • Novel vulnerability combinations (AI chains multiple patterns in unexpected ways)
  • Developers struggle to explain why code is secure
  • Average critical findings: 2.8 per audit
  • Concerning trend: More “unknown unknowns”—vulnerabilities I haven’t seen before

Example: The Proxy Pattern Brian Mentioned

Brian talked about consensus-critical code needing bespoke security. I recently audited a bridge protocol where the developer used an AI-generated proxy upgrade pattern.

The code looked textbook-correct:

  • :white_check_mark: Access control on upgrade function
  • :white_check_mark: Initialization protection
  • :white_check_mark: Storage layout preservation
  • :white_check_mark: Event emission on upgrade

But there was a 4-block window during upgrade where the bridge was vulnerable to a state manipulation attack that could have resulted in infinite minting.

The developer couldn’t see it because they didn’t understand why the proxy pattern works. They just knew “AI said this is the secure proxy pattern.”

The Formal Verification Problem Mike Identified

Mike’s data showed:

  • Commits down 75%
  • But complexity per commit up

From a verification standpoint, this is catastrophic. Here’s why:

Traditional code review assumes:

  • Small, incremental changes
  • Reviewers can reason about diff impact
  • Test coverage catches regressions

AI-generated code breaks these assumptions:

  • Large diffs (AI writes 200 lines at once)
  • Complex interactions hard to reason about
  • Tests also AI-generated (might miss edge cases)

The Math That Should Scare You

If a human writes 50 lines of code with 1% bug probability = 0.5 bugs expected

If AI writes 500 lines of code with 0.5% bug probability = 2.5 bugs expected

But: AI-generated bugs are less obvious, so detection rate is lower. If humans catch 90% of their own bugs but only 60% of AI bugs:

  • Human residual risk: 0.05 bugs
  • AI residual risk: 1.0 bugs (20x higher)

This is consistent with Sarah’s observation of more vulnerabilities per audit despite overall developer experience increasing.

Brian’s Two-Tier Ecosystem Is a Security Bifurcation

Brian’s prediction about protocol vs. application developer split has security implications:

Protocol layer:

  • Hand-written, peer-reviewed, formally verified
  • High security bar maintained
  • Experienced developers who understand threat models

Application layer:

  • AI-assisted, fast iteration
  • Security as afterthought (“we’ll audit later”)
  • Developers who can’t explain their own security model

The problem? Applications compose with protocols. A vulnerable dApp can drain a secure protocol through composability.

We’re creating a system where the foundation is solid but the applications built on top are structurally unsound.

Response to Alex: New Metrics for AI Era

Alex asked what metrics we should trust now. From security perspective:

Deprecated metrics:

  • Lines of code (meaningless with AI)
  • Test coverage % (AI generates tests that pass but don’t cover edge cases)
  • Code review approval count (if reviewers don’t understand code)

Critical new metrics:

  • Formal verification coverage (% of critical paths formally proven safe)
  • Explainability score (can developers explain security model?)
  • Invariant test coverage (not just unit tests, but property-based tests)
  • Third-party security review depth (not just “did you audit” but “how many rounds”)
  • Incident response time (Brian’s point about MTTR)

I’m working on a research framework called “Security Comprehension Index” that measures:

  1. Can the development team explain their security model?
  2. Can they articulate threat scenarios?
  3. Can they walk through why specific code patterns prevent exploits?

Projects that score high are using AI productively. Projects that score low are accumulating security debt.

The Uncomfortable Prediction I Share With Sarah

Sarah predicted:

  • Audit costs increase (confirmed: my rates up 40% YoY)
  • More catastrophic failures (I fear this is coming)
  • Bifurcation in developer quality (Brian’s two-tier system)

I’ll add a fourth prediction: We’ll see the first major exploit explicitly attributed to AI-generated code within 6 months.

Not “developer used AI and made a mistake”—but “AI generated a novel vulnerability pattern that no human would have written.”

This will be a watershed moment for the industry.

Recommendations for This Community

For developers:

  1. Treat AI output as untrusted external input
  2. Every AI-generated line should be reviewed with adversarial mindset
  3. If you can’t explain why code is secure, don’t deploy it
  4. Budget 2-3x more for security audits when using AI tools heavily

For protocols:

  1. Require security comprehension tests during code review
  2. Implement formal verification for critical paths
  3. Run bug bounties before mainnet, not after
  4. Consider AI-usage disclosure in security documentation

For the ecosystem:

  1. We need better static analysis tools for AI-generated code
  2. Research into “AI-typical” vulnerability patterns
  3. Education programs that teach verification alongside generation
  4. Open benchmark datasets for AI security (similar to Paradigm’s EVMBench)

Research Question for Anyone Interested

I’m compiling data on AI-generated vulnerability patterns. If you’ve:

  • Found bugs in AI-generated smart contracts
  • Conducted audits where AI usage was significant
  • Researched AI coding assistant outputs

I’d love to collaborate. Working on a paper tentatively titled “Adversarial Analysis of LLM-Generated Smart Contracts” for submission to IEEE S&P.


Final thought: This discussion demonstrates the maturity of our ecosystem. We’re not panicking about the 75% drop—we’re analyzing what it means and adapting.

Mike brought data. Sarah brought audit frontline experience. Alex brought product wisdom. Brian brought protocol perspective. This is exactly the kind of multi-disciplinary thinking we need.

The question isn’t “is AI good or bad for Web3 development”—it’s “how do we ensure AI makes us productively faster, not dangerously faster?”

Trust but verify. Then verify again.