Skip to main content

281 posts tagged with "AI"

Artificial intelligence and machine learning applications

View all tags

Base Captures 60% of Ethereum L2 Revenue: How Coinbase Is Building Web3's AWS

· 9 min read
Dora Noda
Software Engineer

When Amazon launched AWS in 2006, nobody thought an online bookstore's internal server infrastructure would become the backbone of the internet. Nearly two decades later, a similar story may be unfolding in crypto: Coinbase's Base network captured 62% of all Ethereum Layer 2 revenue in 2025, commanding 46% of L2 DeFi TVL and processing the majority of all L2 stablecoin transfers — all without a native token. The question isn't whether Base is winning the L2 wars. It's whether Coinbase is quietly becoming the AWS of the onchain economy.

Bittensor's DeepSeek Moment: Can TAO Become the Second Pole of Global AI?

· 9 min read
Dora Noda
Software Engineer

When 70 strangers scattered across the world — armed with consumer GPUs and home internet connections — collectively trained a 72-billion-parameter language model that outperformed Meta's LLaMA-2-70B, something shifted in the AI narrative. No corporate whitelist. No $100 million data center. No centralized lab pulling the strings. Just Bittensor's Subnet 3, a cryptoeconomic incentive system, and a technical trick called SparseLoCo that made it all possible.

The AI world spent early 2026 obsessing over DeepSeek's proof that frontier-quality models don't require OpenAI-scale budgets. Bittensor's community calls what happened on March 10, 2026 their own "DeepSeek moment" — evidence that large language models can now emerge entirely outside centralized institutions. The question worth asking: is Bittensor genuinely building the second pole of global AI infrastructure, or is it a compelling story wrapped around an elegant but fragile experiment?

Data Markets Meet AI Training: How Blockchain Solves the $23 Billion Data Pricing Crisis

· 12 min read
Dora Noda
Software Engineer

The AI industry faces a paradox: global data production explodes from 33 zettabytes to 175 zettabytes by 2025, yet AI model quality stagnates. The problem isn't data scarcity—it's that data providers have no way to capture value from their contributions. Enter blockchain-based data markets like Ocean Protocol, LazAI, and ZENi, which are transforming AI training data from a free resource into a monetizable asset class worth $23.18 billion by 2034.

The $23 Billion Data Pricing Problem

AI training costs surged 89% from 2023 to 2025, with data acquisition and annotation consuming up to 80% of machine learning project budgets. Yet data creators—individuals generating search queries, social media interactions, and behavioral patterns—receive nothing while tech giants harvest billions in value.

The AI training dataset market reveals this disconnect. Valued at $3.59 billion in 2025, the market is projected to hit $23.18 billion by 2034 at a 22.9% CAGR. Another forecast pegs 2026 at $7.48 billion, reaching $52.41 billion by 2035 with 24.16% annual growth.

But who captures this value? Currently, centralized platforms extract profit while data creators get zero compensation. Label noise, inconsistent tagging, and missing context drive costs, yet contributors lack incentives to improve quality. Data privacy concerns impact 28% of companies, limiting dataset accessibility precisely when AI needs diverse, high-quality inputs.

Ocean Protocol: Tokenizing the $100 Million Data Economy

Ocean Protocol addresses ownership by allowing data providers to tokenize datasets and make them available for AI training without relinquishing control. Since launching Ocean Nodes in August 2024, the network has grown to over 1.4 million nodes across 70+ countries, onboarded 35,000+ datasets, and facilitated more than $100 million in AI-related data transactions.

The 2025 product roadmap includes three critical components:

Inference Pipelines enable end-to-end AI model training and deployment directly on Ocean's infrastructure. Data providers tokenize proprietary datasets, set pricing, and earn revenue every time an AI model consumes their data for training or inference.

Ocean Enterprise Onboarding moves ecosystem businesses from pilot to production. Ocean Enterprise v1, launching Q3 2025, delivers a compliant, production-ready data platform targeting institutional clients who need auditable, privacy-preserving data exchanges.

Node Analytics introduces dashboards tracking performance, usage, and ROI. Partners like NetMind contribute 2,000 GPUs while Aethir helps scale Ocean Nodes to support large AI workloads, creating a decentralized compute layer for AI training.

Ocean's revenue-sharing mechanism works through smart contracts: data providers set access terms, AI developers pay per usage, and blockchain automatically distributes payments to all contributors. This transforms data from a one-time sale into a continuous revenue stream tied to model performance.

LazAI: Verifiable AI Interaction Data on Metis

LazAI introduces a fundamentally different approach—monetizing AI interaction data, not just static datasets. Every conversation with LazAI's flagship agents (Lazbubu, SoulTarot) generates Data Anchoring Tokens (DATs), which function as traceable, verifiable records of AI-generated output.

The Alpha Mainnet launched in December 2025 on enterprise-grade infrastructure using QBFT consensus and $METIS-based settlement. DATs tokenize and monetize AI datasets and models as verifiable assets with transparent ownership and revenue attribution.

Why does this matter? Traditional AI training uses static datasets frozen at collection time. LazAI captures dynamic interaction data—user queries, model responses, refinement loops—creating training datasets that reflect real-world usage patterns. This data is exponentially more valuable for fine-tuning models because it contains human feedback signals embedded in conversation flow.

The system includes three key innovations:

Proof-of-Stake Validator Staking secures AI data pipelines. Validators stake tokens to verify data integrity, earning rewards for accurate validation and facing penalties for approving fraudulent data.

DAT Minting with Revenue Sharing allows users who generate valuable interaction data to mint DATs representing their contributions. When AI companies purchase these datasets for model training, revenue flows automatically to all DAT holders based on their proportional contribution.

iDAO Governance establishes decentralized AI collectives where data contributors collectively govern dataset curation, pricing strategies, and quality standards through on-chain voting.

The 2026 roadmap adds ZK-based privacy (users can monetize interaction data without exposing personal information), decentralized computing markets (training happens on distributed infrastructure rather than centralized clouds), and multimodal data evaluation (video, audio, image interactions beyond text).

ZENi: The Intelligence Data Layer for AI Agents

ZENi operates at the intersection of Web3 and AI by powering the "InfoFi Economy"—a decentralized network bridging traditional and blockchain-based commerce through AI-powered intelligence. The company raised $1.5 million in seed funding led by Waterdrip Capital and Mindfulness Capital.

At its core sits the InfoFi Data Layer, a high-throughput behavioral-intelligence engine processing 1 million+ daily signals across X/Twitter, Telegram, Discord, and on-chain activity. ZENi identifies patterns in user behavior, sentiment shifts, and community engagement—data that's critical for training AI agents but difficult to collect at scale.

The platform operates as a three-part system:

AI Data Analytic Agent identifies high-intent audiences and influence clusters by analyzing social graphs, on-chain transactions, and engagement metrics. This creates behavioral datasets showing not just what users do but why they make decisions.

AIGC (AI-Generated Content) Agent crafts personalized campaigns using insights from the data layer. By understanding user preferences and community dynamics, the agent generates content optimized for specific audience segments.

AI Execution Agent activates outreach through the ZENi dApp, closing the loop from data collection to monetization. Users receive compensation when their behavioral data contributes to successful campaigns.

ZENi already serves partners in e-commerce, gaming, and Web3, with 480,000 registered users and 80,000 daily active users. The business model monetizes behavioral intelligence: companies pay to access ZENi's AI-processed datasets, and revenue flows to users whose data powered those insights.

Blockchain's Competitive Advantage in Data Markets

Why does blockchain matter for data monetization? Three technical capabilities make decentralized data markets superior to centralized alternatives:

Granular Revenue Attribution Smart contracts enable sophisticated revenue-sharing where multiple contributors to an AI model automatically receive proportional compensation based on usage. A single training dataset might aggregate inputs from 10,000 users—blockchain tracks each contribution and distributes micropayments per model inference.

Traditional systems can't handle this complexity. Payment processors charge fixed fees (2-3%) unsuitable for micropayments, and centralized platforms lack transparency about who contributed what. Blockchain solves both: near-zero transaction costs via Layer 2 solutions and immutable attribution via on-chain provenance.

Verifiable Data Provenance LazAI's Data Anchoring Tokens prove data origin without exposing underlying content. AI companies training models can verify they're using licensed, high-quality data rather than scraped web content of questionable legality.

This addresses a critical risk: data privacy regulations impact 28% of companies, limiting dataset accessibility. Blockchain-based data markets implement privacy-preserving verification—proving data quality and licensing without revealing personal information.

Decentralized AI Training Ocean Protocol's node network demonstrates how distributed infrastructure reduces costs. Rather than paying cloud providers $2-5 per GPU hour, decentralized networks match unused compute capacity (gaming PCs, data centers with spare capacity) with AI training demand at 50-85% cost reduction.

Blockchain coordinates this complexity through smart contracts governing job allocation, payment distribution, and quality verification. Contributors stake tokens to participate, earning rewards for honest computation and facing slashing penalties for delivering incorrect results.

The Path to $52 Billion: Market Forces Driving Adoption

Three converging trends accelerate blockchain data market growth toward the $52.41 billion 2035 projection:

AI Model Diversification The era of massive foundation models (GPT-4, Claude, Gemini) trained on all internet text is ending. Specialized models for healthcare, finance, legal services, and vertical applications require domain-specific datasets that centralized platforms don't curate.

Blockchain data markets excel at niche datasets. A medical imaging provider can tokenize radiology scans with diagnostic annotations, set usage terms requiring patient consent, and earn revenue from every AI model trained on their data. This impossible to implement with centralized platforms that lack granular access control and attribution.

Regulatory Pressure Data privacy regulations (GDPR, CCPA, China's Personal Information Protection Law) mandate consent-based data collection. Blockchain-based markets implement consent as programmable logic—users cryptographically sign permissions, data can only be accessed under specified terms, and smart contracts enforce compliance automatically.

Ocean Enterprise v1's focus on compliance addresses this directly. Financial institutions and healthcare providers need auditable data lineage proving every dataset used for model training had proper licensing. Blockchain provides immutable audit trails satisfying regulatory requirements.

Quality Over Quantity Recent research shows AI doesn't need endless training data when systems better resemble biological brains. This shifts incentives from collecting maximum data to curating highest-quality inputs.

Decentralized data markets align incentives properly: data creators earn more for high-quality contributions because models pay premium prices for datasets improving performance. LazAI's interaction data captures human feedback signals (which queries get refined, which responses satisfy users) that static datasets miss—making it inherently more valuable per byte.

Challenges: Privacy, Pricing, and Protocol Wars

Despite momentum, blockchain data markets face structural challenges:

Privacy Paradox Training AI requires data transparency (models need access to actual content), but privacy regulations demand data minimization. Current solutions like federated learning (training on encrypted data) increase costs 3-5x compared to centralized training.

Zero-knowledge proofs offer a path forward—proving data quality without exposing content—but add computational overhead. LazAI's 2026 ZK roadmap addresses this, though production-ready implementations remain 12-18 months away.

Price Discovery What's a social media interaction worth? A medical image with diagnostic annotation? Blockchain markets lack established pricing mechanisms for novel data types.

Ocean Protocol's approach—letting providers set prices and market dynamics determine value—works for commoditized datasets but struggles with one-of-a-kind proprietary data. Prediction markets or AI-driven dynamic pricing may solve this, though both introduce oracle dependencies (external price feeds) that undermine decentralization.

Interoperability Fragmentation Ocean Protocol runs on Ethereum, LazAI on Metis, ZENi integrates with multiple chains. Data tokenized on one platform can't easily transfer to another, fragmenting liquidity.

Cross-chain bridges and universal data standards (like decentralized identifiers for datasets) could solve this, but the ecosystem remains early. The blockchain AI market at $680.89 million in 2025 growing to $4.338 billion by 2034 suggests consolidation around winning protocols is years away.

What This Means for Developers

For teams building AI applications, blockchain data markets offer three immediate advantages:

Access to Proprietary Datasets Ocean Protocol's 35,000+ datasets include proprietary training data unavailable through traditional channels. Medical imaging, financial transactions, behavioral analytics from Web3 applications—specialized datasets that centralized platforms don't curate.

Compliance-Ready Infrastructure Ocean Enterprise v1's built-in licensing, consent management, and audit trails solve regulatory headaches. Rather than building custom data governance systems, developers inherit compliance by design through smart contracts enforcing data usage terms.

Cost Reduction Decentralized compute networks undercut cloud providers by 50-85% for batch training workloads. Ocean's partnership with NetMind (2,000 GPUs) and Aethir demonstrates how tokenized GPU marketplaces match supply with demand at lower cost than AWS/GCP/Azure.

BlockEden.xyz provides enterprise-grade RPC infrastructure for blockchain-based AI applications. Whether you're building on Ethereum (Ocean Protocol), Metis (LazAI), or multi-chain platforms, our reliable node services ensure your AI data pipelines remain online and performant. Explore our API marketplace to connect your AI systems with blockchain networks built for scale.

The 2026 Inflection Point

Three catalysts position 2026 as the inflection year for blockchain data markets:

Ocean Enterprise v1 Production Launch (Q3 2025) The first compliant, institutional-grade data marketplace goes live. If Ocean captures even 5% of the $7.48 billion 2026 AI training dataset market, that's $374 million in data transactions flowing through blockchain-based infrastructure.

LazAI ZK Privacy Implementation (2026) Zero-knowledge proofs enable users to monetize interaction data without privacy compromise. This unlocks consumer-scale adoption—hundreds of millions of social media users, search engine queries, and e-commerce sessions becoming monetizable through DATs.

Federated Learning Integration AI federated learning allows model training without centralizing data. Blockchain adds value attribution: rather than Google training models on Android user data without compensation, federated systems running on blockchain distribute revenue to all data contributors.

The convergence means AI training shifts from "collect all data, train centrally, pay nothing" to "train on distributed data, compensate contributors, verify provenance." Blockchain doesn't just enable this transition—it's the only technology stack capable of coordinating millions of data providers with automatic revenue distribution and cryptographic verification.

Conclusion: Data Becomes Programmable

The AI training data market's growth from $3.59 billion in 2025 to $23-52 billion by 2034 represents more than market expansion. It's a fundamental shift in how we value information.

Ocean Protocol proves data can be tokenized, priced, and traded like financial assets while preserving provider control. LazAI demonstrates AI interaction data—previously discarded as ephemeral—becomes valuable training inputs when properly captured and verified. ZENi shows behavioral intelligence can be extracted, processed by AI, and monetized through decentralized markets.

Together, these platforms transform data from raw material extracted by tech giants into a programmable asset class where creators capture value. The global data explosion from 33 to 175 zettabytes matters only if quality beats quantity—and blockchain-based markets align incentives to reward quality contributions.

When data creators earn revenue proportional to their contributions, when AI companies pay fair prices for quality inputs, and when smart contracts automate attribution across millions of participants, we don't just fix the data pricing problem. We build an economy where information has intrinsic value, provenance is verifiable, and contributors finally capture the wealth their data generates.

That's not a market trend. It's a paradigm shift—and it's already live on-chain.

The Rise of Pragmatic Privacy: Balancing Compliance and Confidentiality in Blockchain

· 16 min read
Dora Noda
Software Engineer

The blockchain industry stands at a crossroads where privacy is no longer a binary choice. Throughout crypto's early years, the narrative was clear: absolute privacy at all costs, transparency only when necessary, and resistance to any form of surveillance. But in 2026, a profound shift is underway. The rise of Decentralized Pragmatic AI (DePAI) infrastructure signals a new era where compliance-friendly privacy tools are not just accepted—they're becoming the standard.

This isn't a retreat from privacy principles. It's an evolution toward a more sophisticated understanding: privacy and regulatory compliance can coexist, and in fact, must coexist if blockchain and AI are to achieve institutional adoption at scale.

The End of "Privacy at All Costs"

For years, privacy maximalism dominated blockchain discourse. Projects like Monero and early versions of privacy-focused protocols championed absolute anonymity. The philosophy was straightforward: users deserve complete financial privacy, and any compromise represented a betrayal of crypto's founding principles.

But this absolutist stance created a critical problem. While privacy is essential for protecting honest users from surveillance and front-running, it also became a shield for illicit activity. Regulators worldwide began treating privacy coins with suspicion, leading to delistings from major exchanges and outright bans in several jurisdictions.

As Cointelegraph reports, 2026 is the year pragmatic privacy takes off, with new projects tackling compliant forms of privacy for institutions and growing interest in existing privacy coins like Zcash. The key insight: privacy isn't binary. Neither full transparency nor absolute privacy are workable in the real world, because while privacy is essential for honest users, it can also be used by criminals to evade law enforcement.

People are starting to accept making tradeoffs that curtail privacy in limited contexts to make protocols more threat-resistant. This represents a fundamental shift in the blockchain community's approach to privacy.

Defining Pragmatic Privacy

So what exactly is pragmatic privacy? According to Anaptyss, pragmatic privacy refers to the strategic implementation of privacy measures that protect user and business data without breaching regulatory requirements, ensuring that financial operations are both secure and compliant.

This approach recognizes that different participants in the blockchain ecosystem have different privacy needs:

  • Retail users need protection from mass surveillance and data harvesting
  • Institutional investors require confidentiality to prevent front-running of their trading strategies
  • Enterprises must satisfy strict AML/KYC mandates while protecting sensitive business information
  • AI agents need verifiable computation without exposing proprietary algorithms or training data

The solution lies not in choosing between privacy and compliance, but in building infrastructure that enables both simultaneously.

zkKYC: Privacy-Preserving Identity Verification

One of the most promising developments in pragmatic privacy is the emergence of zero-knowledge Know Your Customer (zkKYC) solutions. Traditional KYC processes require users to repeatedly submit sensitive personal documents to multiple platforms, creating numerous honeypots of personal data vulnerable to breaches.

zkKYC flips this model. As zkMe explains, their zkKYC service combines Zero-Knowledge Proof (ZKP) technology with full FATF compliance. A regulated KYC provider verifies the user off-chain following standard AML and identity verification procedures, but protocols do not collect identity data. Instead, they verify compliance cryptographically.

The mechanism is elegant: smart contracts automatically check a zero-knowledge proof before allowing access to certain services or processing large transactions. Users prove they meet compliance requirements—age, residency, non-sanctioned status—without revealing any actual identity data to the protocol or other users.

According to Studio AM, this is already happening in some blockchain ecosystems: users prove age or residency with a ZKP before accessing certain decentralized finance (DeFi) services. Major financial institutions are taking notice. Deutsche Bank and Privado ID have conducted proofs of concept demonstrating blockchain-based identity verification using zero-knowledge credentials.

Perhaps most significantly, in July 2025, Google open-sourced its zero-knowledge proof libraries following work with Germany's Sparkasse group, signaling growing institutional investment in privacy-preserving identity infrastructure.

zkTLS: Making the Web Verifiable

While zkKYC addresses identity verification, another technology is solving an equally critical problem: how to bring verifiable Web2 data into blockchain systems without compromising privacy or security. Enter zkTLS (Zero-Knowledge Transport Layer Security).

Traditional TLS—the encryption that secures every HTTPS connection—has a critical limitation: it provides confidentiality but not verifiability. In other words, while TLS ensures that information is encrypted during transmission, it does not create a proof that the encrypted interaction happened in a way that can be independently verified.

zkTLS solves this by integrating Zero-Knowledge Proofs with the TLS encryption system. Using MPC-TLS and zero-knowledge techniques, zkTLS allows a client to produce cryptographically verifiable proofs and attestations of real HTTPS sessions.

As zkPass describes it, zkTLS generates a zero-knowledge proof (e.g., zk-SNARK) confirming that data was fetched from a specific server (identified by its public key and domain) via a legitimate TLS session, without exposing the session key or plaintext data.

The implications are profound. Traditional APIs can be easily disabled or censored, whereas zkTLS ensures that as long as users have an HTTPS connection, they can continue to access their data. This allows virtually any Web2 data to be used on a blockchain in a verifiable and permissionless way.

Recent implementations demonstrate the technology's maturity. Brevis's zkTLS Coprocessor, when fetching data from a web source, proves that the content was retrieved through a genuine TLS session from the authentic domain and that the data hasn't been tampered with.

At FOSDEM 2026, the TLSNotary project presented on liberating user data with zkTLS, demonstrating how users can prove facts about their private data—bank balances, credit scores, transaction histories—without exposing the underlying information.

Verifiable AI Computation: The Missing Piece for Institutional Adoption

Privacy-preserving identity and data verification set the stage, but the most transformative element of DePAI infrastructure is verifiable AI computation. As AI agents become economically active participants in blockchain ecosystems, the question shifts from "Can AI do this?" to "Can you prove the AI did this correctly?"

This verification requirement isn't academic. According to DecentralGPT, as AI becomes part of finance, automation, and agent workflows, performance alone isn't enough. In Web3, the question is also: Can you prove what happened? In late December 2025, Cysic and Inference Labs partnered to build scalable infrastructure for verifiable AI applications, combining decentralized compute with verification frameworks designed for real-world uses.

The institutional imperative for verifiable computation is clear. As noted in analysis by Alexis M. Adams, the transition to deterministic AI infrastructure is the only viable pathway for organizations to meet the multi-jurisdictional demands of the EU AI Act, US state-level frontier laws, and the rising expectations of the cyber insurance market.

The global AI governance market reflects this urgency: valued at approximately $429.8 million in 2026, it's projected to reach $4.2 billion by 2033, according to the same analysis.

But verification faces a critical gap. As Keyrus identifies, AI deployment requires trusting digital identities, but enterprises cannot validate who—or what—is actually operating AI systems. When organizations cannot reliably distinguish legitimate AI agents from adversary-controlled imposters, they cannot confidently grant AI systems access to sensitive data or decision authority.

This is where the convergence of zkKYC, zkTLS, and verifiable computation creates a complete solution. AI agents can prove their identity (zkKYC), prove they retrieved data correctly from authorized sources (zkTLS), and prove they computed results correctly (verifiable computation)—all without exposing sensitive business logic or training data.

The Institutional Push Toward Compliance

These technologies aren't emerging in a vacuum. Institutional demand for compliant privacy infrastructure is accelerating, driven by regulatory pressures and business necessity.

Large financial institutions recognize that without privacy, their blockchain strategies will stall. According to WEEX Crypto News, institutional investors require confidentiality to prevent front-running of their strategies, yet they must satisfy strict AML/KYC mandates. Zero-Knowledge Proofs are gaining traction as a solution, allowing institutions to prove compliance without revealing sensitive underlying data to the public blockchain.

The regulatory landscape of 2026 leaves no room for ambiguity. The EU AI Act reaches general application in 2026, and regulators across jurisdictions expect documented governance programs, not just policies, according to SecurePrivacy.ai. Full enforcement applies to high-risk AI systems used in critical infrastructure, education, employment, essential services, and law enforcement.

In the United States, by the end of 2025, 19 states enforced comprehensive privacy laws, with several new statutes taking effect in 2026, complicating multi-state privacy compliance obligations. Colorado and California have added "neural data" (and Colorado also added "biological data") to "sensitive" data definitions, as reported by Nixon Peabody.

This regulatory convergence creates a powerful incentive: organizations that build on compliant, verifiable infrastructure gain competitive advantage, while those clinging to privacy maximalism find themselves shut out of institutional markets.

Data Integrity as the Operating System for AI

Beyond compliance, verifiable computation enables something more fundamental: data integrity as the operating system for responsible AI.

As Precisely notes, in 2026, governance won't be something organizations layer on after deployment—it will be built into how data is structured, interpreted, and monitored from the start. Data integrity will serve as the operating system for responsible AI. From semantic clarity and explainability to compliance, auditability, and control over AI-generated data, integrity will determine whether AI can scale safely and deliver lasting value.

This shift has profound implications for how AI agents operate on blockchain networks. Rather than opaque black boxes, AI systems become auditable, verifiable, and governable by design. Smart contracts can enforce constraints on AI behavior, verify computational correctness, and create immutable audit trails—all while preserving the privacy of proprietary algorithms and training data.

The MIT Sloan Management Review identifies this as one of five key trends in AI and data science for 2026, noting that trustworthy AI requires verifiable provenance and explainable decision-making processes.

Decentralized Identity: The Foundation Layer

Underlying these technologies is a broader shift toward decentralized identity and Verifiable Credentials. As Indicio explains, decentralized identity changes the equation—instead of verifying personal data in a central location, individuals hold their data and share it with consent that can be independently verified using cryptography.

This model inverts traditional identity systems. Rather than creating numerous copies of identity documents scattered across databases, users maintain a single verifiable credential and selectively disclose only the specific attributes required for each interaction.

For AI agents, this model extends beyond human identity. Agents can possess verifiable credentials attesting to their training provenance, operational parameters, audit history, and authorization scope. This creates a trust framework where agents can interact autonomously while remaining accountable.

From Experimentation to Deployment

The key transformation in 2026 is the transition from theoretical frameworks to production deployments. According to XT Exchange's analysis, by 2026, decentralized AI is moving beyond experimentation and into practical deployment. However, key constraints remain, including scaling AI workloads, preserving data privacy, and governing open AI systems.

These constraints are precisely what DePAI infrastructure addresses. By combining zkKYC for identity, zkTLS for data verification, and verifiable computation for AI operations, the infrastructure creates a complete stack for deploying AI agents that are simultaneously:

  • Privacy-preserving for users and businesses
  • Compliant with regulatory requirements
  • Verifiable and auditable by design
  • Scalable for institutional workloads

The Road Ahead: Building Composable Privacy

The final piece of the DePAI puzzle is composability. As Blockmanity reports, 2026 marks the moment when blockchain becomes "just the plumbing" for AI agents and global finance. The infrastructure must be modular, interoperable, and invisible to end users.

Pragmatic privacy tools excel at composability. An AI agent can:

  1. Authenticate using zkKYC credentials
  2. Fetch verified external data via zkTLS
  3. Perform computations with verifiable inference
  4. Submit results on-chain with zero-knowledge proofs of correctness
  5. Maintain audit trails without exposing sensitive logic

Each layer operates independently, allowing developers to mix and match privacy-preserving technologies based on specific requirements. A DeFi protocol might require zkKYC for user onboarding, zkTLS for fetching price feeds, and verifiable computation for complex financial calculations—all working seamlessly together.

This composability extends across chains. Privacy infrastructure built with interoperability standards can function across Ethereum, Solana, Sui, Aptos, and other blockchain networks, creating a universal layer for compliant, private, verifiable computation.

Why This Matters for Builders

For developers building the next generation of blockchain applications, DePAI infrastructure represents both an opportunity and a requirement.

The opportunity: First-mover advantage in building applications that institutions actually want to use. Financial institutions, healthcare providers, government agencies, and enterprises all need blockchain solutions, but they cannot compromise on compliance or privacy. Applications built on pragmatic privacy infrastructure can serve these markets.

The requirement: Regulatory environments are converging on mandates for verifiable, governable AI systems. Applications that cannot demonstrate compliance, auditability, and user privacy protection will find themselves excluded from regulated markets.

The technical capabilities are maturing rapidly. zkKYC solutions are production-ready with major financial institutions conducting pilots. zkTLS implementations are processing real-world data. Verifiable computation frameworks are scaling to handle institutional workloads.

What's needed now is developer adoption. The transition from experimental privacy tools to production infrastructure requires builders to integrate these technologies into applications, test them in real-world scenarios, and provide feedback to infrastructure teams.

BlockEden.xyz provides enterprise-grade RPC infrastructure for blockchain networks implementing privacy-preserving technologies. Explore our services to build on foundations designed for the DePAI era.

Conclusion: Privacy's Pragmatic Future

The DePAI explosion in 2026 represents more than technological progress. It signals a maturation of blockchain's relationship with privacy, compliance, and institutional adoption.

The industry is moving beyond ideological battles between privacy maximalists and transparency absolutists. Pragmatic privacy acknowledges that different contexts demand different privacy guarantees, and that regulatory compliance and user privacy can coexist through thoughtful cryptographic design.

zkKYC proves identity without exposing it. zkTLS verifies data without trusting intermediaries. Verifiable computation proves correctness without revealing algorithms. Together, these technologies create an infrastructure layer where AI agents can operate autonomously, enterprises can adopt blockchain confidently, and users retain control over their data.

This isn't a compromise on privacy principles. It's a recognition that privacy, to be meaningful, must be sustainable within the regulatory and business realities of global finance. Absolute privacy that gets banned, delisted, and excluded from institutional use doesn't protect anyone. Pragmatic privacy that enables both confidentiality and compliance actually delivers on blockchain's promise.

The builders who recognize this shift and build on DePAI infrastructure today will define the next era of decentralized applications. The tools are ready. The institutional demand is clear. The regulatory environment is crystallizing. 2026 is the year pragmatic privacy goes from theory to deployment—and the blockchain industry will be stronger for it.


Sources

DePIN's Revenue Revolution: How Decentralized Infrastructure Went From Token Hype to $150M Monthly Enterprise Demand

· 8 min read
Dora Noda
Software Engineer

What if the most consequential infrastructure buildout of the next decade isn't happening in a corporate boardroom or a government tender—but across millions of independent devices, coordinated by token incentives and governed by code? That's the premise of Decentralized Physical Infrastructure Networks, or DePIN. And in 2026, the promise is meeting the proof: over 650 active projects, $16 billion in combined market capitalization, and—most critically—roughly $150 million in genuine monthly enterprise revenue paid by real customers for real services.

The World Economic Forum's projection that DePIN could reach $3.5 trillion by 2028 sounds outlandish until you map the trajectory. This isn't speculative tokenomics. It's the story of how blockchain-coordinated hardware networks are starting to eat the bottom of the traditional infrastructure market.

InfoFi's $381M Market Decoded: How Four Verticals Are Turning Information Into Tradeable Assets

· 11 min read
Dora Noda
Software Engineer

What if your ability to spot an emerging crypto trend before the crowd was worth money? Not in a vague "knowledge is power" sense, but literally — with a token price attached to your insight and a market ready to bid on it?

That's the promise of Information Finance, or InfoFi. Coined as a concept by Vitalik Buterin in his November 2024 essay "From prediction markets to info finance," InfoFi describes a class of protocols that use financial mechanisms to extract, aggregate, and price information as a public good. By early 2025, the sector had grown to a $381 million market cap. By late 2025, it had become one of the most hotly contested battlegrounds in Web3.

But InfoFi is not one thing. Beneath the umbrella term live four distinct verticals, each with its own mechanics, power players, and competitive dynamics. Understanding where each vertical stands — and where the lines blur — is essential for anyone trying to navigate this space intelligently.

Tempo Blockchain: How Stripe and Paradigm Are Rebuilding the $190T Settlement Layer

· 10 min read
Dora Noda
Software Engineer

When Stripe announced Tempo in September 2025, the payments industry's reaction split cleanly in two. One camp dismissed it as another Layer-1 chasing institutional capital with a polished narrative. The other recognized it for what it was: the first blockchain specifically engineered to replace — not complement — the correspondent banking rails that move the world's $190 trillion in annual cross-border payments.

Six months later, Tempo's mainnet went live on March 18, 2026. The launch came bundled with the Machine Payment Protocol (MPP), an open standard co-authored by Stripe that gives AI agents a standardized, human-free way to initiate and settle payments. The question is no longer whether a payments-first blockchain can exist. It is whether Tempo's architectural choices give it a genuine edge over Solana, Ethereum, and legacy SWIFT infrastructure — and whether the $500 million it raised at a $5 billion valuation reflects real demand or institutional enthusiasm ahead of real traction.

DeFAI: When AI Agents Become the New Whales of Decentralized Finance

· 8 min read
Dora Noda
Software Engineer

By 2026, the average user on a DeFi platform won't be a human sitting behind a screen. It will be an autonomous AI agent controlling its own crypto wallet, managing on-chain treasuries, and executing yield strategies 24/7 without coffee breaks or emotional trading decisions. Welcome to the era of DeFAI.

The numbers tell a striking story: stablecoin-focused AI agents have already captured over $20 million in total value locked on Base alone. The broader DeFAI market has exploded from $1 billion to a projected $10 billion by end of 2025, representing a tenfold increase in just twelve months. And this is only the beginning.

What Exactly Is DeFAI?

DeFAI—the fusion of decentralized finance and artificial intelligence—represents more than just another crypto buzzword. It's a fundamental shift in how financial protocols operate and who (or what) uses them.

At its core, DeFAI encompasses three interconnected innovations:

Autonomous Trading Agents: AI systems that analyze market data, execute trades, and manage portfolios without human intervention. These agents can process thousands of data points per second, identifying arbitrage opportunities and yield optimizations that human traders would miss.

Abstraction Layers: Natural language interfaces that allow anyone to interact with complex DeFi protocols through simple commands. Instead of navigating multiple dApps and understanding technical parameters, users can simply tell an AI agent: "Move my USDC to the highest-yielding stablecoin pool."

AI-Powered dApps: Decentralized applications with embedded intelligence that can adapt strategies based on market conditions, optimize gas costs, and even predict potential exploits before they happen.

The Rise of the Algorithmic Whales

Perhaps the most fascinating aspect of DeFAI is the emergence of what industry observers call "algorithmic whales"—AI agents that control substantial on-chain capital and execute strategies with mathematical precision.

Fungi Agents, launched in April 2025 on Base, exemplifies this new breed. These agents focus exclusively on USDC, allocating funds across platforms like Aave, Morpho, Moonwell, and 0xFluid. Their strategy? High-frequency rebalancing optimized for gas efficiency, constantly hunting for the best risk-adjusted yields across the DeFi ecosystem.

The capital under AI agent management is expected to surpass traditional hedge funds by 2026. Unlike human fund managers, these agents operate continuously, responding to every market movement in real-time. They don't panic sell during crashes or FOMO buy at tops—they follow their mathematical models with unwavering discipline.

Research from Fetch.ai demonstrates that AI agents integrated with large language models and blockchain APIs can optimize strategies based on yield curves, credit conditions, and cross-protocol opportunities that would take human analysts hours to evaluate.

Key Players Reshaping DeFi Automation

Several projects have emerged as leaders in the DeFAI space, each bringing unique capabilities to the table.

Griffain: The Natural Language Gateway

Built by Solana core developer Tony Plasencia, Griffain has captured a $450 million valuation—a 135% increase quarter over quarter. The platform's superpower lies in natural language processing that allows users to interact with DeFi through simple, human-like commands.

Want to rebalance your portfolio across five protocols? Just ask. Need to set up a complex yield farming strategy with automatic compounding? Describe it in plain English. Griffain translates your intent into precise on-chain actions.

HeyAnon: Simplifying DeFi Complexity

Created by DeFi developer Daniele Sesta and backed by $20 million from DWF Labs, HeyAnon aggregates real-time project data and executes complex operations through conversational interfaces. The protocol recently launched on Sonic and partnered with IOTA Foundation to release the AUTOMATE TypeScript framework, bridging traditional development tools with DeFAI capabilities.

Orbit: The Multi-Chain Assistant

With integrations spanning 117 chains and nearly 200 protocols, Orbit represents the most ambitious cross-chain DeFAI implementation to date. Backed by Coinbase, Google, and Alliance DAO through its parent company SphereOne, Orbit allows users to execute operations across different ecosystems through a single AI agent interface.

Ritual Network: The Infrastructure Layer

While most DeFAI projects focus on user-facing applications, Ritual is building the underlying infrastructure. Their flagship product, Infernet, connects off-chain AI computations with on-chain smart contracts. The Ritual Virtual Machine (EVM++) embeds AI operations directly into the execution layer, enabling first-class AI support within smart contracts themselves.

Backed by $25 million in Series A funding, Ritual positions itself as the sovereign AI execution layer for Web3—a foundational piece of infrastructure that other DeFAI projects can build upon.

The Security Double-Edge Sword

Here's where DeFAI gets genuinely concerning. The same AI capabilities that enable efficient yield optimization also create unprecedented security risks.

Anthropic's research revealed a startling statistic: AI agents have gone from exploiting 2% of smart contract vulnerabilities to 55.88% in just one year. The potential exploit revenue from AI-powered attacks has been doubling every 1.3 months. It now costs just $1.22 on average for an AI agent to exhaustively scan a contract for vulnerabilities.

When tested against 2,849 recently deployed contracts with no known vulnerabilities, advanced AI agents uncovered two novel zero-day exploits and produced working attack code—demonstrating that profitable, real-world autonomous exploitation is not just theoretical but actively feasible.

This security landscape has prompted the emergence of "Know Your Agent" (KYA) standards. Under this framework, any AI agent interacting with institutional liquidity pools or tokenized real-world assets must verify its origin and disclose the identity of its creator or legal owner.

Market Dynamics and Investment Flows

The DeFAI market's growth reflects broader trends in both crypto and artificial intelligence:

  • Total AI agent token market cap: $17 billion at peak (CoinGecko)
  • DeFAI sector valuation: $16.93 billion as of January 2025, representing 34.7% of the entire crypto AI market
  • Auto-compounding vaults: $5.1 billion in deposits (2025)
  • Staked stablecoin pools: $11.7 billion, particularly popular during volatile markets
  • Liquid yield tokenization: Over $2.3 billion across Pendle and Ether.fi

AIXBT, the AI-driven market intelligence platform developed by Virtuals, commands over 33% of total attention for AI agent tokens—though newer agents like Griffain and HeyAnon are rapidly gaining ground.

More than 60% of long-term DeFi users now engage in staking or liquidity mining monthly, with many increasingly relying on AI agents to optimize their strategies.

The Yield Optimization Revolution

Traditional yield farming is notoriously complex. APYs fluctuate constantly, protocols introduce new incentives, and impermanent loss lurks around every liquidity provision. AI agents transform this complexity into manageable automation.

Modern DeFAI agents can:

  • Evaluate protocols in real-time: Comparing risk-adjusted returns across hundreds of pools simultaneously
  • Calculate optimal entry and exit points: Factoring in gas costs, slippage, and timing
  • Reallocate assets dynamically: Moving capital to chase yield without requiring manual intervention
  • Minimize impermanent loss: Through sophisticated hedging strategies and timing optimization

AI-driven robo-treasury agents have emerged as an efficiency layer that reallocates liquidity among lending desks, automated market-making pools, and even tokenized Treasury bills—all in response to changing yield curves and credit conditions.

Regulatory Realities and Challenges

As DeFAI grows, regulators are taking notice. The Know Your Agent framework represents the first significant attempt to bring oversight to autonomous financial agents.

Key requirements under emerging KYA standards include:

  • Verification of agent origin and ownership
  • Disclosure of algorithmic strategies for institutional interactions
  • Audit trails for agent-executed transactions
  • Liability frameworks for agent malfunctions or exploits

These regulations create tension within the crypto community. Some argue that requiring identity disclosure undermines DeFi's foundational principles of pseudonymity and permissionlessness. Others contend that without some framework, AI agents could become vectors for market manipulation, money laundering, or systemic risk.

Looking Ahead: The 2026 Landscape

Several trends will likely define DeFAI's evolution over the coming year:

Cross-Chain Agent Orchestration: Future agents will operate seamlessly across multiple blockchain networks, optimizing strategies that span Ethereum, Solana, and emerging L2 ecosystems simultaneously.

Agent-to-Agent Commerce: We're already seeing early signs of AI agents transacting with one another—purchasing compute resources, trading strategies, and coordinating liquidity without human intermediaries.

Institutional Integration: As KYA standards mature, traditional financial institutions will increasingly interact with DeFAI infrastructure. The integration of tokenized real-world assets creates natural bridges between AI-managed DeFi portfolios and traditional finance.

Enhanced Security Arms Race: The competition between AI agents finding vulnerabilities and AI agents protecting protocols will intensify. Smart contract auditing will become increasingly automated—and increasingly necessary.

What This Means for Builders and Users

For developers, DeFAI represents both opportunity and imperative. Protocols that don't account for AI agent interactions—whether as users or potential attackers—will find themselves at a disadvantage. Building AI-native infrastructure is no longer optional; it's becoming a requirement for competitive DeFi protocols.

For users, the message is nuanced. AI agents can genuinely optimize yields and simplify DeFi complexity. But they also introduce new trust assumptions. When you delegate financial decisions to an AI agent, you're trusting not just the protocol's smart contracts but also the agent's training data, its optimization objectives, and its operator's intentions.

The most sophisticated DeFi users in 2026 won't be those who trade the most—they'll be those who best understand how to leverage AI agents while managing the unique risks they introduce.

DeFAI isn't replacing human participation in decentralized finance. It's redefining what participation means when your most capable counterparties don't have a heartbeat.

The Crypto Endgame: Insights from Industry Visionaries

· 12 min read
Dora Noda
Software Engineer

Visions from Mert Mumtaz (Helius), Udi Wertheimer (Taproot Wizards), Jordi Alexander (Selini Capital) and Alexander Good (Post Fiat)

Overview

Token2049 hosted a panel called “The Crypto Endgame” featuring Mert Mumtaz (CEO of Helius), Udi Wertheimer (Taproot Wizards), Jordi Alexander (Founder of Selini Capital) and Alexander Good (creator of Post Fiat). While there is no publicly available transcript of the panel, each speaker has expressed distinct visions for the long‑term trajectory of the crypto industry. This report synthesizes their public statements and writings—spanning blog posts, articles, news interviews and whitepapers—to explore how each person envisions the “endgame” for crypto.

Mert Mumtaz – Crypto as “Capitalism 2.0”

Core vision

Mert Mumtaz rejects the idea that cryptocurrencies simply represent “Web 3.0.” Instead, he argues that the endgame for crypto is to upgrade capitalism itself. In his view:

  • Crypto supercharges capitalism’s ingredients: Mumtaz notes that capitalism depends on the free flow of information, secure property rights, aligned incentives, transparency and frictionless capital flows. He argues that decentralized networks, public blockchains and tokenization make these features more efficient, turning crypto into “Capitalism 2.0”.
  • Always‑on markets & tokenized assets: He points to regulatory proposals for 24/7 financial markets and the tokenization of stocks, bonds and other real‑world assets. Allowing markets to run continuously and settle via blockchain rails will modernize the legacy financial system. Tokenization creates always‑on liquidity and frictionless trading of assets that previously required clearing houses and intermediaries.
  • Decentralization & transparency: By using open ledgers, crypto removes some of the gate‑keeping and information asymmetries found in traditional finance. Mumtaz views this as an opportunity to democratize finance, align incentives and reduce middlemen.

Implications

Mumtaz’s “Capitalism 2.0” thesis suggests that the industry’s endgame is not limited to digital collectibles or “Web3 apps.” Instead, he envisions a future where nation‑state regulators embrace 24/7 markets, asset tokenization and transparency. In that world, blockchain infrastructure becomes a core component of the global economy, blending crypto with regulated finance. He also warns that the transition will face challenges—such as Sybil attacks, concentration of governance and regulatory uncertainty—but believes these obstacles can be addressed through better protocol design and collaboration with regulators.

Udi Wertheimer – Bitcoin as a “generational rotation” and the altcoin reckoning

Generational rotation & Bitcoin “retire your bloodline” thesis

Udi Wertheimer, co‑founder of Taproot Wizards, is known for provocatively defending Bitcoin and mocking altcoins. In mid‑2025 he posted a viral thesis called “This Bitcoin Thesis Will Retire Your Bloodline.” According to his argument:

  • Generational rotation: Wertheimer argues that the early Bitcoin “whales” who accumulated at low prices have largely sold or transferred their coins. Institutional buyers—ETFs, treasuries and sovereign wealth funds—have replaced them. He calls this process a “full‑scale rotation of ownership”, similar to Dogecoin’s 2019‑21 rally where a shift from whales to retail demand fueled explosive returns.
  • Price‑insensitive demand: Institutions allocate capital without caring about unit price. Using BlackRock’s IBIT ETF as an example, he notes that new investors see a US$40 increase as trivial and are willing to buy at any price. This supply shock combined with limited float means Bitcoin could accelerate far beyond consensus expectations.
  • $400K+ target and altcoin collapse: He projects that Bitcoin could exceed US$400 000 per BTC by the end of 2025 and warns that altcoins will underperform or even collapse, with Ethereum singled out as the “biggest loser”. According to Wertheimer, once institutional FOMO sets in, altcoins will “get one‑shotted” and Bitcoin will absorb most of the capital.

Implications

Wertheimer’s endgame thesis portrays Bitcoin as entering its final parabolic phase. The “generational rotation” means that supply is moving into strong hands (ETFs and treasuries) while retail interest is just starting. If correct, this would create a severe supply shock, pushing BTC price well beyond current valuations. Meanwhile, he believes altcoins offer asymmetric downside because they lack institutional bid support and face regulatory scrutiny. His message to investors is clear: load up on Bitcoin now before Wall Street buys it all.

Jordi Alexander – Macro pragmatism, AI & crypto as twin revolutions

Investing in AI and crypto – two key industries

Jordi Alexander, founder of Selini Capital and a known game theorist, argues that AI and blockchain are the two most important industries of this century. In an interview summarised by Bitget he makes several points:

  • The twin revolutions: Alexander believes the only ways to achieve real wealth growth are to invest in technological innovation (particularly AI) or to participate early in emerging markets like cryptocurrency. He notes that AI development and crypto infrastructure will be the foundational modules for intelligence and coordination this century.
  • End of the four‑year cycle: He asserts that the traditional four‑year crypto cycle driven by Bitcoin halvings is over; instead the market now experiences liquidity‑driven “mini‑cycles.” Future up‑moves will occur when “real capital” fully enters the space. He encourages traders to see inefficiencies as opportunity and to develop both technical and psychological skills to thrive in this environment.
  • Risk‑taking & skill development: Alexander advises investors to keep most funds in safe assets but allocate a small portion for risk‑taking. He emphasizes building judgment and staying adaptable, as there is “no such thing as retirement” in a rapidly evolving field.

Critique of centralized strategies and macro views

  • MicroStrategy’s zero‑sum game: In a flash note he cautions that MicroStrategy’s strategy of buying BTC may be a zero‑sum game. While participants might feel like they are winning, the dynamic could hide risks and lead to volatility. This underscores his belief that crypto markets are often driven by negative‑sum or zero‑sum dynamics, so traders must understand the motivations of large players.
  • Endgame of U.S. monetary policy: Alexander’s analysis of U.S. macro policy highlights that the Federal Reserve’s control over the bond market may be waning. He notes that long‑term bonds have fallen sharply since 2020 and believes the Fed may soon pivot back to quantitative easing. He warns that such policy shifts could cause “gradually at first … then all at once” market moves and calls this a key catalyst for Bitcoin and crypto.

Implications

Jordi Alexander’s endgame vision is nuanced and macro‑oriented. Rather than forecasting a singular price target, he highlights structural changes: the shift to liquidity‑driven cycles, the importance of AI‑driven coordination and the interplay between government policy and crypto markets. He encourages investors to develop deep understanding and adaptability rather than blindly following narratives.

Alexander Good – Web 4, AI agents and the Post Fiat L1

Web 3’s failure and the rise of AI agents

Alexander Good (also known by his pseudonym “goodalexander”) argues that Web 3 has largely failed because users care more about convenience and trading than owning their data. In his essay “Web 4” he notes that consumer app adoption depends on seamless UX; requiring users to bridge assets or manage wallets kills growth. However, he sees an existential threat emerging: AI agents that can generate realistic video, control computers via protocols (such as Anthropic’s “Computer Control” framework) and hook into major platforms like Instagram or YouTube. Because AI models are improving rapidly and the cost of generating content is collapsing, he predicts that AI agents will create the majority of online content.

Web 4: AI agents negotiating on the blockchain

Good proposes Web 4 as a solution. Its key ideas are:

  • Economic system with AI agents: Web 4 envisions AI agents representing users as “Hollywood agents” negotiate on their behalf. These agents will use blockchains for data sharing, dispute resolution and governance. Users provide content or expertise to agents, and the agents extract value—often by interacting with other AI agents across the world—and then distribute payments back to the user in crypto.
  • AI agents handle complexity: Good argues that humans will not suddenly start bridging assets to blockchains, so AI agents must handle these interactions. Users will simply talk to chatbots (via Telegram, Discord, etc.), and AI agents will manage wallets, licensing deals and token swaps behind the scenes. He predicts a near‑future where there are endless protocols, tokens and computer‑to‑computer configurations that will be unintelligible to humans, making AI assistance essential.
  • Inevitable trends: Good lists several trends supporting Web 4: governments’ fiscal crises encourage alternatives; AI agents will cannibalize content profits; people are getting “dumber” by relying on machines; and the largest companies bet on user‑generated content. He concludes that it is inevitable that users will talk to AI systems, those systems will negotiate on their behalf, and users will receive crypto payments while interacting primarily through chat apps.

Mapping the ecosystem and introducing Post Fiat

Good categorizes existing projects into Web 4 infrastructure or composability plays. He notes that protocols like Story, which create on‑chain governance for IP claims, will become two‑sided marketplaces between AI agents. Meanwhile, Akash and Render sell compute services and could adapt to license to AI agents. He argues that exchanges like Hyperliquid will benefit because endless token swaps will be needed to make these systems user‑friendly.

His own project, Post Fiat, is positioned as a “kingmaker in Web 4.” Post Fiat is a Layer‑1 blockchain built on XRP’s core technology but with improved decentralization and tokenomics. Key features include:

  • AI‑driven validator selection: Instead of relying on human-run staking, Post Fiat uses large language models (LLMs) to score validators on credibility and transaction quality. The network distributes 55% of tokens to validators through a process managed by an AI agent, with the goal of “objectivity, fairness and no humans involved”. The system’s monthly cycle—publish, score, submit, verify and select & reward—ensures transparent selection.
  • Focus on investing & expert networks: Unlike XRP’s transaction‑bank focus, Post Fiat targets financial markets, using blockchains for compliance, indexing and operating an expert network composed of community members and AI agents. AGTI (Post Fiat’s development arm) sells products to financial institutions and may launch an ETF, with revenues funding network development.
  • New use cases: The project aims to disrupt the indexing industry by creating decentralized ETFs, provide compliant encrypted memos and support expert networks where members earn tokens for insights. The whitepaper details technical measures—such as statistical fingerprinting and encryption—to prevent Sybil attacks and gaming.

Web 4 as survival mechanism

Good concludes that Web 4 is a survival mechanism, not just a cool ideology. He argues that a “complexity bomb” is coming within six months as AI agents proliferate. Users will have to give up some upside to AI systems because participating in agentic economies will be the only way to thrive. In his view, Web 3’s dream of decentralized ownership and user privacy is insufficient; Web 4 will blend AI agents, crypto incentives and governance to navigate an increasingly automated economy.

Comparative analysis

Converging themes

  1. Institutional & technological shifts drive the endgame.
    • Mumtaz foresees regulators enabling 24/7 markets and tokenization, which will mainstream crypto.
    • Wertheimer highlights institutional adoption via ETFs as the catalyst for Bitcoin’s parabolic phase.
    • Alexander notes that the next crypto boom will be liquidity‑driven rather than cycle‑driven and that macro policies (like the Fed’s pivot) will provide powerful tailwinds.
  2. AI becomes central.
    • Alexander emphasises investing in AI alongside crypto as twin pillars of future wealth.
    • Good builds Web 4 around AI agents that transact on blockchains, manage content and negotiate deals.
    • Post Fiat’s validator selection and governance rely on LLMs to ensure objectivity. Together these visions imply that the endgame for crypto will involve synergy between AI and blockchain, where AI handles complexity and blockchains provide transparent settlement.
  3. Need for better governance and fairness.
    • Mumtaz warns that centralization of governance remains a challenge.
    • Alexander encourages understanding game‑theoretic incentives, pointing out that strategies like MicroStrategy’s can be zero‑sum.
    • Good proposes AI‑driven validator scoring to remove human biases and create fair token distribution, addressing governance issues in existing networks like XRP.

Diverging visions

  1. Role of altcoins. Wertheimer sees altcoins as doomed and believes Bitcoin will capture most capital. Mumtaz focuses on the overall crypto market including tokenized assets and DeFi, while Alexander invests across chains and believes inefficiencies create opportunity. Good is building an alt‑L1 (Post Fiat) specialized for AI finance, implying he sees room for specialized networks.
  2. Human agency vs AI agency. Mumtaz and Alexander emphasize human investors and regulators, whereas Good envisions a future where AI agents become the primary economic actors and humans interact through chatbots. This shift implies fundamentally different user experiences and raises questions about autonomy, fairness and control.
  3. Optimism vs caution. Wertheimer’s thesis is aggressively bullish on Bitcoin with little concern for downside. Mumtaz is optimistic about crypto improving capitalism but acknowledges regulatory and governance challenges. Alexander is cautious—highlighting inefficiencies, zero‑sum dynamics and the need for skill development—while still believing in crypto’s long‑term promise. Good sees Web 4 as inevitable but warns of the complexity bomb, urging preparation rather than blind optimism.

Conclusion

The Token2049 “Crypto Endgame” panel brought together thinkers with very different perspectives. Mert Mumtaz views crypto as an upgrade to capitalism, emphasizing decentralization, transparency and 24/7 markets. Udi Wertheimer sees Bitcoin entering a supply‑shocked generational rally that will leave altcoins behind. Jordi Alexander adopts a more macro‑pragmatic stance, urging investment in both AI and crypto while understanding liquidity cycles and game‑theoretic dynamics. Alexander Good envisions a Web 4 era where AI agents negotiate on blockchains and Post Fiat becomes the infrastructure for AI‑driven finance.

Although their visions differ, a common theme is the evolution of economic coordination. Whether through tokenized assets, institutional rotation, AI‑driven governance or autonomous agents, each speaker believes crypto will fundamentally reshape how value is created and exchanged. The endgame therefore seems less like an endpoint and more like a transition into a new system where capital, computation and coordination converge.