As a data engineer who’s spent the last few years building pipelines for blockchain analytics, I’ve been closely watching The Graph’s evolution. Their 2026 technical roadmap just dropped, and it’s making me rethink everything about how we approach blockchain data infrastructure.
What The Graph Just Announced
The Graph is pivoting from being “just” a subgraph indexing network into a full-scale multi-service data platform. They’re launching six specialized products on their Horizon protocol layer:
- Subgraphs - Enhanced developer support with cost and scaling efficiencies
- Substreams - High-performance, low-latency data streaming for DeFi protocols, DePIN, AI infrastructure, and institutional analytics
- Token API - Multi-chain support (currently 10 chains, expanding)
- Tycho - Real-time DeFi liquidity aggregation that provides consistent pricing/quotes across multiple DEXs
- SQL Platform (Amp) - Enterprise-grade analytics engine for institutional use cases
- AI Services - Natural language query integration with Claude, ChatGPT, and Cursor through Subgraph MCP and agent-to-agent (A2A) protocols
After processing 1.27 trillion queries, they’re positioning themselves as the data backbone for a projected $47 billion agentic AI economy.
The AI Agent Shift
Here’s what really caught my attention: 37% of new Token API users are AI agents, not human developers.
Think about that for a second. More than a third of their new users aren’t people—they’re autonomous systems querying blockchain data. The x402 protocol they’re building enables AI agents to autonomously query the network and pay per-query without requiring manual setup or API key management.
Q2 2026 will deliver the x402-compliant gateway with full AI support. We’re talking about AI agents that can query blockchain data through natural language, get structured responses, and pay for each query automatically.
The Core Question: Infrastructure or Just Better APIs?
This is where I start to get conflicted. On one hand, The Graph is solving real problems:
- Indexing is hard - Running your own nodes, writing custom indexers, maintaining infrastructure 24/7
- Consistency across chains - Their abstraction layer works across 10+ blockchains with a unified interface
- Decentralization - Unlike running a centralized indexer, The Graph’s protocol distributes the work
But on the other hand… is this essential infrastructure or expensive middleware?
When I look at competing solutions (Covalent, Ormi, Envio, Goldsky, Ponder), they’re all solving similar problems with different trade-offs. Some are faster, some are cheaper, some offer more customization. The Graph’s differentiator is now “AI-queryable” and “natural language interfaces.”
But if you strip away the AI buzzwords, they’re fundamentally doing what blockchain indexers have always done: reading blocks, parsing events, storing structured data, and exposing query interfaces.
The Build vs Buy Analysis
As someone who’s built custom indexers and used The Graph, here’s my cost-benefit breakdown:
When The Graph makes sense:
- You need multi-chain support without maintaining nodes for each network
- Your team is small and can’t dedicate engineers to infrastructure
- You value decentralization and censorship resistance
- You need proven reliability (they’ve handled 1.27T queries)
When building in-house makes sense:
- You have specific performance requirements The Graph can’t meet
- Your queries are complex and don’t fit the GraphQL model well
- You’re analyzing proprietary data or need custom transformations
- At scale, you can amortize infrastructure costs across high query volume
For my current project analyzing MEV patterns, I ended up running a hybrid setup: using The Graph for standard queries (token transfers, DEX swaps) but running custom indexers for specialized MEV detection that requires microsecond precision and custom algorithms.
What About This “AI-Queryable” Angle?
The natural language query integration is undeniably convenient. Instead of writing GraphQL, you could theoretically ask Claude to show you the top tokens by trading volume.
But here’s my concern: convenience that hides complexity can be dangerous in production systems.
When you’re building financial applications on blockchain data, you need to understand exactly what data you’re getting, how fresh it is, what assumptions the indexer made, and what could go wrong. Natural language queries abstract away those details.
Maybe that’s fine for prototyping or internal analytics. But for production systems handling real money? I want explicit queries where I control every parameter.
My Take: It’s Complicated
After thinking through this, I don’t think it’s binary. The Graph isn’t “just APIs” and it’s not pure essential infrastructure either. It’s commodity infrastructure with increasing specialization.
The parallel I’d draw is AWS. You can run your own data centers. Many companies did in the early 2000s. But AWS commoditized infrastructure and let companies focus on their actual business logic. The trade-off was vendor dependence and less control.
The Graph is making a similar bet: that most blockchain applications don’t need custom data infrastructure. They need reliable, decentralized access to blockchain data so they can focus on building products.
The question is whether blockchain data access becomes a commodity like cloud computing, or whether it remains specialized enough that custom solutions win.
What Do You Think?
For those of you building on blockchain data:
- Are you using The Graph, competitors, or building your own indexers?
- How do you evaluate the build vs buy decision for data infrastructure?
- Does the AI integration genuinely add value, or is it just marketing?
- At what point does middleware become so essential that it’s infrastructure?
I’m genuinely curious about your experiences and how you’re approaching this.
Sources: