I’ve been analyzing RPC traffic patterns for the past few months, and I recently came across some data from Syndica that really made me pause and think about what we’re actually building here.
The stat that caught my attention: 96% of all calls made to Solana nodes are read operations. For every transaction submission (write), there are 25 read calls—balance checks, transaction lookups, account queries, program state reads.
This isn’t unique to Solana. When I looked at Ethereum and other chains, the pattern holds. The vast majority of blockchain infrastructure usage is reading data, not writing it.
So what does this mean?
On one hand, you could argue we’ve built the world’s most expensive, slowest database. Traditional databases are optimized for read-heavy workloads with techniques like caching, indexing, and read replicas. They handle billions of queries per second at millisecond latency.
Blockchains? We’re celebrating thousands of TPS and sub-second finality.
But here’s where it gets interesting: maybe that’s exactly the point.
The Infrastructure Response
What I find fascinating is how the industry is responding to this reality. Syndica built Sig, a read-optimized Solana validator client written in Zig specifically because they recognized this 96% read pattern. Early benchmarks show 50-70% performance improvements, with some operations running 1.5-4x faster than existing implementations.
RPC providers are splitting into specialized services:
- Read-optimized nodes: Archive nodes, fast-path routing for balance checks and state queries
- Write-optimized infrastructure: Focused on transaction submission, mempool optimization, MEV protection
- Hybrid approaches: Trying to balance both
The Database Question
Here’s my question for the community: If blockchain usage is 96% reads, are we just building really slow, really expensive databases with consensus as a feature?
Or is the correct mental model something else entirely? Maybe blockchains are less like databases and more like:
- Public ledgers where write scarcity is the feature, and reads are meant to be cheap/abundant
- Shared state machines where consensus on writes is valuable, but reads are just local queries
- Verification systems where anyone can independently verify writes, making reads inherently cheap
What I’m watching
As a data engineer, I’m tracking a few developments:
-
Client specialization: Will we see more read-optimized clients like Sig? What about write-optimized consensus clients?
-
Pricing models: Current RPC pricing often charges for read volume, but if reads are 96% of usage and should be cheap, does this pricing make sense?
-
Infrastructure architecture: Should we separate read and write infrastructure entirely? What are the security implications?
-
Developer experience: Most devs don’t think about read vs write optimization when building dApps—should they?
My take (tentative)
I’m increasingly thinking that the 96% read statistic isn’t a bug, it’s a feature. The value proposition of blockchain isn’t fast reads or writes—it’s verifiable, uncensorable writes with abundant, permissionless reads.
Traditional databases optimize for both read and write performance because both are expensive. Blockchains deliberately make writes expensive (consensus overhead) while making reads cheap (anyone can run a node).
The question is: are we building infrastructure that acknowledges this reality? Or are we still trying to make blockchains into better databases?
What do you think? Am I overthinking this? Is the 96% read pattern actually revealing something fundamental about blockchain architecture, or is it just a technical detail we need to optimize?
Data sources: Syndica’s Sig announcement, Solana Compass, personal analysis of RPC traffic patterns