Danksharding Explained: Exclusive Guide vs Proto-Danksharding
Ethereum’s roadmap revolves around one core promise: scale without sacrificing security. Danksharding and proto-danksharding are key milestones on that path....
In this article

Why scaling Ethereum hinges on data availability
Rollups like Optimism, Arbitrum, and zkSync batch thousands of transactions and publish proof and data back to Ethereum. The cost bottleneck isn’t computation; it’s posting the data so anyone can verify the rollup’s state. Lowering this “data availability” cost has an immediate impact on user fees, from swaps to NFT mints.
Enter data blobs—large, temporary chunks of data that rollups can publish cheaply without clogging the main chain. This idea sits at the heart of proto-danksharding and danksharding.
Proto-danksharding (EIP-4844): the practical first step
Proto-danksharding, introduced by EIP-4844, adds a new transaction type that carries “blobs.” Blobs are not accessible to the EVM directly and expire after a short retention window. They’re verified with cryptographic commitments (KZG commitments) but don’t impose long-term storage costs on nodes. The immediate effect: rollups get a cheaper lane for their data.
Think of a rollup posting a 100 kB batch. Before EIP-4844, it had to cram that into calldata—expensive and permanent. With blobs, the same batch rides a discounted lane designed precisely for short-lived data.
Danksharding: the full design
Danksharding is the endgame architecture. It scales blob capacity by orders of magnitude using a single, unified market for block space and data availability. Instead of sharding execution (many mini-chains), Ethereum keeps a single execution layer but shards data availability. Validators sample small slices of the blob data to check that the full dataset is available—this is data availability sampling (DAS).
The core pieces include a single proposer, a builder market, large numbers of blobs per block, commitments for each blob, and lightweight sampling by validators. Together, these changes let Ethereum carry massive rollup throughput without each node downloading every byte.
Key differences at a glance
The table below summarizes how proto-danksharding compares with full danksharding across design goals and mechanics.
| Aspect | Proto-danksharding (EIP-4844) | Danksharding (Full) |
|---|---|---|
| Purpose | Immediate fee relief via blobs | Massive, sustained scale for rollups |
| Data mechanism | Limited number of blobs per block | Much higher blob throughput with sampling |
| Verification | KZG commitments; no DAS | KZG commitments plus data availability sampling |
| Node load | Manageable; full download of included blobs | Lower per-node bandwidth via sampling |
| Execution model | Single chain; execution unchanged | Single chain; data availability “sharded” |
| Maturity | Live (post-4844) | Planned; requires protocol upgrades and DAS |
Both phases preserve Ethereum’s single execution environment. The leap from proto- to full danksharding is about scale and verification strategy, not changing how smart contracts run.
How blobs work in practice
Each blob is a large data chunk attached to a transaction. It’s committed with a polynomial commitment (KZG), so nodes can verify integrity without inspecting every byte. Blobs are pruned after a set period, which keeps state bloat in check. Rollups extract these blobs to reconstruct batches, then prove correctness on-chain.
A quick scenario: a DEX on a rollup executes 40,000 swaps in ten seconds. The rollup batches them, posts one blob with the batch data, and a validity or fraud-proof path anchors it. Users pay cents instead of dollars because the data lane is priced for throughput.
Why data availability sampling matters
Full danksharding leans on data availability sampling so validators don’t need to download every blob. Each validator randomly samples small pieces from many blobs. If the network as a whole can retrieve enough samples, it’s statistically safe to assume the full data is available. If a proposer tries to hide data, enough samples fail and the block is rejected.
This approach unlocks high blob counts without melting bandwidth. It’s what turns “cheaper” into “truly scalable.”
Benefits for users and builders
The most visible win is lower fees on rollups. That cascades into better UX: cheaper swaps, faster settlements, and more predictable costs. Builders get room to design richer applications—think on-chain games streaming frequent state updates or social protocols with high write volumes—without fee shock.
On the infrastructure side, node operators avoid long-term storage burdens, since blobs expire. The network stays lean while still carrying a flood of short-lived, verifiable data.
Risks and open questions
Every scaling upgrade shifts incentives and attack surfaces. Two areas draw scrutiny: proposer-builder separation (PBS) markets and blob pricing. Concentration among block builders could tilt power dynamics if not checked by protocol and competition. Meanwhile, blob fees need adaptive tuning so demand spikes don’t starve execution or vice versa.
There’s also the choreography of upgrades. Introducing data availability sampling, strengthening light clients, and refining fee markets must land in the right order to avoid regressions in security or liveness.
What changes for developers
Smart contracts don’t read blobs directly. If you’re building a rollup, you’ll integrate blob submission in your sequencer and batcher. If you’re building on a rollup, you just enjoy lower fees. Tooling upgrades focus on:
- Sequencer pipelines that package batches into blobs
- Indexers that fetch and cache blob data before expiry
- Monitoring for blob inclusion, fees, and data integrity
A wallet UX tweak helps too: show blob-related fee components separately from gas so users see where savings come from.
From proto to full: the expected path
The journey proceeds in measured steps. Proto-danksharding established the economics and plumbing for blobs. Full danksharding layers in data availability sampling and bigger blob budgets.
- Harden the blob market: watch uptake, fee dynamics, and relayer behavior under EIP-4844.
- Roll out sampling: enable validators and light clients to perform DAS at scale.
- Increase capacity: raise blob counts per block as measurements confirm network health.
Each step is observable. Client teams and researchers track bandwidth, propagation delays, and inclusion risk before dialing up throughput.
How this affects L2 competition
Lower data availability costs compress fee differences across optimistic and zk-rollups. The battleground shifts to proof systems, UX, and ecosystem incentives. A zk-rollup posting frequent validity proofs might pass more savings to users once blob capacity scales further. An optimistic rollup could lean on higher throughput and shared liquidity.
Expect more apps to default to L2, while mainnet focuses on settlement, security, and value storage.
Bottom line for readers
Proto-danksharding is the live, tangible upgrade that introduced blobs and cut rollup costs. Danksharding is the full vision that adds data availability sampling and unlocks far higher throughput. Both preserve Ethereum’s single execution model and push most activity to rollups, where users get speed and low fees without abandoning mainnet security.
If you’re a user, watch L2 fees—they’ll trend down as capacity scales. If you’re a builder, design for blobs now and prepare for a world where data is cheap, verifiable, and short-lived by design.

