Ethereum's rollups need data availability (DA) to remain permissionless. Many production rollups today ensure DA by storing transactions directly in Ethereum calldata. Doing so inherits the security of Ethereum. This, however, comes with the drawback of high costs. Rollups that use Ethereum for DA must compete with all other network activity for the same available space per block, thus driving up L2 transaction costs.
Proto-danksharding (described in EIP-4844) is a soon-to-be upgrade that increases the data available in each Ethereum block from ~100KB to 1MB. More space means cheaper storage. Cheaper storage means cheaper DA. Cheaper DA means cheaper txs for users.
How does proto-danksharding make this increase in block storage economically viable for nodes? Rather than storing data blobs in the execution layer, proto-danksharding has raw data stored temporarily in the consensus layer, with only the commitments stored in the execution layer. Raw blob data in the consensus layer can then be deleted after a small time window to avoid blowing up disk space requirements for nodes.
These blob commitments are KZG vector commitments. Higher level reasoning about these commitments (eg. "transactions
The first step in writing these circuits is proving KZG multi-opens. In particular, we need Halo2 circuits to prove that a given KZG commitment
KZG is instantiated over a pairing friendly curve
Blobs are treated as vectors
It is
We'd like to prove that a commitment
Our prover implementation is included in the demo below. We do not cover any of the intuition behind KZG's correctness and evaluation binding. For more information, we recommend Dankrad's explainer.
We use the monomial basis from the trusted setup even though EIP4844 specifies that
Using the monomial basis for evaluating
We implement the verifier in a KZGChip
, which comes with a PolyChip
to do polynomial evaluations. They're both merged into Axiom's halo2 fork.
The verifier checks whether a given commitment
Interpolating
We need to take special care to ensure that these polynomials are properly constrained since our interpolation and polynomial multiplication is unconstrained. Think "single arrow" in circom.
We ran benchmarks on an M1 Max with 10 cores and 64GB of RAM. The table below shows metrics for blob length
Poly Eval | MSM | Pairing Check | |
---|---|---|---|
Num advice cells | 5M | 25M | 9M |
Circuit compilation | 10s | 59s | 23s |
Proving time | 11s | 61s | 28s |
Verification time | 1ms | 1ms | 1ms |
Notice that the pairing check is a heavy component in our circuit. There's a well-established technique for optimizing it that we saw as worthwhile. A pairing has two primary steps- a miller loop and a final exponentiation. If we define
Since we know 1) how to invert either side by negating any of the inputs to
This rearrangement only requires one final exponentiation, thus shaving off the number of required context cells by 30%. We implemented this efficient pairing check in PR#65 and incorporated it into our project.
The KZGChip
is a general-purpose primitive that can be used for any construction built on top of KZG. Aside from our immediate use case with proto-danksharding, we're also excited for future work to use the chip for Halo2 composition with KZG-based proof systems (including itself).
You can begin by cloning our repository.
git clone git@github.com:lyronctk/kzg-blob.git
You can run our demo with the following command. It runs through what a sample workflow would look like for a rollup proving transaction inclusion in a blob commitment.
cargo run --example demo
We first randomly initializes 16 random transactions.
[
DemoTx { from: "0xJDp5PvyZPVfOt3YDqabZwi4mYA9LVC4O8Q1NUUaj", to: "0xKUGZ8LOdC2HL6Q6dDNtMb0Arwv9yRk9CL38RrZ4N", gas_limit: 12173, max_fee_per_gas: 126, max_priority_fee_per_gas: 14, nonce: 22, value: 123008058146 },
...,
DemoTx { from: "0x7selZXOPMPraS5guZACuJW4YexSpzVZi6bZxhsES", to: "0xXd225ynUdqKTlmZlFtfVbpVWhBQeoSX1hjTeT4Vj", gas_limit: 16162, max_fee_per_gas: 423, max_priority_fee_per_gas: 14, nonce: 25, value: 62855985777 }
]
The transactions are then packed into elements in
[
0x00006000009dc030000000000000002a000000000000002a00006000009dc000,
...,
0x00000000000001a70000000000003f22000000000000002a000000000000002a
]
We then commit to the blob data, producing a
(0x196decccff1936530fe5b88d5e529b9460f056a515342b33bed55a88c740e16c, 0x0c93b3a9f6b70434c0fe06f8f758f6c0eba61583342606e89e51be0b765e85dc)
The commitment
One limitation with the current chip is that it's specific to the BN254 curve. An immediate next step following this project is adding support for other pairing-friendly curves, most notably BLS12-381 since this is what's officially used for proto-danksharding.
Upcoming use cases will also require concurrent reads from multiple blob commitments. Concretely, this refers to opening a set of commitments