This post outlines optimistic rollup: a construction which enables autonomous smart contracts on layer 2 (L2) using the OVM. The construction borrows heavily from both plasma and zkRollup designs, and builds on shadow chains as described by Vitalik. This construction resembles plasma but trades off some scalability to enable running fully general (eg. Solidity) smart contracts in layer 2, secured by layer 1. Scalability is proportional to the bandwidth of data availability oracles which include Eth1, Eth2, or even Bitcoin Cash or ETC — providing a near term scalable EVM-like chain in layer 2.
Let’s start with some intuitions for how optimistic rollup works end to end on mainnet Ethereum, then dive in deep.
The following is a chronicle of the life of an optimistic rollup smart contract… named Fred:
- Developer writes a Solidity contract named Fred. Hello Fred!
- Developer sends transaction off-chain to a bonded aggregator (a layer 2 block producer) which deploys the contract.
— Anyone with a bond may become an aggregator.
— There are multiple aggregators on the same chain.
— Fees are paid however the aggregator wants (account abstraction / meta transactions).
— Developer gets an instant guarantee that the transaction will be included or else the aggregator loses their bond.
- Aggregator locally applies the transaction & computes the new state root.
- Aggregator submits an Ethereum transaction (paying gas) which contains the transaction & state root (an optimistic rollup block).
- If anyone downloads the block & finds that it is invalid, they may prove the invalidity with
verify_state_transition(prev_state, block, witness)which:
— Slashes the malicious aggregator & any aggregator who built on top of the invalid block.
— Rewards the prover with a portion of the aggregator’s bond.
- Fred the smart contract is safe, happy & secure knowing her deployment transaction is now a part of every valid future optimistic rollup state. Plus Fred can be sent mainnet ERC20’s deposited into L2! Yay!
That’s it! The behavior of users & smart contracts should be very similar to what we see today on Ethereum mainnet, except, it scales! Now let’s explore how this whole thing is possible.
Optimistic Rollup In Depth
To begin let’s define what it means to create a permissionless smart contract platform like Ethereum. There are three properties we must satisfy to build one of these lovely state machines:
- Available head state — Any relevant party can download the current head state.
- Valid head state — The head state is valid (eg. no invalid state transitions).
- Live head state — Any interested party can submit transactions which transition the head state.
You’ll notice that Ethereum layer 1 satisfies these three properties because we believe 1) miners do not mine on unavailable blocks, 2) miners do not mine on invalid blocks*; and 3) not all miners will censor transactions. However, it doesn’t currently scale.
On the other hand, under some similar security assumptions, optimistic rollup can provide all three guarantees at scale. To understand the construction & security assumptions we’ll go over each property we’d like to ensure individually.
#1: Available head state
Optimistic rollup uses classic rollup techniques (outlined here) to ensure data availability of the current state. The technique is simple — block producers (called aggregators) pass all blocks which include transactions & state roots through calldata (ie. the input to an Ethereum function) on Ethereum mainnet. The calldata block is then merklized & a single 32 byte state root is stored. For reference, calldata is 2,000 gas per 32 bytes while storage is 20,000 gas. Additionally, the gas cost of calldata will be reduced by almost 5x in the Istanbul hard fork.
Notably, we can use data availability oracles other than the Ethereum mainnet including Bitcoin Cash and Eth2. With Eth2 phase 1 all shards can serve as data availability oracles, scaling TPS linearly in the number of shards. This is enough throughput that we will hit other scalability bottlenecks before we run out of available data, for example state computation.
Here we assume honest majority on Ethereum mainnet. In addition, if we use Eth2, ETC, or Bitcoin Cash, we similarly inherit their honest majority assumptions.
Under these assumptions, using a trusted availability oracle to publish all transactions we can ensure that anyone can compute the current head state, satisfying property #1.
#2: Valid head state
The next property we need to ensure is a valid head state. In zkRollup we use zero-knowledge proofs to ensure validity. While this is a great solution in the long term, for now it is not possible to create efficient zkProofs for arbitrary state transitions. However, there’s still hope for a general purpose EVM-style state machine! We can use a cryptoeconomic validity game similar to plasma / truebit.
Cryptoeconomic Validity Game
At a high level the block submission & validity game is as follows:
- Aggregators post a security deposit to start producing blocks.
- Each block contains
[access_list, transactions, post_state_root].
- All blocks are committed to a
ROLLUP_CHAINcontract by a bonded aggregator on a first come first serve basis (or round robin if desired).
- Anyone may prove a block invalid, winning a portion of the aggregator’s security deposit.
To prove a block invalid you must prove one of the three following properties:
1. INVALID_BLOCK: The committed block is *invalid*.
This is calculated with `is_valid_transition(prev_state, block, witness) => boolean`
2. SKIPPED_VALID_BLOCK: The committed block "skipped" a valid block.
3. INVALID_PARENT: The committed block's parent is invalid.
These three state transition validity conditions can be visualized as:
There are a few interesting properties that fall out of this state validity game:
- Pluggable validity checkers: We can define different validity checkers for
is_valid_transition(…)allowing us to use different VMs to run smart contracts including EVM and WASM.
- Only one valid chain: Blocks are submitted to Ethereum which gives us a total ordering of transactions & blocks. This enables us to deterministically decide the “head” block, and thereby require aggregators to prune invalid blocks before submitting a new block.
- Sharded validation: This validity game can be played out at an individual UTXO basis. Instead of invalidating full blocks, we partially invalidate them — similar to Plasma Cash. Note that this does not require proving all invalid transitions up front for a single block. Partial block invalidation means we can validate only UTXOs for contracts we care about to secure our state. To learn more about how UTXOs enable parallelism check out this Cryptoeconomics.study video!
A Note on Watchtowers
One challenge to adoption of L2 has been the added complexity of watchtowers. Users contracting watchtowers adds yet another entity to manage to an already complex system. Thankfully, watchtowers are naturally incentivized by the optimistic rollup cryptoeconomic validity game! All data is available & therefore anyone running a full node stands to gain the security bond of all aggregators who build on their invalid chain. This risk incentivizes aggregators to be watchtowers, validating the chain they are building on — mitigating the verifiers dilemma.
A Note on Plasma
Many plasma constructions also rely on cryptoeconomic validity games; however, in plasma autonomous smart contract state enforcement is impossible without zkProofs or a fisherman’s game during a data withholding attack (the data availability problem). Thankfully rollup gets around the data availability problem by posting the minimal information needed to compute state transitions on-chain. Still, plasma is critical if we want to scale up to transactions per second in the hundreds of thousands (and more) — a long term necessity but not a requirement in the medium term for many smart contracts.
- This cryptoeconomic validity game requires a single rational verifier assumption. We can say it is a “rational” verifier as opposed to “honest” because they are economically incentivized by the forfeiture of the aggregator bond to prove invalidity.
- Additionally we assume the mainnet is live, meaning it is not censoring all incoming transactions attempting to prove invalidity. Note that the aggregator unbonding period is in some sense the liveness assumption on the mainnet (eg. if we require a 1 month unbonding period, then invalidity must be proven within that month to forfeit that bond).
Under these assumptions, all invalid blocks / state transitions will be discarded leaving us with a *single* valid head state, satisfying property #2.
#3: Live Head State
The final property we must satisfy is liveness, often known as censorship resistance. The key insights which ensures this are:
- Anyone with a bond size above
MINIMUM_BOND_SIZEmay become an aggregator for the same rollup chain.
- Because honest aggregators may prune invalid blocks, the chain does not halt in the event of an invalid block.
With these two properties we’ve already got liveness! Honest aggregators may always submit blocks which fork around invalid blocks & so even if there’s just one non-censoring aggregator your transaction will eventually get through — similar to mainnet.
A Note on Liveness vs Instant Confirmations
One property we really want is instant confirmations. This way we can give users sub-second feedback that their transaction will be processed. We can achieve this by designating short-lived aggregator monopolies on blocks. The downside is that it trades off censorship resistance because now a single party can censor for some period of time. Would love to hear about any research on this tradeoff!
With two security assumptions we get liveness:
- There exists a non-censoring aggregator.
- Mainnet Ethereum is not censoring.
Under these assumptions, the optimistic rollup chain will be able to progress & mutate the head state based on any valid user transactions, satisfying property #3.
Now all three properties are satisfied & we’ve got a permissionless smart contract platform in Ethereum L2!
The following estimates are purely based on data availability. In practice other bottlenecks could be hit, one being state calculation. However, this does provide a useful upper bound.
ERC20 Transfers with ETH1 Data availability
Calculations are based on this little call-data calculation python script.
Note that these ERC20 transfers are calldata optimized. Additionally note that the nice thing about Optimistic Rollup is we aren’t limited to ERC20 transfers!
~100 TPS without EIP 2028
~450 TPS with EIP 2028 (coming in October 2019)
BLS Signature / SNARK-ed Signatures
~400 TPS without EIP 2028
~2000 TPS with EIP 2028 (coming in October 2019)
With external availability oracles (eg. ETH2, Bitcoin Cash)
~linear in relation to the amount of throughput the availability oracle can handle.
That’s a lot more than 2000 TPS!
Optimistic Rollup vs Plasma
Optimistic Rollup shares much in common with Plasma. Both use aggregators to commit to blocks on mainnet with a cryptoeconomic validity game ensuring safety. The sole divergence is whether or not we have an availability receipt ensuring block availability.
The similarities between the two solutions allows for lots of shared infrastructure & code between the two constructions. In a mature layer 2 ecosystem it’s likely that we will see rollup, plasma, and state channels all working together in the same client (a smart wallet). Oh, have I mentioned the OVM? 😁
Yay Optimistic Rollup 🎉
Optimistic Rollup occupies a nice niche in the space of layer 2 constructions. It trades off some scalability for general purpose smart contracts, simplicity, & security. Plus being able to run secure smart contracts means that it can even be used to adjudicate other layer 2 solutions like plasma and state channels!
Call it “the layer 1 of layer 2s.”
Anyway, enough research — time to implement a robust, comprehensive, and user friendly Ethereum layer 2! 😍
Special thanks to Vitalik Buterin for working through these ideas with me and for coming up with much of this.
Additionally, thank you Ben Jones for much of this and Jinglan Wang, Kevin Ho & Jesse Walden for edits.
Update: Shout out to the great and related work by John Adler in his article on Merged Consensus which can be compared to optimistic rollups — Check it out here! Plus this proposal to improve Eth1’s ability to act as a data availability oracle — more tps!