A basic REST proxy server to interact with the EigenDA network:
- POST routes: submit a payload (rollup txs, state-diffs, or anything really) that will be encoded into an EigenDA blob and submitted to the EigenDA disperser to make available for 2 weeks. A DA certificate of availability will be returned, which can be used to validate the availability and query the payload back.
- GET routes: submit a DA Certificate to retrieve its respective blob from the EigenDA network, which will be decoded, validated, and returned as a response.
V1 Integration Guide | V2 Integration Spec | EigenDA Repo
This service wraps the high-level EigenDA client, exposing endpoints for interacting with the EigenDA disperser in conformance to the OP Alt-DA server spec, and adding disperser verification logic. This simplifies integrating EigenDA into various rollup frameworks by minimizing the footprint of changes needed within their respective services.
Features:
- Exposes an API for dispersing blobs to EigenDA and retrieving blobs from EigenDA via the EigenDA disperser
- Handles BN254 field element encoding/decoding
- Performs KZG verification during retrieval to ensure that data returned from the EigenDA disperser is correct.
- Performs KZG verification during dispersal to ensure that DA certificates returned from the EigenDA disperser have correct KZG commitments.
- Performs DA certificate verification during dispersal to ensure that DA certificates have been properly bridged to Ethereum by the disperser.
- Performs DA certificate verification during retrieval to ensure that data represented by bad DA certificates do not become part of the canonical chain.
- Compatibility with Optimism's alt-da commitment type with eigenda backend.
- Compatibility with Optimism's keccak-256 commitment type with S3 storage.
For testing purposes, proxy provides a fully in-memory backend that mocks a real backing EigenDA network. Here's how to start the proxy in this mode and interact with it:
# Start the proxy with memstore backend enabled
$ docker run --rm -p 3100:3100 ghcr.io/layr-labs/eigenda-proxy:latest --memstore.enabled --port 3100
# In another terminal... submit a payload save the returned cert in hex format
$ CERT_HEX=$(curl -X POST -d my-eigenda-payload "http://127.0.0.1:3100/put?commitment_mode=standard" | xxd -p | tr -d ' \n')
# Finally retrieve the payload using the cert
$ curl "http://127.0.0.1:3100/get/$CERT_HEX?commitment_mode=standard"
We build and publish containers on every release to ghcr.io/layr-labs/eigenda-proxy. You can also build from source by running make
.
The source of truth for the routes is defined by our gorilla mux router in ./server/routing.go. We offer two sets of POST/GET routes.
TODO
These routes are specific to optimism rollups and follow op's altda-server spec. Do note that the op spec is wrong in that their altda client and server implementation actually return <commitment_bytes> on the POST routes, not <hex_encoded_commitment>. The below routes are correct.
Request:
POST /put/<hex_encoded_commitment>
Content-Type: application/octet-stream
Body: <preimage_bytes>
Response:
200 OK
Where the <hex_encoded_commitment> for keccak commitments is the keccak256 hash of the preimage_bytes, prepended with 0x00
.
Request:
POST /put
Content-Type: application/octet-stream
Body: <preimage_bytes>
Response:
200 OK
Content-Type: application/octet-stream
Body: <commitment_bytes>
Where the <commitment_bytes> is the serialized versioned DA certificate of the blob.
Both altda commitment forms above share the same GET route to retrieve the preimage_bytes.
Request:
GET /get/<hex_encoded_commitment>
Response:
200 OK
Content-Type: application/octet-stream
Body: <preimage_bytes>
We also provide network-specific example env configuration files in .env.example.holesky
and .env.example.mainnet
as a place to get started:
- Copy example env file:
cp .env.example.holesky .env
- Update env file, setting
EIGENDA_PROXY_SIGNER_PRIVATE_KEY_HEX
. On mainnet you will also need to setEIGENDA_PROXY_ETH_RPC
. - Pass into binary:
ENV_PATH=.env ./bin/eigenda-proxy --addr 127.0.0.1 --port 3100
## Setup new keypair for EigenDA authentication
$ cast wallet new --json > keypair.json
## Extract keypair private key and remove 0x prefix
$ PRIVATE_KEY=$(jq -r '.[0].private_key' keypair.json | tail -c +3)
## If running V1 against testnet or mainnet, register the keypair ETH address and wait for approval: https://forms.gle/niMzQqj1JEzqHEny9
## Run EigenDA Proxy with EigenDA V1 backend
$ ./bin/eigenda-proxy \
--port 3100 \
--eigenda.disperser-rpc disperser-holesky.eigenda.xyz:443 \
--eigenda.signer-private-key-hex $PRIVATE_KEY \
--eigenda.eth-rpc https://ethereum-holesky-rpc.publicnode.com \
--eigenda.svc-manager-addr 0xD4A7E1Bd8015057293f0D0A557088c286942e84b
Below is a list of the main high-level features offered for configuring the eigenda-proxy. These features are controlled via flags and/or env vars. To view the extensive list of available flags/env-vars to configure a given version of eigenda-proxy, run eigenda-proxy --help
.
In order for the EigenDA Proxy to avoid a trust assumption on the EigenDA disperser, the proxy offers a DA cert verification feature which ensures that:
- The DA cert's batch hash can be computed locally and matches the one persisted on-chain in the
ServiceManager
contract - The DA cert's blob inclusion proof can be successfully verified against the blob-batch merkle root
- The DA cert's quorum params are adequately defined and expressed when compared to their on-chain counterparts
- The DA cert's quorum ids map to valid quorums
To target this feature, use the CLI flags --eigenda-svc-manager-addr
, --eigenda-eth-rpc
.
An optional --eigenda.confirmation-depth
flag can be provided to specify a number of ETH block confirmations to wait for the confirmBatch to have landed onchain before returning the cert to the batcher after having dispersed a blob in the put route. The flag value can either be the string 'finalized' or a number:
finalized
: Wait for the confirmBatch transaction to be finalized on-chain before returning the cert to the batcher
0
: Verify the cert immediately upon blob confirmation and return the cert
N where 0<N<64
: Wait N
blocks before returning the cert to the batcher
The default value is 8. Using 0 is dangerous: see troubleshooting the batch-hash-mismatch error.
An ephemeral memory store backend can be used for faster feedback testing when testing rollup integrations. To target this feature, use the CLI flags --memstore.enabled
, --memstore.expiration
.
An optional --routing.concurrent-write-routines
flag can be provided to enable asynchronous processing for secondary writes - allowing for more efficient dispersals in the presence of a hefty secondary routing layer. This flag specifies the number of write routines spun-up with supported thread counts in range [1, 100)
.
An optional storage fallback CLI flag --routing.fallback-targets
can be leveraged to ensure resiliency when reading. When enabled, a blob is persisted to a fallback target after being successfully dispersed. Fallback targets use the keccak256 hash of the existing EigenDA commitment as their key, for succinctness. In the event that blobs cannot be read from EigenDA, they will then be retrieved in linear order from the provided fallback targets.
An optional storage caching CLI flag --routing.cache-targets
can be leveraged to ensure less redundancy and more optimal reading. When enabled, a blob is persisted to each cache target after being successfully dispersed using the keccak256 hash of the existing EigenDA commitment for the fallback target key. This ensure second order keys are succinct. Upon a blob retrieval request, the cached targets are first referenced to read the blob data before referring to EigenDA.
In the event that the EigenDA disperser or network is down, the proxy will return a 503 (Service Unavailable) status code as a response to POST requests, which rollup batchers can use to failover and start submitting blobs to the L1 chain instead. For more info, see our failover designs for op-stack and for arbitrum.
This behavior is turned on by default, but configurable via the --eigenda.confirmation-timeout
flag (set to 15 mins by default currently). If a blob is not confirmed within this time, the proxy will return a 503 status code. This should be set long enough to accomodate for the disperser's batching interval (typically 10 minutes), signature gathering, and onchain submission.
In order to disperse to the EigenDA V1 network in production, or at high throughput on testnet, please register your authentication ethereum address through this form. Your EigenDA authentication keypair address should not be associated with any funds anywhere. For EigenDA V2, please see our payments doc.
A normal (non-archival) Ethereum node is sufficient for running the proxy with cert verification turned on. This is because all parameters that are read from the chain are either:
- immutable (eg: securityThresholds), or
- are upgradeable but have all the historical versions available in contract storage (eg: versioninedBlobParams)
In order to compute (and in our current implementation also verify) KZG commitments, G1 SRS points of size equivalent to the blob size are needed. The points must be loaded into the binary by using the --eigenda.g1-path flag. A 32MiB G1 SRS file is available under ./resources/g1.point. This file is also copied inside our distributed docker images, at <WORKDIR>/resources/g1.point. The --eigenda.g1-path
flag's default value is the relative path resources/g1.point
, which will work when running the binary from the repo's root directory, as well as inside the container.
The following specs are recommended for running on a single production server:
- 1-2 cores CPU
- 4 GB RAM
To the see list of available metrics, run ./bin/eigenda-proxy doc metrics
To quickly set up monitoring dashboard, add eigenda-proxy metrics endpoint to a reachable prometheus server config as a scrape target, add prometheus datasource to Grafana to, and import the existing Grafana dashboard JSON file
Warning: the below diagrams describe EigenDA V2 interactions. EigenDA V1 is very similar, but has slight discrepancies.
The proxy fundamentally acts as a bridge between the rollup nodes and the EigenDA network. The following sequence diagram illustrates the lifecycle of a rollup payload (compressed batch of txs or state transitions), as it gets transformed to an EigenDA blob and dispersed to the network. The received EigenDA cert is then published to the rollup batcher-inbox, to be retrieved by rollup validators, and used to retrieve and validate the corresponding blob, which can then be decoded into the original payload and used by the rollup's stack.
The rollup payload is submitted via a POST request to the proxy. Proxy encodes the payload into a blob and submits it to the EigenDA disperser. After the DACertificate is available via the GetBlobStatus endpoint, it is encoded using the requested commitment schema and sent back to the rollup sequencer. The sequencer then submits the commitment to the rollup's batcher inbox.
Validator nodes proceed with the exact reverse process as that used by the sequencer in the posting blobs section. The rollup validator submits a GET request to the proxy with the DACert in the body. The proxy validates the cert, fetches the corresponding blob from EigenDA, validates it, decodes it back into the rollup payload, and returns it the rollup node.
Warning: the name
commitment
here refers to the piece of data sent to the rollup's batcher inbox (see op spec's description), not to blobs' KZG commitment. The Rollup commitment consists of a few-byte header (described below) followed by aDACertificate
, which contains all the information necessary to retrieve and validate an EigenDA blob. TheDACertificate
itself contains the KZG commitment to the blob.
Currently, there are two commitment modes supported with unique encoding schemas for each. The version byte
is shared for all modes and denotes which version of the EigenDA DACertificate
is being used/requested. The following versions are currently supported:
0x00
: EigenDA V1 certificate type (i.e, dispersal blob info struct with verification against service manager)0x01
: EigenDA V2 certificate type
For alt-da
Optimism rollups using EigenDA, the following commitment schemas are supported by our proxy:
commitment_type (byte) | da_layer_byte | version_byte | payload |
---|---|---|---|
0x00 | keccak_commitment | ||
0x01 | 0x00 | 0x00 | eigenda_cert_v1 |
0x01 | 0x00 | 0x01 | eigenda_cert_v2 |
keccak256
(commitment_type 0x00) uses an S3 storage backend with where a simple keccak commitment of the blob
is used as the key. For generic
commitments, we only support da_layer_byte
0x00 which represents EigenDA.
For standard clients (i.e, clients/standard_client/client.go
) communicating with proxy (e.g, arbitrum nitro), the following commitment schema is supported:
version_byte | payload |
---|---|
0x00 | eigenda_cert_v1 |
0x01 | eigenda_cert_v2 |
eigenda_cert_v0
is an RLP-encoded EigenDA V1 certificate. eigenda_cert_v1
works similarly.
Browse our Makefile for a list of available commands such as make
for building the binary and make docker-build
to build the docker image.
Unit tests can be ran via invoking make test-unit
.
Integration tests against op framework can be ran via make test-e2e-local
. These tests use the op-e2e framework for asserting correct interaction behaviors with batch submission and state derivation.
These tests also assert E2E client <-> server interactions using simple/op clients.
A holesky integration test can be ran using make test-e2e-holesky
to assert proper dispersal/retrieval against a public network. Please note that EigenDA Holesky network which is subject to rate-limiting and slow confirmation times (i.e, >10 minutes per blob confirmation). Please advise EigenDA's inabox if you'd like to spin-up a local DA network for faster iteration testing.
This E2E test will fuzz the proxy client server integration and op client keccak256 with malformed inputs. This is never meant to be fuzzed with EigenDA. Run with make test-e2e-fuzz
.