Documentation
Narwhal: Achieving Scalability and Throughput

Narwhal: Achieving Scalability and Throughput

The key insight from the Narwhal & Tusk (opens in a new tab) paper is that scalable, high-throughput consensus systems should separate data dissemination from the ordering mechanism for reliability.

Narwhal employs a highly scalable and efficient Directed Acyclic Graph (DAG) structure, which, in conjunction with Tusk/Bullshark, processes over 100,000 transactions per second (tps) in a geo-replicated environment while maintaining latency below three seconds. In contrast, Hotstuff handles fewer than 2,000 tps under similar conditions.

Narwhal's symmetric data dissemination among validators ensures fault resilience, as the system's performance is only impacted when faulty validators fail to disseminate data.

Each validator in Narwhal consists of multiple workers and a primary node. Workers continuously exchange data batches and forward batch digests to their primaries. The primaries construct a round-based DAG using these digests. Crucially, data dissemination by workers occurs at network speed, independent of the DAG construction pace by the primaries. Validators use a broadcast protocol (opens in a new tab) ensuring reliable communication with a linear message count on the critical path.

DAG

One of the key advantages of DAG-based consensus is that it introduces zero additional communication overhead. Each validator independently examines its local view of the DAG and fully orders all vertices without sending any extra messages. The structure of the DAG itself is interpreted as the consensus protocol, where a vertex represents a proposal and an edge signifies a vote.

A significant challenge in this context arises from the asynchronous nature of the network, meaning that different validators may see slightly different DAGs at any given time. The primary difficulty lies in ensuring that all validators ultimately agree on the same total order of transactions.

Narwhal

During each round (refer to Figure 2 in the Narwhal & Tusk paper for a visual guide):

  1. Every validator constructs and dispatches a message to all other validators containing the metadata for the DAG vertex, which includes batch digests and n-f references to vertices from the previous round.

  2. Upon receiving this message, a validator responds with a signature if:

  • Its workers have stored the data corresponding to the digests in the vertex (ensuring data availability).

  • It has not yet responded to this validator in the current round (ensuring non-equivocation).

  1. The sender aggregates these signatures to form a quorum certificate from n-f responses and includes this certificate in its vertex for the round.

  2. A validator progresses to the next round once it has received n-f vertices accompanied by valid certificates.

Quorum certificates serve three main purposes:

  1. Non-equivocation: Validators sign only one vertex per round, ensuring consistency in the DAG. Byzantine validators cannot generate conflicting quorum certificates for the same round.

  2. Data availability: Validators sign only if they have stored the corresponding data, guaranteeing that data can be retrieved later by any validator.

  3. History availability: Certificates of blocks from the previous round ensure the availability of the entire causal history. A new validator can join by learning the certificates from the prior round, simplifying data management.

By decoupling data dissemination from metadata ordering, Narwhal ensures network-speed throughput, regardless of the DAG construction or consensus latency. Data dissemination happens at network speed, while the DAG references all disseminated data. Once a DAG vertex is committed, its entire causal history (including most disseminated data) is ordered.

2024 Scalaris