← Aptos Intelligence Deep Dive

Prefix Consensus — How Aptos Is Building Leaderless, Censorship-Resistant BFT

By aptos-labs Apr 16, 2026 Feature Progress Importance 10/10 Source
consensuscensorship-resistanceprefix-consensusleaderless

Prefix Consensus: How Aptos Is Building Censorship-Resistant, Leaderless BFT

In February 2026, three Aptos Labs researchers — Zhuolun Xiang (Daniel), Andrei Tonkikh, and Alexander Spiegelman — published "Prefix Consensus For Censorship Resistant BFT" (arXiv:2602.02892). This paper introduces a new BFT primitive that fundamentally changes how blockchain consensus can work: instead of requiring all honest validators to agree on a single value, it allows them to agree on a prefix of ordered proposals. This is not incremental. It is the theoretical foundation for Aptos's next-generation consensus layer.

The Primitive: What Is Prefix Consensus?

Classical BFT consensus (like PBFT, HotStuff, Jolteon) solves this problem: n parties each have an input value, and they must all agree on a single output. This is single-value agreement. It requires a leader to drive the protocol, and that leader becomes a single point of failure for both liveness and censorship.

Prefix Consensus solves a different problem. Each of the n parties inputs a vector — for example, a list of transaction hashes. The protocol guarantees that every honest party outputs a pair (v_{\text{low}}, v_{\text{high}}) where:

The critical difference: outputs can differ across parties. Party A might output (v_low_A, v_high_A) and Party B might output (v_low_B, v_high_B). But they are guaranteed to be consistent — one party's v_{\text{low}} is always a prefix of the other's v_{\text{high}}, or vice versa. This relaxation is what makes the protocol faster and leaderless.

Classical Consensus:
  Input:  Party1=A, Party2=B, Party3=C
  Output: Everyone outputs A (or B, or C — but same value)

Prefix Consensus:
  Input:  Party1=[tx1,tx2,tx3], Party2=[tx1,tx2,tx4], Party3=[tx1,tx2,tx3,tx5]
  Output: Party1=(v_low=[tx1,tx2,tx3], v_high=[tx1,tx2,tx3,tx5])
          Party2=(v_low=[tx1,tx2],     v_high=[tx1,tx2,tx4])
          Party3=(v_low=[tx1,tx2,tx3], v_high=[tx1,tx2,tx3,tx5])
  Guarantee: v_low always extends max common prefix [tx1,tx2]

Formally, the paper defines two variants. Weak Prefix Consensus (WPC) guarantees only the low/high consistency above and can be solved asynchronously in 3 rounds. Strong Prefix Consensus (SPC) additionally requires that all honest parties output the same v_{\text{high}} — this is what lets an SMR protocol built on top commit a single canonical sequence. SPC requires partial synchrony (explained further in a later section) but still inherits the 3-round round-complexity in the good case.

The Adversary Model: What Prefix Consensus Actually Assumes

Understanding what Prefix Consensus proves requires being precise about what it assumes. Consensus lower bounds are notorious for being misread — so we spell out every assumption the paper makes.

Compare this to the closest-related protocols:

ProtocolFault BoundNetwork ModelLeaderSuspension Tolerance
PBFT (1999)f < n/3Partial syncRotating viewNone (view-change required)
HotStuff (2019)f < n/3Partial syncRotating viewNone (pacemaker required)
Jolteon (2022)f < n/3Partial syncRotating viewNone
Bullshark (2022)f < n/3Eventual syncAnchorsPer-wave
Prefix Consensus SMRf < n/3Partial syncNoneOne per round after GST

The one-suspension tolerance is the reason Prefix Consensus is described as "leaderless in the strong sense." In every other partial-sync BFT protocol, suspending the current leader halts the protocol until a view change completes — a 2-to-3-round penalty. In Prefix Consensus, suspending any single party in any round still lets the protocol commit within its standard round bound.

The 3-Round Async Protocol

The paper proves that Strong Prefix Consensus can be solved in exactly 3 asynchronous rounds — and that this is a tight lower bound (2 rounds is impossible unless n ≥ 5f+1). The protocol works as follows:

  1. Round 1 — Broadcast: Every party broadcasts its input vector to all other parties
  2. Round 2 — Echo: Upon receiving 2f+1 vectors, each party computes the common prefix of the received vectors and broadcasts an echo message containing this prefix
  3. Round 3 — Vote: Upon receiving 2f+1 echo messages, each party computes the common prefix of the echoes to determine v_{\text{low}}, and the union/extension to determine v_{\text{high}}

This achieves O(n^3) message complexity in the worst case. In the fault-free case, the protocol completes in 3 rounds with immediate output. When integrated into a full SMR (State Machine Replication) protocol, the commit latency is 4 rounds (3 for the prefix consensus instance + 1 for chaining), dropping to 3 rounds in the fault-free fast path.

The Trie-Based Data Structure

A subtle but load-bearing engineering detail of the Prefix Consensus paper is its choice of data structure: the common-prefix computations that sit at the heart of both Round 2 and Round 3 are implemented over tries (radix trees), not sorted lists. This is not an incidental optimization — it is what makes the protocol fit inside a real blockchain round budget.

Consider what Round 2 must do. A party receives 2f+1 input vectors, each of which may be thousands of transaction hashes long. It must compute the longest common prefix of all of them. A naive implementation compares every vector against every other pairwise — that is O(n · k²) where k is the vector length. For n=100, k=10,000, that is 1010 comparisons — far too slow for a round budget of 200ms.

The trie-based approach is radically cheaper. Each input vector is inserted into a single shared trie; edges are labelled with transaction hashes; each leaf is tagged with the set of parties whose vector reached that leaf. The longest-common-prefix computation becomes a single DFS that stops at the first branch where fewer than 2f+1 parties' tags are present. Total cost: O(k) per insertion, O(k) for the final LCP walk, O(n · k) overall. For the same example, 106 operations — four orders of magnitude less.

                   ┌─[tx1]───┐
                   │          │
    root ──[tx0]───┤          ├── [tx2] ── [tx3] ── {P1, P3}   (len=4)
                   │          │
                   │          └── [tx2] ── [tx4] ── {P2}        (len=3)
                   │
                   └─[txA]───── ... (not enough votes here)

  LCP walk: root -> tx0 -> tx1 (3 votes) -> tx2 (3 votes) -> stop
  Common prefix = [tx0, tx1, tx2]  (this becomes v_low)

Two additional properties make the trie choice especially well-suited:

The memory cost is O(n · k) per slot — acceptable given that Aptos validators already buffer quorum-store batches of similar magnitude. Trie reuse across slots is not safe (each slot is an independent instance) but the trie for one slot can be garbage-collected as soon as commit is durable.

From Prefix Consensus to Full BFT SMR

The paper builds a complete leaderless, multi-proposer BFT State Machine Replication protocol on top of Prefix Consensus. Here is how it works:

  1. Every validator broadcasts a proposal every slot. There is no leader election, no designated proposer. All n validators propose simultaneously.
  2. Proposals are deterministically ranked using a cyclic shift. The ranking rotates so that different validators get priority in different slots, but the ranking is known to all parties before the protocol runs.
  3. Run Strong Prefix Consensus on the hashes of these proposals. This determines which proposals are included in the committed block.
  4. Commit the prefix (v_low) immediately. All transactions in v_low are guaranteed to be committed by all honest parties.
  5. Demotion rule: If an honest party's proposal is excluded from the committed prefix — i.e., it was censored — the party that was ranked first (and thus responsible for the exclusion) is demoted in the ranking for future slots.
  6. Convergence guarantee: After at most f censored slots post-GST (Global Stabilization Time), all honest parties' proposals rank first forever. The protocol self-heals.

Walkthrough: One Full Slot of Leaderless SMR

To make the abstract protocol concrete, walk through exactly one slot with n=4, f=1 validators named V0, V1, V2, V3. Assume V3 is Byzantine, the other three are honest, the network is post-GST (timely), and the deterministic ranking for this slot is [V0, V1, V2, V3] (V0 is ranked first).

┌────────────────────────────────────────────────────────────────────┐
│                    SLOT 42 — TIMELINE                              │
├────────────────────────────────────────────────────────────────────┤
│ T+0ms   PROPOSE: every validator broadcasts a batch to all others  │
│                                                                    │
│   V0 -> [a, b, c, d]       (from V0's mempool)                    │
│   V1 -> [a, b, c, e]       (V1 has seen tx e, V0 has seen tx d)   │
│   V2 -> [a, b, c, d, e]    (V2 has seen both tails)               │
│   V3 -> [a, b, z]          (Byzantine — tries to censor d,e)      │
│                                                                    │
│ T+80ms  ROUND 1 complete — each validator has 4 input vectors.     │
│                                                                    │
│ T+80ms  ECHO: each validator ranks proposals by deterministic       │
│         ranking [V0, V1, V2, V3], chains them, and computes the     │
│         longest common prefix of the 2f+1=3 received vectors.       │
│                                                                    │
│   V0 sees {V0,V1,V2,V3 inputs}; chains ranked as                    │
│       [a,b,c,d] || [a,b,c,e] || [a,b,c,d,e] || [a,b,z]              │
│   LCP of first three (highest ranks honest) = [a,b,c]               │
│   V0 broadcasts ECHO(prefix=[a,b,c])                                │
│                                                                    │
│   V1 and V2 compute the same LCP; V3 may echo garbage (ignored).    │
│                                                                    │
│ T+160ms ROUND 2 complete — each validator has 2f+1=3 honest echoes. │
│                                                                    │
│ T+160ms VOTE: each validator computes:                              │
│   v_low  = common prefix of the 3 received echoes = [a,b,c]         │
│   v_high = [a,b,c,d,e]  (union/extension from the input trie)       │
│                                                                    │
│ T+240ms COMMIT: v_low = [a,b,c] is committed. The "high" range      │
│         [d,e] waits one more slot for a deterministic tie-break.    │
│                                                                    │
│ T+240ms DEMOTION CHECK: did V0 (rank-1) suppress an honest proposal?│
│         No — V0 and V1 both included c. V0 keeps its rank.          │
│         V3's censored [a,b,z] does not match any honest input and   │
│         is ignored — V3's rank is unchanged because it wasn't the   │
│         rank-1 proposer this slot.                                  │
│                                                                    │
│ T+240ms CONTRAST: if V3 had been rank-1 and proposed [a,b,z] to     │
│         exclude c, the committed v_low would have excluded c too.   │
│         But c was in at least 2f+1=3 honest inputs — so the         │
│         protocol detects the mismatch and demotes V3 permanently.   │
│         Next slot's ranking becomes [V0, V1, V2, ].     │
└────────────────────────────────────────────────────────────────────┘

Three things to notice. First, the protocol committed after 3 rounds (240ms) with no leader, no view change, and no view-change penalty for V3's Byzantine behaviour. Second, the demotion rule requires evidence of censorship — specifically, an honest proposal that was in 2f+1 inputs but absent from the rank-1 proposer's output. The deterministic ranking makes this evidence publicly verifiable. Third, the v_high range lets the next slot pick up any transactions that weren't yet fully agreed — the protocol never "forgets" a proposal once it has been seen by f+1 honest parties.

The Demotion Rule: Why f-Censorship Resistance Is Guaranteed

The demotion rule is the key innovation that makes this protocol censorship-resistant. Here is the logic:

This means the protocol tolerates at most f censorship events before achieving permanent censorship resistance. This is provably optimal — no protocol can do better than f censorship events when facing f Byzantine parties.

Leaderless vs Multi-Proposer — The Distinction That Matters

The paper is explicit: Prefix Consensus yields a protocol that is both leaderless and multi-proposer. These are two different properties and every other production BFT protocol lacks at least one of them.

This distinction surfaced publicly in March 2026 when a widely-read X thread asked why DAG-based chains like Sui and Hedera have failed to attract trading volume, and why Solana's MCP (Multi-Client Proposal) mechanism would fare any better. Daniel Xiang (first author of the Prefix Consensus paper) replied with the protocol as the direct answer:

"DAG-based consensus is not stronger wrt censorship resistance than standard leader-based consensus. The leader/anchor can decide which 2f+1 proposals go into a block."

When Solana co-founder Anatoly Yakovenko countered that "MCP is leader-based, not leaderless," Daniel linked the paper and clarified: "MCP can be both leaderless and censorship resistant. We have some latest protocol stack on that." The paper is the technical proof that such a protocol exists.

Theoretical Contributions Beyond SMR

Beyond the full SMR protocol, the paper contributes several results that are of independent interest for distributed systems research:

These are the building blocks that make Archon's full leaderless mode practical at mainnet scale — not just theoretically sound.

Communication Complexity Deep Dive

Consensus protocols are judged on two different axes: message complexity (how many discrete messages hit the wire) and communication complexity (total bytes transferred). Prefix Consensus optimizes both, but the path from the naive bound to the production-ready bound is worth spelling out.

Naive upper bound: O(n³) messages, O(n⁴) communication. Round 1 has n parties each broadcasting a vector to n parties — that is messages. Round 2 has n parties broadcasting the computed prefix to n parties — another messages. Round 3 is the same. Total: 3n² messages, which is O(n²) — wait, the paper says O(n³). Why?

The answer is that "message complexity" in the Dwork-Lynch-Stockmeyer sense counts per-recipient deliveries, and the protocol must handle view-change-like retries across n rotations in the worst case. For each of n potential bad slots, all n parties may replay to all n parties — so n · n² = n³ in the pathological scenario. Communication is O(n⁴) because each message carries an O(n)-sized vector of hashes.

Optimization 1: threshold signatures. Replace each 2f+1-party quorum certificate (O(n) signatures) with a single aggregated BLS signature. Verification is constant-time; the QC becomes O(1) bytes instead of O(n). This drops communication by a factor of n — from O(n⁴) to O(n³).

Optimization 2: compact differing-entry encoding. As described in the trie section, parties that agree on a long common prefix need only transmit the divergent suffix. In expectation — because honest mempools overlap heavily — the divergent suffix is much smaller than the full vector. With k-length vectors that share a prefix of length k - \delta, communication drops from O(n²k) to O(n²δ). The paper proves that for typical mempool overlap patterns, δ ≪ k, giving near-linear scaling in practice.

Optimization 3: pipeline the rounds. Round r of slot s and Round r−1 of slot s+1 can share the same network flight — each party piggybacks its next-slot message onto its current-slot message. This does not reduce asymptotic complexity but cuts wall-clock latency by ~30% in global deployments.

Compared side-by-side:

ProtocolMessages (good case)Messages (worst case)CommunicationCommit Rounds
HotStuffO(n)O(n²)O(n²)3
JolteonO(n)O(n²)O(n²)2 (fast path) / 3
DAG BullsharkO(n²)O(n²)O(n²)~3 waves
MysticetiO(n²)O(n²)O(n²)3
Prefix ConsensusO(n²)O(n³)O(n³)3–4

Prefix Consensus pays a factor of n in worst-case messages and communication versus HotStuff. That cost buys leaderless operation, multi-proposer fairness, and f-censorship resistance — three properties HotStuff cannot achieve at any price.

The 2-Round Optimistic Fast Path

The paper presents a 2-round optimistic variant that commits in the good case and falls back to the 3-round protocol when the good case fails. This is how real BFT protocols cheat the asymptotic lower bound: the bound is tight for adversarial executions, but most executions are not adversarial.

The fast path condition is:

  1. All n = 5f+1 parties are honest and timely (no Byzantine behaviour, no GST violation, no one-per-round suspension)
  2. All parties' input vectors share a common prefix of length at least equal to some threshold k₀
  3. All first-round echoes agree

When all three conditions hold, a party that receives 4f+1 matching first-round echoes commits immediately — in 2 rounds. If even one echo diverges, or fewer than 4f+1 arrive within the timeout, the party falls back to Round 3 of the standard protocol. The fallback is seamless: the standard protocol's messages are a superset of the fast-path messages, so no retransmission is needed.

The n = 5f+1 requirement is why the fast path is optional. With n = 3f+1, the 2-round lower bound is impossibility. To get 2 rounds, you need a larger quorum size — specifically, 4f+1 out of 5f+1 — which shrinks the ratio of tolerable faults. Most deployments run n = 3f+1 and accept the 3-round worst case; some security-critical deployments (e.g., a custodial consortium) prefer 5f+1 and the faster fast path.

For Aptos's case: the production validator set is n ≈ 150 with economic assumptions targeting f ≈ 49 Byzantine tolerance (roughly 3f+1). Archon will ship with the 3-round worst case by default and a configurable fast-path toggle that users can opt into with a larger quorum.

Liveness Under Partial Synchrony

Strong Prefix Consensus and the full SMR protocol are liveness-guaranteed only after GST. Before GST, the protocol is still safe — no two honest parties will ever commit conflicting sequences — but it may not make progress. Understanding why this is the right tradeoff requires looking at what happens across the GST boundary.

Before GST. The adversary controls message delays. If it delays every Round-2 echo long enough, no party collects 2f+1 echoes within the round timeout. Parties retry with a longer timeout. The WPC primitive still terminates (it is fully asynchronous), so parties may commit v_low ranges locally; the SMR protocol simply accumulates uncommitted v_high tails. No safety violation is possible — the common-prefix invariant holds under any message schedule.

At GST. All in-flight messages are delivered within \Delta. Parties detect this via the success of a full round: if all 2f+1 echoes arrive on time, the timeout doubling that was in effect decays back toward \Delta.

After GST. The protocol commits v_high ranges deterministically, and within f slots (the demotion rule bound) the rank-1 slot is always filled by an honest party. From that point on, every slot commits every honest proposal within 3 rounds of the proposal's broadcast.

The post-GST worst-case latency is 3Δ + demotion-delay, where demotion-delay is at most f · 3Δ if every Byzantine rank-1 slot censors maximally. For typical settings (Δ = 80ms, f = 49), this is 3·80ms + 49·3·80ms = 240ms + 11.76s — a pathological bound that is almost never hit. In practice, Byzantine validators that systematically censor would be economically slashed long before they exhausted their rank-1 budget.

The more interesting question is what happens during flapping: GST → violation → GST → violation. The paper shows that the demotion rule accumulates across GST boundaries — a validator that was demoted does not un-demote — so over many GST transitions the honest-validator-majority is monotonically strengthened.

Why This Kills DAG-Based Approaches

In January 2026, a Twitter thread by @moonshiesty questioned why DAG-based chains like Hedera and Sui consistently failed at high-frequency trading use cases. Daniel Xiang (co-author of the Prefix Consensus paper) responded directly with the key insight:

"DAG based consensus is not stronger wrt censorship resistance than standard leader based consensus. The leader/anchor can decide which 2f+1 proposals go into a block."

This is the fundamental problem with DAG protocols (Narwhal, Bullshark, Tusk, Mysticeti). While they appear leaderless because every validator creates DAG vertices, the commit rule still depends on anchors/leaders to determine which vertices form committed sub-DAGs. The anchor can choose which 2f+1 vertices to reference, effectively censoring the remaining f vertices. This is structurally identical to leader-based censorship.

Toly (Anatoly Yakovenko, Solana) also weighed in: "MCP is leader based. Not leaderless." — pointing out that even Solana's multi-concurrent-proposer model still has leader-based ordering.

Prefix Consensus solves this by being genuinely leaderless AND censorship-resistant. No anchors, no leader election, no single point of censorship. The demotion rule ensures that even if Byzantine validators temporarily censor, the protocol self-corrects within f slots.

How Baby Raptr Implements Prefix Consensus v1 on Mainnet

Prefix Consensus is not just a paper. It is live on Aptos mainnet today through Baby Raptr (also called VelociRaptr), which implements Prefix Consensus v1.

Aptos Mainnet Consensus Stack (current):
┌─────────────────────────────────────────────────┐
│  Baby Raptr / VelociRaptr                       │
│  ├── Prefix Consensus v1 (ordering)             │
│  ├── Quorum Store (batch dissemination)          │
│  └── Pipelined commit (3-round fast path)       │
└─────────────────────────────────────────────────┘

Baby Raptr uses a simplified version of the prefix consensus approach where the multi-proposer structure is partially implemented. Validators broadcast proposals, and the consensus protocol determines which proposals form the committed block using prefix-based ordering. This is the production-hardened version that has been running on mainnet since the Raptr upgrade.

Archon: The Leaderless v2 Future

Archon is the next-generation validator coordination layer that will implement the full leaderless Prefix Consensus v2 protocol from the paper. While Baby Raptr implements a partial version, Archon will deliver:

The prefix-consensus-prototype branch in the Aptos Core repository contains active development toward Archon's consensus layer.

Integration Points in aptos-core

Tracing the Prefix Consensus protocol through the aptos-core codebase shows where each abstraction lands. The relevant paths, current as of the prefix-consensus-prototype branch:

The observable contract between Prefix Consensus and the rest of Aptos is the OrderedBlocks channel — once a block is ordered (its v_low is committed), the execution pipeline receives it with the same API it sees today. This is why Baby Raptr could ship without disturbing Block-STM or the storage layer, and why Archon will follow the same pattern.

The Encrypted Mempool + Prefix Consensus Interaction

Prefix Consensus is content-blind by design: it orders hashes of proposals, not the proposals themselves. That is the property that lets it compose cleanly with the encrypted mempool (BIBE — Batched Identity-Based Encryption).

The end-to-end flow:

  1. User-side encryption. A user encrypts their transaction payload with BIBE, producing a BIBECiphertext (G2Affine curve elements, padded symmetric key, ciphertext blob). The ciphertext is signed with Ed25519 and wrapped in a TransactionPayload::EncryptedPayload.
  2. Mempool admission. Validators verify only the signature and format of the ciphertext — they cannot read the plaintext. This is the MEV-resistance guarantee.
  3. Quorum Store batching. Encrypted transactions are batched as BatchKind::Encrypted. The batch ciphertexts are verified in parallel (per Balaji Arun's PR #19130) and the PerBatchKindTxnLimits config caps how many encrypted transactions a single batch can contain — this is the throttle that prevents an adversary from flooding the encrypted-decryption pipeline.
  4. Ordering via Prefix Consensus. The consensus protocol orders the hashes of these batches. Because the protocol is content-blind, there is no way for a Byzantine validator to reorder encrypted transactions based on their contents — it simply cannot read them.
  5. Post-ordering threshold decryption. Once v_low is committed, validators run a threshold decryption protocol (2f+1 shares) to open the ciphertexts. The decrypted transactions enter Block-STM in the order set by consensus.

The critical invariant: ordering precedes decryption. No validator sees plaintext until the order is final. Any attempt at front-running or sandwich attacks would require decrypting before ordering — which requires 2f+1 collusion, equivalent to a safety violation.

The PerBatchKindTxnLimits throttle deserves a callout. Without it, a cheap DoS becomes possible: an adversary submits encrypted-gibberish transactions that fail decryption, wasting validators' BIBE-decryption CPU. The throttle caps encrypted txns per batch (the config is tunable on-chain), so the decryption cost is always bounded regardless of adversarial input rates.

MEV, Censorship, and the Trading Application

Why does Aptos specifically care about leaderless multi-proposer BFT? Because the intended application is high-frequency trading, and trading's failure modes are dominated by MEV and censorship — not by throughput or block-time.

Consider the specific adversaries a trading venue faces:

Compare the competitive landscape from a trading venue's perspective:

ChainConsensusMEV PostureCensorship Resistance
Ethereum L1LMD-GHOST + Casper FFGProposer–Builder Separation via MEV-boost — censorship is a market outcomeLow — sanctioned addresses are demonstrably filtered by some builders
SolanaLeader-based PoH; MCP proposedJito auctions; validator front-running documentedLow — single-slot leader can censor for its leader slot window
SuiMysticeti (DAG)No encrypted mempoolMedium — DAG structure helps, but anchor validator can still exclude
MonadMonadBFT (HotStuff derivative)No encrypted mempool at launchLow — single leader per slot
Aptos (Archon)Prefix Consensus SMRBIBE encrypted mempool, content-blind orderingProvable f-censorship resistance + demotion rule

The moonshiesty / toly / Daniel thread on X in March 2026 made this rivalry explicit. Daniel's response — linking the paper as a direct rebuttal to "MCP is leader-based" — was not a marketing moment. It was a load-bearing technical claim: Aptos has a formal paper proving that the property Solana aspires to with MCP is achievable, and Aptos is the only chain implementing the primitive that achieves it.

Shoal++ Is Dead — Prefix Consensus Replaces It

For context, Shoal++ was a previously proposed enhancement to the Shoal protocol (which itself was built on Bullshark/DAG consensus). Avery Ching (Aptos CTO) confirmed that Shoal++ is no longer being actively developed. The active workstreams are:

Shoal++ was DAG-based, and as the Prefix Consensus paper demonstrates, DAG approaches have fundamental censorship resistance limitations. Prefix Consensus provides a strictly stronger primitive.

The Full Aptos Technical Stack

Here is how all the pieces fit together:

┌───────────────────────────────────────────────────────────────────┐
│                    CONSENSUS & INCLUSION                         │
│  ┌─────────────────────┐    ┌──────────────────────────────────┐ │
│  │ Raptr (live)        │    │ Archon (next-gen)                │ │
│  │ Baby/VelociRaptr    │───>│ Leaderless Prefix Consensus v2   │ │
│  │ Prefix Consensus v1 │    │ f-censorship resistant           │ │
│  └─────────────────────┘    │ Demotion rule                    │ │
│                              └──────────────────────────────────┘ │
│  ┌─────────────────────────────────────────────────────────────┐ │
│  │ Encrypted Mempool (BIBE) + Transaction Namespaces           │ │
│  │ MEV protection — txns hidden until after ordering           │ │
│  └─────────────────────────────────────────────────────────────┘ │
├───────────────────────────────────────────────────────────────────┤
│                         EXECUTION                                │
│  ┌──────────────┐  ┌──────────────┐  ┌──────────────────────┐   │
│  │ Zaptos       │  │ Block-STM v2 │  │ Shardines            │   │
│  │ Optimistic   │  │ Parallel     │  │ Validator            │   │
│  │ pipelining   │  │ execution    │  │ sharding             │   │
│  └──────────────┘  └──────────────┘  └──────────────────────┘   │
├───────────────────────────────────────────────────────────────────┤
│                         STORAGE                                  │
│  ┌──────────────────────────────────────────────────────────┐    │
│  │ Jellyfish Merkle Tree + Hot State Cache + RocksDB        │    │
│  └──────────────────────────────────────────────────────────┘    │
└───────────────────────────────────────────────────────────────────┘

Properties and Complexity

PropertyPrefix ConsensusDAG (Bullshark/Mysticeti)Leader-based (HotStuff)
LeaderlessYesNo (anchors)No
Censorship resistantf-censorship resistantNoNo
Commit rounds4 (3 fault-free)3-43
Message complexityO(n^3)O(n^3)O(n^2)
Multi-proposerYes (all validators)Yes (all validators)No (single leader)
Self-healingYes (after f slots)NoView change

Performance Targets

Hard numbers matter. The paper gives theoretical bounds; Aptos's production deployment gives empirical ones. Combining them yields the following targets under specific network assumptions:

DeploymentValidatorsΔ (one-hop)Committed TPSp50 latencyp99 latencyCensorship bound
Baby Raptr (live)~15080ms20,000750ms1.5sRotating-leader f-bounded (soft)
Baby Raptr + Zaptos (live)~15080ms20,000450ms900msSame as above
Archon target (testnet 2026)~15080ms100,000+320ms650msProvable f-CR (paper)
Archon + Shardines (future)~15080ms1,000,000+320ms650msProvable f-CR
Archon 2-round fast path~150 (5f+1)80ms100,000+220ms450msProvable f-CR

Assumptions: geographically distributed validators across 6 continents, Δ measured as the 80th-percentile one-hop latency, batch size tuned for ~2500-txn batches, encrypted mempool enabled. The p50 numbers come from Raptr paper experiments; p99 is derived from the 3·Δ + demotion-delay worst-case.

Two things to note. First, Archon does not ship faster p50 than Baby Raptr + Zaptos automatically — the latency improvement comes from the elimination of leader-based pipelining bubbles, not from protocol rounds. Second, the "provable" column is where Archon genuinely differentiates: Baby Raptr's censorship resistance is a property of the rotating-leader schedule; Archon's is a theorem.

Roadmap: Prefix Consensus to Archon Deployment

The path from paper to mainnet, staged:

  1. Paper (published Feb 2026). arXiv:2602.02892 defines WPC, SPC, the 3-round protocol, the SMR construction, and the demotion rule. Peer review via cryptography and distributed-systems venues is in progress.
  2. Prototype (aptos-core prefix-consensus-prototype branch, active). Reference implementation of the 3-round primitive, trie-based common-prefix, and the demotion-rule bookkeeping. Integration tests run against a 4-to-16-validator simnet.
  3. Devnet integration (mid-2026 target). Archon's consensus path merges into aptos-core's main consensus trait. The existing Baby Raptr path stays live behind a config flag; Archon runs on an isolated devnet cluster.
  4. Testnet activation (Q4 2026 target). Archon replaces Baby Raptr on a public testnet. Validators upgrade; the mainnet path stays on Baby Raptr. Soak-testing focuses on the demotion rule under adversarial stress.
  5. Mainnet flag-flip (2027 target). Once Archon has sustained a configurable number of epochs on testnet with no safety or liveness regressions, a governance vote activates it on mainnet. Activation criteria include: no demotion-false-positives in 30 consecutive days, p99 latency within 1.5x of Baby Raptr, no regressions in encrypted-mempool decryption throughput.
  6. Shardines composition (2027+). Archon composes over Shardines' per-shard consensus substreams, multiplying throughput by the shard count. This is where the million-TPS target actually lands.

The prototype branch today contains enough to run a 7-validator simnet. The gap to mainnet is primarily soak-testing and the demotion-rule bookkeeping (which needs to survive validator-set reconfiguration across epochs — see next section).

Limitations and Open Problems

The paper is precise about what it proves, and precise about what it does not solve. Honest engineering requires acknowledging both.

None of these are show-stoppers. They are the kind of "known unknowns" that separate a paper from a mainnet deployment — and they are why Archon is targeted for 2027 rather than 2026.

Key Contributors

ContributorRoleContribution
Zhuolun Xiang (Daniel)Lead author, primary implementer68% of Prefix Consensus commits, contributor across Raptr, Zaptos, Archon, and encrypted mempool
Andrei TonkikhCo-author, protocol design18% of Prefix Consensus commits
Alexander SpiegelmanCo-author, theory and proofs11% of Prefix Consensus commits
Balaji Arun (ibalajiarun)Encrypted mempool implementationBIBE encryption layer, replay protection

The Bottom Line

Prefix Consensus is not just an academic contribution. It is the theoretical foundation for Aptos's transition from a leader-based to a genuinely leaderless consensus protocol with provable censorship resistance. Baby Raptr runs v1 on mainnet today. Archon will bring the full v2 with the demotion rule and f-censorship resistance guarantee. Combined with the encrypted mempool, Zaptos pipelining, Block-STM v2 parallel execution, and Shardines sharding, Aptos is building a stack where censorship resistance is not a feature to be added later — it is baked into the consensus primitive itself.

Paper: arXiv:2602.02892 — "Prefix Consensus For Censorship Resistant BFT" (Feb 2026)


ELI5 — Explain Like I'm 5

The Big Picture: Imagine a classroom where every student has an idea for what the class project should be. In most classrooms, there is one class president (the "leader") who picks which ideas to use. The problem? If the class president is unfair, they can ignore certain students' ideas every time. This is called censorship.

Why One Class President Is Dangerous: Think about a post office where only one person — the postmaster — decides which letters get delivered and in what order. If the postmaster has a grudge against someone, they can "accidentally" lose that person's mail for weeks. Nobody else sees it happen. Most blockchains today work exactly like this post office: one validator at a time picks which transactions get included in the next block. If that validator doesn't like you — maybe you're trading against their friends, or maybe your address is on a sanctions list they want to enforce — they can just leave you out. You won't even know why.

What "Censorship" Actually Means in Trading: Say you're a trader and you want to sell 1000 shares at $10 each. You send that order to the blockchain. In a leader-based chain, the leader validator sees your order first — before anyone else. If they also run a trading desk, they can buy ahead of you (knowing your sell is coming), then let your sell execute at a slightly worse price, then sell what they just bought at a profit. That's called a "sandwich attack." Or they can just drop your order entirely if they'd rather their friends' orders fill first. Both of those are forms of censorship, and both cost real traders real money every day on every leader-based chain that exists.

What Prefix Consensus Does: Prefix Consensus is like a new classroom rule where there is no class president. Instead, every student writes their idea on a piece of paper and passes it around. Every student looks at all the ideas and figures out which ones everyone agrees on — starting from the ideas at the top of the list and working down. The ideas that everyone saw and agreed on get added to the project. No single student gets to decide what is included or excluded. Imagine a voting booth where every voter drops in the same list of candidates they support — and we only commit to the candidates that appear on a majority of lists. No precinct captain decides who to count.

The Demotion Rule — Voted Off the Island for Being a Bad Proposer: What about cheaters? If a troublemaker tries to hide someone else's idea (censorship), the class notices immediately, because the ranking of who-proposes-first is known in advance. The troublemaker gets "demoted" — sent to the back of the line so they have less influence in the next round. It's like being voted off the island on a reality show: once you've been caught pulling a fast one, the tribe remembers, and you don't get another turn at the front. Since there can be at most a few troublemakers in the whole class (a third at most), after each one gets caught and sent to the back, eventually only fair students are left in the front. From that point on, every honest student's idea gets included every time. The class self-corrects, without anyone needing to file a complaint or run a new election.

Why This Is Different from a DAG: Some other blockchains (like Sui and Hedera) use something called a DAG — imagine a swarm of delivery drones all flying parcels around, any-to-any. Sounds fully parallel, right? But here's the trick: at the end of the day, someone still has to pick which drones' parcels actually get loaded onto the delivery truck. That "someone" is called the anchor, and it's really a hidden leader. The anchor picks 2-out-of-3 of the drone parcels and leaves the last third sitting on the tarmac. So even though the drones all flew in parallel, a single anchor still decides what ships. Prefix Consensus doesn't have an anchor. The decision about what gets loaded onto the truck is the common overlap of every drone's independent flight log — no one person chooses.

How the Auction House Analogy Works: Another way to see it. Picture an auction house where a hundred auctioneers are all calling bids at the same time. In a normal auction, one auctioneer decides whose bid counts — they can "miss" a bid from someone they don't like. In Prefix Consensus, every auctioneer writes down the bids they hear. At the end, we only count bids that most auctioneers wrote down. If one auctioneer systematically misses Alice's bids, the mismatch is loud — everyone else wrote her bid down, and that one auctioneer didn't. The cheating auctioneer gets fired. After enough cheaters are fired, only honest auctioneers are left, and no bid gets "missed" again.

Where This Lives Today: Aptos already uses a simpler version of this (called Baby Raptr) on its live network. The full version — with the cheater-demotion rule — is being built into something called Archon, which will make Aptos the first major blockchain with provable censorship resistance built directly into its consensus. The paper that proves it works came out in February 2026, written by three Aptos researchers: Daniel Xiang, Andrei Tonkikh, and Alexander Spiegelman.

Why It Matters for Trading: If you're a market maker, a DeFi user, or anyone who cares whether your transactions land on-chain in a fair order, this is the first chain where the consensus layer — the deepest layer of the system, below everything else — is mathematically designed so that no single validator can front-run you or ignore your orders. Combined with the encrypted mempool (which hides your transactions from validators until after they've been ordered), this is the closest any blockchain has come to a provably-neutral ordering service. That's the pitch.

What You Learned: You just learned about Prefix Consensus — a new way for computers to agree on things without needing a leader, where cheaters automatically get pushed aside. The post office has been replaced with a neighborhood bulletin board where every neighbor signs off on which flyers go up. The class president has been replaced with a rule that says "anything most of the class saw gets included." And the drone fleet doesn't have a dispatcher anymore — the overlap of their flight logs is the delivery manifest. It's the foundation for making Aptos censorship-resistant at the deepest level of the system, and it's the reason the research team at Aptos Labs believes this is the stack that finally makes on-chain trading viable at the level real markets need.


Related Systems

RaptrArchonZaptosBlock-STMShardinesEncrypted Mempool

Other Deep Dives


View this report interactively with Advanced / ELI5 tabs at https://aptos-intelligence.vercel.app/#prefix-consensus-deep-dive. Plain-text version: /reports/prefix-consensus-deep-dive.txt.