ProjectsBlogpscoAbout

42 | Strong Eventual Consistency

18 September, 2022 - 44 min read

An Introduction to the Album

Perhaps the most important part of a post-emo song is its totally incomprehensible song title. Not the arrangement, lyrics, unique tuning. No, it's the title which must be devoid of any relation to the contents or meaning of the song itself.

When viewed at a glance, any punk rock tracklist worth its salt ought to be jarring, immersion-breaking, and puzzlingly disparate from message of the song, if any such meaning is even forthcoming. But, over time, ~eventually~ this pattern becomes an appreciable aesthetic touch, and strongly enough so that the zany titles become integral to the eidetic essence of an album or sequence of songs. To borrow the language of the papers discussed later on, the song titles are consistently offputting, with several bands appearing to pride themselves on incoherent discographies, which converge on a reliably pleasant listening experience.

† - I know at least a few readers will take offense with my liberal blending of genres, but they're welcome to respond with their own blog post.1

The analogy for the actual subject matter of discussion for this post kind of fizzles out here, but I'm going to lean into it regardless, trying (and failing) to draw connections between red-herring titles and conflict resolution in distributed systems.

I actually had no strong motivation to write about eventually consistent data types. I thought it would be a quirky playlist name, based on a phrase I vaguely remember reading several years ago in a paper about ¿key-value stores? or something like that. But I couldn't remember their significance, nor how they worked. So, naturally, I've spent the past couple weeks re-familiarizing myself with that paper, leapfrogging to contemporary developments, and naaturally convinving myself that I understand conflict resolution as thoroughly as a hostage negotiator. This is the result of that rabbithole.

Scott Pilgrim v. My GPA

(The Byzantine Problem)

In distributed systems striving for eventual consistency (definition forthcoming), The Byzantine Problem is an analogy describing the need for a resolution protocol for imperfect information or component failure. Oftentimes, such protocols rely on consensus voting, proof of work, or prioritization of "leading" trustworthy nodes in the system.

A compromised component in a distributed system can stochastically exhibit symptoms of success and failure in completing whatever the distributed task is. A distributed task might be something along the lines of tallying views on a YouTube video,2 a simple fork-join task, or coordination of battle strategy against Ottoman combatants.

This presents the problem of isolating, or at the very least mitigating, the spread of compromised information to the rest of the system, and this introduces the measure of severity of such a breach to a system's integrity: Byzantine Fault Tolerance – the resilience of a system on the whole to such conditions.

In the titular metaphor, a group of generals must come to a consensus about whether to attack or to retreat. Each component of the distributed system (commonly referred to as a replica in the literature) may have its own strategic preference or signal (in this example, Attack\mathbf{A}ttack or Retreat\mathbf{R}etreat - though signals need not be binary in general). To establish urgency in our hypothetical, we asssume that a coordinated attack or retreat by all the generals (the system on the whole) is favorale to a fractured mobilization as the result of miscommunication, where each general acts according to their own strategy rather than that of the whole army.

To complicate the situation, some generals may be selectively treacherous (probabilistically, intentionally, etc. The Internet is a big place, and lots can go wrong), sending signal AA to some partition of their confederates, and signal R\mathbf{R} to the others in an attempt to sew chaos on behalf of the Turks.

Additionally, as the army is physically separated (distributed) the integrity of their messages is at the whim of their messengers. Messengers can be bribed, intercepted, or be UDP packets: they might not correctly reflect the intent of the sender, nor even be received. Worse yet, they might not be consistent in their success or failures, making this a Byzantine problem.

Words That Rhyme with Different, Etc.

(Characteristics of Byzantine Problems)

  • A Byzantine Fault occurs if any component in the system presents different symptoms to different observers.
  • A Byzantine Failure occurs if a system on the whole is compromised due to Byzantine Faults in a system where consensus is required.
  • A Byztantine Fault Tolerance is achieved if non-faulty components have a majority agreement on strategy (or whatever the subject of consensus is). "Majority," here, is also loosely defined. Some systems need only reach 51% consensus to function properly (such as a blockchain), others might require a 2/3 super majority (like a different blockchain), and some might need perfect accuracy (like sales of a precisely controlled commodity).

Formally: Given a system of nn components, where t<nt < n components are truthful, as well as a peer to peer communication channel, whenever component AA tries to broadcast a value XX, the other components are allowed to discuss with each other and verify the integrity of AA's broadcast, the system will eventually settle on a common value YY.

Such a system is said to resist Byzantine Faults if AA broadcasts some value xx and:

  1. if AA is truthful, then all components agree on the value xx
  2. in any case, all truthful components agree on the same consensus value yy.

i just want to kickflip into the sunset and disappear

(Trivial Solutions to the Coordination Problems )

A default initial value or assumed vote can be given to all null or missing messages from unresponsive (or faulty!) components deemed untrustworthy or unreliable. Interpretation of the null value can be prearranged by the constituents of a system.

  • Such a prearrangement fails for the Byzantine since, not only can reaching consensus be prevented by incomplete or faulty signals with otherwise perfect knowledge of the system, but reaching consensus may be prevented about whether or not a component is faulty in the first place.

Lvl. 2 Pidgey in a Masterball

(Solutions to Byzantine Problems)

The two primary solutions to Byzantine Problems I want to discuss here are Conflict-free Replicated Data Types (CRDTs)3 and more sophisticated conflict-resolving data structures called HashGraphs.

J-Kobb Schlepper (My Dog At My Paypal Account)

(Intro to CRDTs)

CRDTs were developed in response to the existential problem posed by the need for eventual consistency that cannot be ameliorated by angsty music alone. Instead we have this "simple, theoretically sound approach," covered at length in the original 2011 paper by Shapiro et al: "A comprehensive study of Convergent and Commutative Replicated Data Types."4 CRDTs in general aim to do away with the need for reaching consensus. In doing so, they indirectly address the Byzantine Problem by defining a protocol for offline, stateful consensus.

In the offline scenarios specifically, if attempts are made to talk off-machine, we want to enqueue outbound messages and give reponses deadlines rather than throwing an exception wherever possible. In the case where we cannot adhere to this rule, we need to gracefully handle the offline scenario. E.g., when talking to a centralized authorization service, a user may be locked out if we (acting as the Auth Service) can't verify their security access permissions due to a network outage. Once we are again able to verify their requisite permissions, and a validity indow for the answer, we should cache the response to prevent a lapse in availability for them in the future.

It is not possible to come to consensus with a system that cannot be communicated with. Even under a resolution protocol which uses default null values for missing votes as would be the case for outbound messages in the above offline example, only one side of the partition can be in the majority (Servers A and B), and the minority does not want to lose its state outright.

It is fruitless to strive for agreement in such scenarios, but a consistent view of the past with distributed conflict resolution is achievable.

A protocol like Paxos for achieving consenus involves seeking a majority vote of a leader, and voting on the next move that the state machine is to take. On the event of a hopefully-exceptional network partition, updates propagating across the network need to terminate in the minority partition (Server C), lest the state be lost when the partition (hopefully) heals.

There are numerous drawbacks to this framework, the most obvious being the assumption that partitions are infrequent and/or brief.

dragon ball z budokai tenkaichi 4

(Types of CRDTs)

According to the seminal paper, for replicas to return the same result for all query operations after such a partition is introduced, two conditions must be met:

  • safety: If the causal history of replicas ii and jj is the same, the abstract state of these replicas must be equivalent
  • liveness: If some element ee is in the causal history of ii, then it will eventually be in the causal history of jj

Inductively, the pairwise eventual consistency defined by these two poperties guarantees convergence for any non-empty subset of replicas contingent on the eventual receipt of all updates.

Fuck, Dantooine Is Big

(Example of a CRDT)

Suppose we have several servers spread across the world – in warzones, hotbeds of natural disaster, and other areas prone to network outages like my mom's basement whenever I hop on the sticks in Warzone: say Aleppo, Abescon, and Alexandria.

We have the simple task of utmost importance: track the highest value observed by any of the severs (my killstreak 😤 😤) across our shared database. Our goal is to provide a protocol with properties of commutativity, associativty, and idempotency. (this all sounds awfully functional to me).

In this trivial example, a merge function like max(a,b)max(a,b) satisfies our needs. Within each replica, if a<ba < b, we can ignore, dismiss, discard value aa in favor of bb. After each update, a replica will broadcast to all its peers indicating that a new maximum value has been observed (loadout drop is inbound). These updates will converge on the system's known maximum value. This is similar to the common leader election algorithm, except in this case, the updates are continuous, whereas an election has a discrete end once all votes are caster and tallied, wherease this system can always observe a new, higher value.

Each server keeps track of the highest observed value for itself as well as for all other replicas, and updates those counters indexed by replica. For example, Server A's internal count might be:

{
    "A": 11,
    "B": 20
}

and Server B, having received one final update from Server C before my dominance got the home router taken offline:

{
    "A": 4,
    "B": 28,
    "C": 8
}

When A and B next communicate with one another, each will attempt to merge their respective states by taking the maximum value observed for the same key by two replicas:

merge({ "A": 11, "B": 20}, { "A": 4, "B": 28, "C": 8}) 
= {
    "A": 11,
    "B": 28,
    "C": 8
}

(Joe Gets Kicked Out of School for Using) Drugs With Friends (But Says This Isn't a Problem)

(More Formal Defintions)

There are two main categories of CRDTs:

  • commutative - which rely on replicated operations – designed to commute
  • convergence - which rely on replicating state, executing operations on a primary component

In both cases, the higher order goal is to avoid the need for coordination by ensuring that actions taken independently can't conflict with each other and thusly can be composed at later point in time [eventually].

Visited Salmon, I Mean Transit Balcony

(State-Based CRDTs)

State-Based CRDTs maintain replicated state across a distributed system via composable Least Upper Bound functions which must be defined over their constituent data. In the prior example, the max()max(\cdot) function is a literal LUB, but our data will usually have a more complex structure than singular numbers.

Formally, our set of data values together within a LUB-based partial order function forms a join semilattice5 where the "join" is a property of the LUB which govers how the lattices may be merged. If values only ever increase, the lattice is monotonic, and said to be a CRDT.

Vanilla-Scented Laser Beams

(Operation-Based CRDTs)

This variety of object assumes a channel which maintains a linear ordering of message delivery to communicate between replicas. Operations outside of this delivery order are called "concurrent," and if they are commutative, than all operations consistent with the order of the deliver channel are equivalent, so the distributed state will be convergent across all replicas.

Hey Ken, Someone Methodically Mushed the Donuts

(Nuff Bout CRDTs – this would be an interlude)

CRDTs avoid the need for consensus by instead relying on definitions for behavior in light of conflicting observations.

It's feasible to due away with git merge conflicts entirely if the version control system is treated as a CRDT with some rule which enforces a Least Upper Bound on the commit history. For example, simply take the most recent provided timestamp of each conflicting event to be the "maximum." This obviously doesn't work in practice, but it's possible.

Umbrellas and Beersocks

(The FLP Theorem)

Before diving into the advancements offered by HashGraphs, it's first worthwhile to survey problems of distributed consensus and pickup some useful terminology. The most relevant to the discussion about HashGraphs which do rely on consensus (sort of) being the eponymous Fischer-Lynch-Paterson Theorem introduced in "Impossibility of Distributed Consensus with One Faulty Process."6

The problem the authors present is that getting reliable, asynchronous procceses –amidst some faulty ones– to agree on even just one binary value is very hard. They note that every protocol in this asynchronous case can be a victim of possibile nontermination with even just one faulty process.

The example problem instance they use is referred to as the "transaction commit problem" for a distributed database which is roughly similar to the aforementioned git merge conflict problem: all the database managers that have participated in processing a transaction must agree on whether or not to accept the transaction's results on the database.

Whatever decision is made must be executed by all managers in order to maintain consistency across the replicas i.e. consensus must be reached. If all of the managers are reliable, then the solution is trivial. However, even non-Byzantine faults such as crashes, network partitions or outright failures, lost, distorted, and/or duplicate messages present problems for the trivial (arguably naive) scenario where replicas trust one another.

The titular crux of the paper is that no completely asynchronous protocol can tolerate even a single unnancounced process death. This constraint is so overbearing that the authors don't even grapple with the Byzantine complications, and further assume reliable message channels. They make these nifty assumptions to show the fragility of distributed systems in general, conversely making their main theorem as widely applicable and robust as possible.

The asynchronicity of the systems being considered is also integral to their asserted fragility. No assumptions are made about relative speeds of processes, or delays in their receipt. Processes don't have access to a universal, synchronized clock, so timeout-based algorithms are not applicable. Therefore, process death is indistinguishable from a slow process or high network latency.

Thrashville 1/3

(Model)

Processes are state machines with potentially infinitely many states which communicate via messages sent between each other. In one discrete time step, a process may attempt to:

  • Attempt to receive a message by performing a local computation as to whether an message was delivered to it
  • Send arbitrarily finite amounts of messages to other processes

If any truthy processes receive the message, then via an assumed atomic broadcast capability (from the generous assumption of a reliable communication channel), all truthy processes eventually will too. In other words, every message is eventually delivered as long as the other processes make sufficiently infinite attmpts to receive it.

There's some lore on the matter:

The asynchronous commit protocols in current use all seem to have a “window of vulnerability”- an interval of time during the execution of the algorithm in which the delay or inaccessibility of a single process can cause the entire algorithm to wait indefinitely. It follows from our impossibility result that every commit protocol has such a “window,” confirming a widely believed tenet in the folklore.

A Consensus Protocol PP is an asynchronous system of N2N \geq 2 proccesses.

  • Each process pPp \in P has an input register xpx_p, an output register yp{b,0,1}y_p \in \{b, 0, 1\}, and an unbounded amount of internal storage.
  • Values in the input and output registers, together with the program counter and internal storage, comprise the internal state of a process.
  • Initial States prescribe fixed starting values for all but the input register. The output register is initially set to bb.
  • Decision States are those in which the output register has a value of 0 or 1.
  • pp acts deterministically according to a transition function which cannot change the value of the output register; ypy_p is "write-once."
  • Messages between processes are tuples of the form (p,m)(p, m) where

    • pp is the target processes
    • mm is the message value in some alphabet MM
  • The Buffer is a multiset of messages that have been dispatched but not yet received, and supports two operations:

    • send(p,m)send(p,m): places (p,m)(p, m) in the message buffer
    • receive(p)receive(p): may delete some message (p,m)(p,m) from the bufferif it exists and returns mm, in which case we say that the message was delivered, otherwise it returns a special null marker \varnothing indicating that the buffer was unchanged (and the process, message combination was not found)

The buffer can act non-deterministicallly (at the hands of non-truthy processes, perhaps), subject only to the condition that if receive(p)receive(p) is performed infinitely-many times, then the buffer is eventually emptied: all messages are delivered.

A Configuration CC of the system PP is comprised of the internal state of each process, together with the contents of the message buffer.

  • An Initial Configuration is one in which each process starts at an initial state and the message buffer is empty.
  • A Step takes one configuration to another via primitive step by a single process pp and occurs in two phases:

    • First, receive(p)receive(p) is performed on the message buffer in CC to obtain m{M}m \in \{M \cup \varnothing\}
    • Second, depending on pp's internal state CC, as well as mm, the process pp enters a new internal state and sends a finite set of messages to other processes.

Since processes are deterministic, the step is completely described by the event e=(p,m)e = (p, m) which can be thought of as receipt of mm by pp. We say e(c)e(c) denotes the resulting config, or that ee can be applied to CC.

  • (p,)(p, \varnothing) can always be applied to CC, so it is always passible to take another step. This event is the identity on CC. (are your categorical senses tingling yet?)

A Schedule from CC is a possibly infinite sequence of σ\sigma events that can be applied in order, starting from CC. The associated sequence of steps is called a run. If σ\sigma is finite, σ(C)\sigma(C) is the resulting configuration, said to be reachable from CC. We only need consider configurations accessible from some initial configuration.

  • Lemma: Suppose that from some configuration CC, the schedules σ1,σ2\sigma_1, \sigma_2 lead to C1,C2C_1, C_2 respectively. If the sets of processes taking steps in these schedules are disjoint, then σ1(C2)\sigma_1(C_2) and σ2(C1)\sigma_2(C_1) lead to the same configuration C3C_3

A configuration CC has decision value vv if some process pp is in a decision state such that yp=vy_p = v.

A consensus protocol is said to be partially correct iff

  1. No accessibe configuration has more than one decision value
  2. For each v{0,1}v \in \{0,1\},8 some accessible configuation has decision value vv

A process pp is non-faulty ("truthy") in a run provided that it takes infinitely many steps, and is faulty 🎩 otherwise. A run is admissable provided that at most one process is faulty and all messages to the other truthy processes eventually received.

The main theorem of the paper shows that every partially correct protocol for the consensus problem has some admissable run that is not a deciding run. Therefore, no consensus protocol it totally correct in spite of one fault.

The rest of the the paper includes supporting lemmas and further stipulations about conditions of failure, which are all fascinating in their own right, but need not be replicated here.

A natural and important problem of fault-tolerant cooperative computing cannot be solved in a totally asynchronous model of computation. These results do not show that such problems cannot be “solved” in practice; rather, they point up the need for more refined models of distributed computing that better reflect realistic assumptions about processor and communication timings, and for less stringent requirements on the solution to such problems

frankie thinks she's punk rock but she's just a POSER!!! (demo)

(Swirlds HashGraph – AKA Solving the Problem in Practice)

With the FLP Theorem in our back pocket, we are ready to confront the HashGraph approach to achieving fair, fast and Byzantine-Fault-tolerant conensus developed by Swirlds. The distributed S/PaaS company published their white paper including proofs of the fault-tolerance of their HashGraph approach in 2016,9 and hot dayum it seems pretty cool.

The abstract claims that their proposal for a new kind of replicated state machine

achieves fairness, in the sense that it is difficult for an attacker to manipulate which of two transactions will be chosen to be first in the consensus order. It has complete asynchrony, no leaders, no round robin, no proof-ofwork, eventual consensus with probability one, and high speed in the absence of faults. It is based on a gossip protocol, in which the participants don’t just gossip about transactions. They gossip about gossip. They jointly build a hashgraph reflecting all of the gossip events. This allows Byzantine agreement to be achieved through virtual voting. Alice does not send Bob a vote over the Internet. Instead, Bob calculates what vote Alice would have sent, based on his knowledge of what Alice knows. This yields fair Byzantine agreement on a total order for all transactions, with very little communication overhead beyond the transactions themselves.

Whereas the authors of the FLP theorem offered generous assumptions about the friendliness of the system at play, Swirlds assumes the opposite: that the system and its external conditions be as hostile as possible. The rhetorical purpose being the same – to prove the most robust form of applicability and/or correctness as possible. They use a strong definition of "Byzantine" to mean that just under 1/31/3 of all components within the system can be adversaries colluding with one another, deleting or corrupting or delaying messages, who also can have full control of the network (and thus, the communication channel) with the same caveat offered by the FLP theorem, that if a truthy component infinitely attempts to send a message to another component, it must be received eventually.

The system is totally asynchronous. It is assumed that for any honest members Alice and Bob, Alice will eventually try to sync with Bob, and if Alice repeatedly tries to send Bob a message, she will eventually succeed. No other assumptions are made about network reliability or network speed or timeout periods. Specifically, the attacker is allowed to completely control the network, deleting and delaying messages arbitrarily, subject to the constraint that a message between honest members that is sent repeatedly must eventually have a copy of it get through.

In spite of the FLP theorem above, the authors assert that their completely asynchronoush, nondeterministic HashGraph achieves consensus with a probability of 1.

They acknowledge other consensus algorithms like proof of work blockchains, and leader-based voting, as well as other Byzantine protocols which attempt to address the pitfalls of the former two, and point out that even the lattermost approaches require up to O(n3)O(n^3) message exchanges to achieve consensus.

HashGraph sends no votes at all over the network, because all voting is virtual.

Welcome to Castle Irwell

(Core Mechanisms)

It's worth noting that a HashGraph is a Graph of Hashes, and not the other way round: something akin to a HashMap of Graphs, though that's also fun to think about. They outline the core mechanisms of the HashGraph as follows.

Eventualities

(Events)

  • Generally, transactions under dispute are called events. An event xx is defined to be an ancestor of event yy if xx is yy, or a parent of yy, or a parent of a parent of yy and so on. It is also a self-ancestor of yy if xx is yy, or a self-parent of yy, or a self-parent of a self-parent of yy and so on.
  • The round of an event xx is defined to be r+ir + i, where rr is the maximum round number of the parents of xx (or 11 if it has no parents), and ii is defined to be 11 if xx can strongly see more than 2n/32n/3 witnesses in round rr (or 00 if it can’t).

    • Lemma: If hashgraphs AA and BB are consistent, and AA decides a Byzantine agreement election with result vv in round rr and BB has not decided prior to rr, then BB will decide vv in round r+2r + 2 or before.
    • This lemma provides a bound on the eventuality of strong consistency, though implicit assumptions of network partitions or lackthereof need still be considered.

Had 2 Try

(Example of Round Delineation)

Here, all events except for a5,d3,d4a_5, d_3, d_4 are in the initial round (round 1). Those 3 nodes are able to strong see a supermajority of all nodes in round r1r-1 (a1,b1,c1,d1a_1, b_1, c_1, d_1), therefore they all sign themselves as being members of round 2, and all of their parent nodes will also have at least round 2 status.

  • The round received number of an event xx is defined to be the first round where all unique famous witnesses are descendants of xx. If any event strongly sees a supermajority of events from the current or previous round r,r1r, r-1, it advances its round counter to rr+1r \leftarrow r + 1.
  • A pair of events (x,y)(x, y) is a fork if xx and yy have the same creator, but neither is a self-ancestor of the other.

    • Lemma: If the pair of events (x,y)(x, y) is a fork, and xx is strongly seen by event zz in hashgraph AA, then yy will not be strongly seen by any event in any hashgraph BB that is consistent with AA.
  • An honest ("truthy") member tries to sync infinitely often with every other member, creates a valid event after each sync (with hashes of the latest self-parent and other-parent), and never creates two events that are forks with each other.

Any member can created a signed transaction at any time. All members get a copy of it (eventually), and the system converges on Byzantine agreement of the linear ordering of those transactions.

It is not enough to ensure that every member knows every event. It is also necessary to agree on a linear ordering of the events, and thus of the transactions recorded inside the events.

Which is particularly challenging in the asynchronous setting, as illustrated by the FLP theorem: every such exchange or reliance on an ACK introduces an oportunity for unavoidable failure.

Our Love Is Dog

(Fork Cheating Example)

Suppose Bob creates an event xx with a certain self-parent hash pointing to his previous event zz. Then Bob creates a new event yy, but gives it a self-parent hash of zz, instead of giving it a self-parent hash of xx as he should. This means that the events by Bob in the hashgraph will no longer be a chain, as they should be. They will now be a tree, because he has created a fork. If Bob gossips xx to Alice and yy to Carol, then for a while, Alice and Carol may not be aware of the fork. And Alice may calculate a virtual vote for xx that is different from Carol’s virtual vote for yy. So it is possible for a fork to be spread across consistent hashgraphs. In this case, there may be a moment when Alice has a hashgraph containing xx but not yy, and Carol has a hashgraph with yy and not xx, and so a fork exists, but neither member is yet aware of the fact that it is a fork.

The hashgraph consensus algorithm prevents this attack by using the concept of one state seeing another, and the concept of one state strongly seeing another. These are based on definitions of ancestor and self-ancestor such that every event is considered to be both an ancestor and self-ancestor of itself.

shlonkey kong

(HashGraph)

The HashGraph is data structure that records who gossiped to whom, and in what order. Every member has a copy of the hashgraph. If Alice and Bob both have the same hashgraph, then they can calculate a total order on the events according to any deterministic function of that hashgraph, and they will both get the same answer. Therefore, consensus is achieved, even without sending vote messages.

Whereas git maintains a graph of cryptographic hashes representing versions in a tree, where edges are the diffs between them, it stores no record of how the components communicated. The HashGraph primarily records the history of how its components communicated. "Alice send Bob all the events that she knows that he doesn't"

Suppose Alice has hashgraph AA and Bob has hashgraph BB. These hashgraphs may be slightly different at any given moment, but they will always be consistent. Consistent means that if AA and BB both contain event xx, then they will both contain exactly the same set of ancestors for xx, and will both contain exactly the same set of edges between those ancestors. If Alice knows of xx and Bob does not, and both of them are honest and actively participating, then we would expect Bob to learn of xx fairly quickly, through the gossip protocol. The consensus algorithm assumes that will happen eventually, but does not make any assumptions about how fast it will happen. The protocol is completely asynchronous, and does not make assumptions about timeout periods, or the speed of gossip, or the rate at which progress is made.

  • Hashgraphs AA and BB are consistent iff for any event xx contained in both hashgraphs, both contain the same set of ancestors for xx, with the same parent and self-parent edges between those ancestors.

    • Lemma: All members have consistent HashGraphs.

if ur leffen then im chillindude

(Gossip about Gossip)

The HashGraph is spread through the gossip protocol. the information being gossiped about is the history of the gossip itself. This uses very little bandwidth overhead beyond simply gossiping the events alone.

  • A member such as Alice will choose another member at random, such as Bob, and then Alice will tell Bob all of the information she knows so far. Alice then repeats with a different random member. Bob repeatedly does the same, and all other members do the same. In this way, if a single member becomes aware of new information, it will spread exponentially fast through the community until every member is aware of it.

Gossiping a hashgraph gives the participants a great deal of information. If a new transaction is placed in the payload of an event, it will quickly spread to all members, until every member knows it. Alice will learn of the transaction. And she will know exactly when Bob learned of the transaction. And she will know exactly when Carol learned of the fact that Bob had learned of that transaction. Deep chains of such reasoning become possible when all members have a copy of the hashgraph. As the hashgraph grows upward, the different members may have slightly different subsets of the new events near the top, but they will quickly converge to having exactly the same events lower down in the hashgraph. Furthermore, if Alice and Bob happen to both have a given event, then they are guaranteed to also both have all its ancestors. And they will agree on all the edges in the subgraph of those ancestors. All of this allows powerful algorithms to run locally, including for Byzantine fault tolerance.

Gossip about gossip requires little overhead communication (compared to gossip about the transactions themselves) as components comunicate only the diffs of their signatures of observing new transactions. Even these signatures can be compressed: "With appropriate compression, this can be sent in very few bytes, adding only a few percent to the size of the message being sent."

Meet Me in Montauk

(Virtual Voting)

Every member has a/is a copy of the HashGraph, so Alice can compute what Bob would have sent her, if they had been adhering to a traditional Byzantine consensus protocol which required vote-sending. Every member can reach agreement on any number of decisions, without a single vote ever being sent. The HashGraph alone is sufficient. So zero bandwidth is used, beyond simply gossiping the HashGraph.

Virtual voting has several benefits. In addition to saving bandwidth, it ensures that members always calculate their votes according to the rules. If Alice is honest, she will calculate virtual votes for the virtual Bob that are honest. Even if the real Bob is a cheater, he cannot attack Alice by making the virtual Bob vote incorrectly

  • Lemma: For any single YES/NO question, consensus is achieved eventually with probability 11

Voting on the acceptance of a proposed event occurs as follows:

  1. An event xrx_r in round rr computes whether it is able to see the prior transaction yr<ry_{r' < r}
  2. xrx_r computes whether it is able to strongly see the yr<ry_{r' < r}
  3. xrx_r computes whether it is able to strongly see a supermajority of events
  4. If so, it advance to a new round, and checks if a supermajority of nodes in r+1r+1 can strongly see a witness of rr
  5. If so, the witness of rr is considered famous, and yr<ry_{r' < r} is committed

The "virtual" part occurs during (3) and (4) by computing the visibility vectors of other nodes locally.

If any event strongly sees a supermajority of events from the current or previous round r,r1r, r-1, it advances its round counter to rr+1r \leftarrow r + 1.

Art School Wannabe

(Famous Witnesses)

A witness is the first event created by a member in a round

The community could put a list of nn transactions into order by running separate Byzantine agreement protocols on O(nlogn)O(n \log n) different yes/no questions of the form “did event xx come before event yy?” A much faster approach is to pick just a few events (vertices in the hashgraph), to be called witnesses, and define a witness to be famous if the hashgraph shows that most members received it fairly soon after it was created. Then it’s sufficient to run the Byzantine agreement protocol only for witnesses, deciding for each witness the single question “is this witness famous?” Once Byzantine agreement is reached on the exact set of famous witnesses, it is easy to derive from the hashgraph a fair total order for all events.

Shred Cruz?

(Seeing and Strongly Seeing)

  • An event xx can see event yy if yy is an ancestor of xx, and the ancestos of xx do not include a fork by the creator of yy

Imagine if I had the patience to make an interactive graph for this. It would be so sparkly. Instead you must use your imagination.

  • a4\color{blue}a_4 can see b1\color{purple}b_1 by way of e.g. a4a3a2b1\color{blue}a_4\color{black} \rightarrow \color{blue}a_3\color{black} \rightarrow \color{blue}a_2\color{black} \rightarrow \color{purple}b_1\color{black}
  • a4\color{blue}a_4 can see c1\color{olive}c_1 by way of e.g. a4b2c2c1\color{blue}a_4\color{black} \rightarrow \color{purple}b_2\color{black} \rightarrow \color{olive}c_2\color{black} \rightarrow \color{olive}c_1\color{black}
  • a4\color{blue}a_4 can see d1\color{orange}d_1 by way of e.g. a4a3c3d2d1\color{blue}a_4\color{black} \rightarrow \color{blue}a_3\color{black} \rightarrow \color{olive}c_3\color{black} \rightarrow \color{orange}d_2\color{black} \rightarrow \color{orange}d_1\color{black}

Etc.


  • An event xx can srongly see event yy if xx can see yy and there is a set SS of events by more than 2/32/3 of the members such that xx can see every event in SS, and every event in SS can see yy

Given any two vertices xx and yy in the HashGraph, it can be immediately calculated10 whether xx can strongly see yy, which is defined to be true if they are connected by multiple directed paths passing through enough members. This concept allows the key lemma to be proved: that if Alice and Bob are both able to calculate Carol’s virtual vote on a given question, then Alice and Bob get the same answer. That lemma forms the foundation for the rest of the mathematical proof of Byzantine agreement with probability one.

If there are n>1n > 1 members, a witness event ww can strongly see an event xx, if ww can see more than 2n/32n/3 events by different members, each of which can see xx.

Consider the following state snapshot of a HashGraph:

b5\color{purple}b_5 has three paths that strongly see c1\color{olive}c_1 (grouped by which witnesses they pass through):

  • (B,C\color{purple}B, \color{olive}C): b5b4b3b2c1\color{purple}b_5\color{black} \rightarrow \color{purple}b_4\color{black} \rightarrow \color{purple}b_3\color{black} \rightarrow \color{purple}b_2\color{black} \rightarrow \color{olive}c_1\color{black}
  • (B,D,C\color{purple}B, \color{orange}D, \color{olive}C): b5b4c3d2c1\color{purple}b_5\color{black} \rightarrow \color{purple}b_4\color{black} \rightarrow \color{olive}c_3\color{black} \rightarrow \color{purplorange}d_2\color{black} \rightarrow \color{olive}c_1\color{black}
  • (B,E,C\color{purple}B, \color{red}E, \color{olive}C): b5e3e2c1\color{purple}b_5\color{black} \rightarrow \color{red}e_3\color{black} \rightarrow \color{red}e_2\color{black} \rightarrow \color{olive}c_1\color{black}

While none of the paths individually traverse a super majority of nodes, together they collectively go through four nodes B, C, D, and E, thus b5\color{purple}b_5 strongly sees c1\color{olive}c_1.


Zepplin V (The House That Ewald Built)

(Fairness)

The authors stipulate that tt should be difficult for a small group of attackers to unfairly influence the order of transactions chosen as the consensus, and that the order of transactions needs to be preserved.

For some applications, the exact order does not matter, but for a stock market it can be critically important that this decision be made fairly.11

As noted, an inherent flaw with the leader-reliant algorithms such as those of Paxos or Raft is that –regardless of the means used to select a hopefully-truthy leader– that component becomes a single point of failure, and therefore an easy target.12

In any case, the leader could arbitrarily decide to ignore Alice or Bob's reported transactions for a period of time, delaying one of them, to force one transaction to come after another. If the gial is distributed trust, then no single individual can be trusted

Additionally, given the nature of consensus needing to be reached by at least 2n/32n/3 witnesses in a round, the structure of the HashGraph acts as a deterrent to attacks predicated on slowing the network. Failures resulting from (real) "total control" over the network are not considered as flaws of the protocol:

The Byzantine proofs assume the attackers control the internet, and can delay arbitrary messages. If attackers actually had that power, they could simply disconnect Alice from the internet for as long as it takes for Bob to send a transaction and have it recorded. This could be done on the real internet by launching a denial of service attack, flooding every computer with packets from Bob in an attempt to prevent Alice from communicating. Of course, this would also be effective if Alice were communicating with a central server, so it could be considered more a failure of the internet than a failure of the consensus system.

(Amateur Cartography)

Fastness (A HashGraph has 99 pace)

Throughout the paper, the authors provide ample discussion for means of compression or further reducing the overhead of the HashGraph without ever compromising the correctness of the structure on the whole.

The concern for overhead is almost humorously gratuitous. The priority for performative resolution of Byzantine problems is highlighted by the discussion of packing witnesses and visibility vectors into a single integer for optimal ALU consumption.

Naruto Themed Sexting

(Example of How it Works)

First, we start with the initial states for a HashGraph with four members. Each node has it's own root event:

Suppose AA receives an event c1c_1 from CC and updates itself:

Immediately following receipt of c1c_1, AA creates its own event a2a_2 which fathers a1,c1a_1, c_1.

Lemma: "fathering"

QED


Imagine that, in the meantine, CC also receives event d1d_1 from DD. CC will sign a new transaction indicating receipt of d1d_1 and is ready to offer their new gossip to other members.

Eventually, all nodes receive some events from one another such that the overall state of each member resembles the following:

As the gossip about the gossip unfolds, each member's local copies of the HashGraph may not be identical. However, all events observed by a member will be recorded in their own copy.

A Random Exercise in Impermanence (The Collector)

(Exchange of Information)

Recall that whenever a member broadcasts a synchronization update, the only information that need be exchanged (ignoring the accompanying signatures because the authors have thoroughly convinced the reader that you can pretty much send the size of that siggy to 0) is the delta between what member AA knows and what that member knows another member BB does not know.

Consider the final state of the previous example:

AA's next broadcast to BB will only include the new information that AA has and AA's internal representation of BB does not, and vice versa. The actual delivery and receipt of the state diffs of different members of the hashgraph is as follows.

So that the resulting states of the involved members are:

Bono!! Bono!!

(Conclusion)

So, that's that's pretty much it? Ended up with far more subsections than songs in the playlist, so forgive the deviation from form (some song titles are in fact relevant smh). What're you gonna do, vote about it?.

c u in da ballpit

(Swirld's Functional Appendix)

term/function definition description
ee ={p,h,t,i,s}= \{p, h, t, i, s\} an event tuple
pp =payload(e)= payload(e) the "payload" data, e.g. a list of transactions
hh =hashes(e)= hashes(e) list of hashes of the event's parents, starting with the self-parent
tt =time(e)= time(e) creator's claimed datetime of the event's creation
ii =creator(e)= creator(e) creator's ID
ss =signature(e)= signature(e) creator's digital signature of {p,h,t,i}\{p, h, t, i\}
nn the number of members in the HashGraph
cc the frequency of coin rounds
dd rounds delayed before start of delection
EE the set of all events in the HashGraph
E0E_0 =E{}= E \cup \{\varnothing\} EE joined with the null message marker
T\mathbb{T} the set of all possible (time,date)(time, date) pairs
B\mathbb{B} {True,False}\{True, False\}
N\mathbb{N} the big natural: 1,2,3,...1, 2, 3, ...
parents(x):EE2parents(x) : E \rarr E^2 set of events that are parents of xx
selfParent(x):EE0selfParent(x) : E \rarr E_0 the self-parent of event xx, or \varnothing if none
ancestor(x,y):E×EBancestor(x, y) : E \times E \rarr \mathbb{B} =x=yzparents(x),ancestor(z,y)= x = y \lor \exists z \in parents(x), ancestor(z,y) true if xx can reach yy by following 00 or more parent edges
selfAncestor(x,y):E×EBselfAncestor(x, y) : E \times E \rarr \mathbb{B} =x=y(selfParent(x)selfAncestor(selfParent(x),y))= x = y \lor (selfParent(x) \neq \varnothing \land selfAncestor(selfParent(x), y)) true if xx can reach yy by following 00 or more selfParent edges
manyCreators(S):2EBmanyCreators(S) : 2^E \rarr \mathbb{B} =S>2n/3x,yS,(xy    creator(x)creator(y))=\vert S \vert > 2n/3 \land \forall x,y \in S, (x \neq y \implies creator(x) \neq creator(y)) true if the set of events SS has more than 2n/32n/3 events, and all have distincy creators
see(x,y):E×EBsee(x, y) : E \times E \rarr \mathbb{B} =ancestor(x,y)¬(a,bE,creator(y)=creator(a)=creator(b)ancestor(x,a)ancestor(x,b)¬selfAncestor(a,b)¬selfAncestor(b,a))=ancestor(x,y) \land \\ \lnot (\exists a,b \in E, creator(y) = creator(a) = creator(b) \land \\\quad ancestor(x, a) \land \\\quad ancestor(x,b) \land \\ \quad \lnot selfAncestor(a,b) \land \\\quad \lnot selfAncestor(b,a)) true if yy is an ancestor of xx, but no fork of yy is an ancestor of xx
stronglySee(x,y):E×EBstronglySee(x, y) : E \times E \rarr \mathbb{B} =see(x,y)(SE,manyCreators(S)(zS    (see(x,z)see(z,y)))=see(x,y) \land \\ (\exists S \subseteq E, manyCreators(S) \\\quad \land (z \in S \implies (see(x,z) \land see(z,y))) true if xx can see events by more than 2n/32n/3 creators, each of which sees y
selfParentRound(x):ENselfParentRound(x) : E \rarr \mathbb{N} ={1if selfParent(x)=round(selfParent(x))o.w.= \begin{cases} 1 &\text{if } selfParent(x) = \varnothing \\ round(selfParent(x)) &\text{o.w.} \end{cases} the maximum created round of all parents of xx ? 1 if none
round(x):ENround(x) : E \rarr \mathbb{N} =max({selfParentRound(x)}{r+1SE,manyCreators(S)(yS,round(y)=rstronglySee(x,y))})= \max (\{selfParentRound(x)\} \cup \{r+1 \vert \exists S \subseteq E, manyCreators(S) \\ \quad \land (\forall y \in S, round(y) = r \land stronglySee(x,y))\}) the created round of xx
witness(x):EBwitness(x) : E \rarr \mathbb{B} =(selfParent(x)=(round(x)>round(selfParent(x))=(selfParent(x) = \varnothing \lor (round(x) > round(selfParent(x)) true if xx has a greater created round than its selfParent
diff(x,y):EIdiff(x, y) : E \rarr \mathbb{I} round(x)round(y)round(x) - round(y)
votes(x,y,v):E×E×BNvotes(x, y, v) : E \times E \times \mathbb{B} \rarr \mathbb{N} ={zEdiff(x,z)=1witness(z)stronglySee(x,z)vote(x,y)=v}= \vert \{ z \in E \vert diff(x,z) = 1 \land \\ witness(z) \land \\ stronglySee(x,z) \land \\ vote(x,y) = v \} \vert the number of votes equal to vv about the fame of witness yy collected by witness xx from the previous round
fractTrue(x,y):E×ERfractTrue(x, y) : E \times E \rarr \mathbb{R} =vote(x,y,true)max(1,votes(x,y,true)+votes(x,y,false))=\frac{vote(x,y, true)}{\max (1, votes(x,y,true) + votes(x,y, false))} fraction of "ye" votes, regarding the fame of witness yy, collected by witness x$, from witnesses in the previous round
decide(x,y):E×EBdecide(x, y) : E \times E \rarr \mathbb{B} =(selfParent(x)decide(selfParent(x),y))(witness(x)witness(y)diff(x,y)>d(diff(x,y)modc>0)(vB,votes(x,y,v)>2n/3)))=(selfParent(x) \neq \varnothing \land \\ decide(selfParent(x), y)) \lor (witness(x) \land \\ witness(y) \land \\ diff(x,y) > d \land \\ (diff(x,y) \mod c > 0) \land \\ (\exists v \in B, votes (x,y,v) > 2n/3))) true if xx (or its self ancestor) "decided" for the elction for witness yy (and therefore that member will never change its vote about yy again)
copyVote(x,y):E×EBcopyVote(x, y) : E \times E \rarr \mathbb{B} =(¬witness(x))(selfParent(x)decide(selfParent(x),y))=(\lnot witness(x)) \lor (selfParent(x) \neq \varnothing \land decide(selfParent(x), y)) true if xx should simply copy its selfParent's vote about the fame of witness yy (or xx is not a witness, or has already decided earlier)
vote(x,y):E×ERvote(x, y) : E \times E \rarr \mathbb{R} ={vote(selfParent(x),y)if copyVote(x,y)see(x,y)if ¬copyVote(x,y)diff(x,y)=d1=middleBit(signature(x))if ¬copyVote(x,y)diff(x,y)d(diff(x,y)modc=0)(13fractTrue(x,y)23)fractTrue(x,y)1/2o.w.=\begin{cases} vote(selfParent(x), y) &\text{if } copyVote(x,y) \\ see(x, y) &\text{if } \lnot copyVote(x,y) \land diff(x,y) = d \\ 1 = middleBit(signature(x)) &\text{if } \lnot copyVote(x,y) \land diff(x,y) \neq d \\ &\quad \land (diff(x,y) \mod c = 0) \\ &\quad \land (\frac{1}{3} \leq fractTrue(x,y) \leq \frac{2}{3}) \\ fractTrue(x,y) \geq 1/2 &\text{o.w.} \end{cases} the vote by witnes xx about the fame of witness yy (true for famous, false for not)
famous(x,y):EBfamous(x, y) : E \rarr \mathbb{B} =yE,decide((y,x)vote(y,x)= \exists y \in E, decide((y,x) \land vote(y,x) true if xx is famous (i.e., has had its fame decided by someone, and their vote was true)
uniqueFamous(x,y):EBuniqueFamous(x, y) : E \rarr \mathbb{B} =famous(x)¬yE,yxfamous(y)round(x)=round(y)creator(x)=creator(y)=famous(x) \land \lnot \exists y \in E, y \neq x \land \\ famous(y) \land round(x) = round(y) \land \\ creator(x) = creator(y) true if xx is famous and is the only famous witness in that round by that creator
roundsDecided(x,y):NBroundsDecided(x, y) : \mathbb{N} \rarr \mathbb{B} =xE,((round(x)rwitness(x))    yE,decide(y,x))= \forall x \in E, ((round(x) \leq r \land witness(x)) \\ \implies \exists y \in E, decide(y,x)) true if all known witnesses had their fame decided, for both round rr and all earlier rounds
roundReceived(x,y):ENroundReceived(x, y) : E \rarr \mathbb{N} =min({rNroundsDecided(r)(yE,(round(y)=runiqueFamous(y))    ancestor(y,x))})=\min (\{r \in \mathbb N \vert roundsDecided(r) \land \\ (\forall y \in E, (round(y) = r \land uniqueFamous(y)) \\ \implies ancestor(y,x)) \}) the round received for event xx
timeReceived(x,y):ETtimeReceived(x, y) : E \rarr \mathbb{T} =median({time(y)yancestor(y,x)(zE,round(z)=roundReceived(x)unqieuFamous(z)selfAncestor(z,y))¬(wE,selfAncestor(y,w)ancestor(w,x))})= median( \{ time(y) \vert y \in ancestor(y,x) \land \\ (\exists z \in E, round(z) =roundReceived(x) \land \\ unqieuFamous(z) \land \\ selfAncestor(z,y)) \land \\ \lnot (\exists w \in E, selfAncestor(y,w) \land \\ ancestor(w,x)) \}) the consensus timestamp for event xx

Kawasaki Backflip


  1. please please please someone start a blogging flame war with me.

  2. "Why Computers Can't Count Sometimes." The King, Tom Scott.

  3. Acceptable pronounciations include "credit", "Crudités", and "Kurds"

  4. Shapiro et al. "A comprehensive study of Convergent and Commutative Replicated Data Types." 2011.

  5. A join-semilattice is just a set that implements Comparable with a total ordering: x,yS  z\forall x,y \in S \; \exists z, where zz is the greatest lower bound of {x,y}\{x, y\}

  6. Fischer, Lynch, Paterson. "Impossibility of Distributed Consensus with One Faulty Process." Journal Association for Computing Machinery, 1985.

  7. The size is entirely arbitrary. In the original paper, the authors use the simplest case of 1-bit to underscore the fragility of systems on the whole.

  8. Again, arbitrary domain, just needs to match the above.

  9. Baird, Leemon. "THE SWIRLDS HASHGRAPH CONSENSUS ALGORITHM: FAIR, FAST, BYZANTINE FAULT TOLERANCE." Swirlds. 2016.

  10. Though it seems like this would actually an O(k/n)O(k/n) computation, where kk is the total number of transactions, as xx needs to traverse the HashGraph to find yy, the authors point out that for the purpose of sight (and other properties used in the proofs of tolerance) the number of temporally-prior nodes that any given node xx needs to view in order to "see" another node yy is actually negligibly small in terms of complexity. This sight need only look as far as the last Famous Witness which indicates that all transactions from rounds prior to that famous witness are agreed upon by the system on the whole. Therefore, the search space for xyx \rarr y is bounded by 1 round.

  11. Flash Boys by Michael Lewis is a great book about why even milliseconds of either innacuracy or advantage in an exchange market can be hugely exploited by high frequency traders.

  12. It's worth noting that this point of contention is largely a theoretical one. Leader/follower-based consensus protocols are still prevalent in the industry and you don't see the world going up in flames about it.