Skip to content

Latest commit

 

History

History
101 lines (76 loc) · 3.85 KB

Readme.md

File metadata and controls

101 lines (76 loc) · 3.85 KB

Distributed Key-Value store

This is a distributed Key-Value store with basic functionality to better understand how distributed databases work.

I was reading Designing Data-Intensive applications, and in order to better understand some of the concepts described in the book, I decided to build a distributed database with replication and master election.

A single master will handle writes and replicate them to n slave nodes, that can handle reads (the master can also handle reads).

It is far from perfect and definitely not intended for production usage. Interesting evolutions would be towards reducing inconsistencies, probably following the path of eventual consistency.

Functionality

All nodes, slave or master:

  • READ a value
  • maintain a list of all available nodes

slave:

  • check the master node's health
  • elect a new master (while agreeing with other nodes)
  • be promoted to master
  • join a master:
    • get an ID and a list of available nodes from the master
    • update asynchronously until fully up to date with master

master

  • check health of the slave nodes
  • WRITE value (new or existing)
  • push a WRITE query to all slaves
  • push an update of the list of nodes when a node joins or leaves

OUT OF SCOPE:

  • database client (but the HTTP API is available)
  • proxy/load balancer to redirect queries to the correct nodes (random or closest node for reads, master for writes) - this could be included in the client, in the nodes or as an external tool (service discovery would be a good option)
  • security and authentication
  • consistency: there are a few cases where data can vary between nodes, and I have not implemented any fix for these cases. Possible solutions:
    • store a transaction log and replay the logs on slaves in the same order
    • periodically check for discrepancies, apply value from master

Design choices

Single write master, many read slaves - no need for a conflict resolver, still increases availability Automatic failover by electing a new master when the old one stops responding for 5 sec (low value used for testing). All un-replicated writes will be lost.

Asynchronous replication on new writes - this could cause inconsistencies if writes do not get handled in the same order in every node: two successive writes on the same key, played on different orders on two nodes, would result in different values in these nodes.

Async replication when a node joins : all the data is copied from the master to the joining node by dribs and drabs. This could cause inconsistencies if the data gets updated while replicating.

Abstract the Storage and Transport layers so it can use different implementations

Basic HTTP queries rather than gRPC + protobuf because I didn't want to include any dependency.

Dependencies:

None

Improvements/To-do list

  • check that queries (i.e. write queries, list update queries) come from master and not another slave
  • persist data to disk
  • one of the following:
    • regularly check for entropy and fix errors
    • implement a transaction log or another way to order all the queries
  • dockerize for easier deployment
  • abstract making HTTP queries because it takes 20 lines to make a POST query
  • better ID generation

References: