aboutsummaryrefslogtreecommitdiff
path: root/src/table
diff options
context:
space:
mode:
Diffstat (limited to 'src/table')
-rw-r--r--src/table/crdt/lww.rs2
-rw-r--r--src/table/crdt/map.rs1
-rw-r--r--src/table/gc.rs2
-rw-r--r--src/table/lib.rs6
-rw-r--r--src/table/replication/fullcopy.rs12
-rw-r--r--src/table/replication/mod.rs6
-rw-r--r--src/table/replication/parameters.rs9
-rw-r--r--src/table/replication/sharded.rs17
-rw-r--r--src/table/schema.rs17
9 files changed, 47 insertions, 25 deletions
diff --git a/src/table/crdt/lww.rs b/src/table/crdt/lww.rs
index 25ecdb07..3b1b2406 100644
--- a/src/table/crdt/lww.rs
+++ b/src/table/crdt/lww.rs
@@ -34,7 +34,7 @@ use crate::crdt::crdt::*;
/// and may differ from what you observed with your atomic clock!
///
/// This scheme is used by AWS S3 or Soundcloud and often without knowing
-/// in entreprise when reconciliating databases with ad-hoc scripts.
+/// in enterprise when reconciliating databases with ad-hoc scripts.
#[derive(Clone, Debug, Serialize, Deserialize, PartialEq)]
pub struct LWW<T> {
ts: u64,
diff --git a/src/table/crdt/map.rs b/src/table/crdt/map.rs
index 1193e6db..c4a30a26 100644
--- a/src/table/crdt/map.rs
+++ b/src/table/crdt/map.rs
@@ -37,6 +37,7 @@ where
Self { vals: vec![(k, v)] }
}
+ /// Add a value to the map
pub fn put(&mut self, k: K, v: V) {
self.merge(&Self::put_mutator(k, v));
}
diff --git a/src/table/gc.rs b/src/table/gc.rs
index e52bf599..694a3789 100644
--- a/src/table/gc.rs
+++ b/src/table/gc.rs
@@ -74,7 +74,7 @@ where
while !*must_exit.borrow() {
match self.gc_loop_iter().await {
Ok(true) => {
- // Stuff was done, loop imediately
+ // Stuff was done, loop immediately
continue;
}
Ok(false) => {
diff --git a/src/table/lib.rs b/src/table/lib.rs
index 3b73163b..c3e14ab8 100644
--- a/src/table/lib.rs
+++ b/src/table/lib.rs
@@ -8,10 +8,10 @@ pub mod schema;
pub mod util;
pub mod data;
-pub mod gc;
-pub mod merkle;
+mod gc;
+mod merkle;
pub mod replication;
-pub mod sync;
+mod sync;
pub mod table;
pub use schema::*;
diff --git a/src/table/replication/fullcopy.rs b/src/table/replication/fullcopy.rs
index bd658f63..a6b4c98c 100644
--- a/src/table/replication/fullcopy.rs
+++ b/src/table/replication/fullcopy.rs
@@ -6,19 +6,19 @@ use garage_util::data::*;
use crate::replication::*;
+/// Full replication schema: all nodes store everything
+/// Writes are disseminated in an epidemic manner in the network
+/// Advantage: do all reads locally, extremely fast
+/// Inconvenient: only suitable to reasonably small tables
#[derive(Clone)]
pub struct TableFullReplication {
+ /// The membership manager of this node
pub system: Arc<System>,
+ /// Max number of faults allowed while replicating a record
pub max_faults: usize,
}
impl TableReplication for TableFullReplication {
- // Full replication schema: all nodes store everything
- // Writes are disseminated in an epidemic manner in the network
-
- // Advantage: do all reads locally, extremely fast
- // Inconvenient: only suitable to reasonably small tables
-
fn read_nodes(&self, _hash: &Hash) -> Vec<UUID> {
vec![self.system.id]
}
diff --git a/src/table/replication/mod.rs b/src/table/replication/mod.rs
index d43d7f19..dfcb026a 100644
--- a/src/table/replication/mod.rs
+++ b/src/table/replication/mod.rs
@@ -1,6 +1,8 @@
mod parameters;
-pub mod fullcopy;
-pub mod sharded;
+mod fullcopy;
+mod sharded;
+pub use fullcopy::TableFullReplication;
pub use parameters::*;
+pub use sharded::TableShardedReplication;
diff --git a/src/table/replication/parameters.rs b/src/table/replication/parameters.rs
index e46bd172..c2c78c8b 100644
--- a/src/table/replication/parameters.rs
+++ b/src/table/replication/parameters.rs
@@ -2,20 +2,25 @@ use garage_rpc::ring::*;
use garage_util::data::*;
+/// Trait to describe how a table shall be replicated
pub trait TableReplication: Send + Sync {
// See examples in table_sharded.rs and table_fullcopy.rs
// To understand various replication methods
- // Which nodes to send reads from
+ /// Which nodes to send read requests to
fn read_nodes(&self, hash: &Hash) -> Vec<UUID>;
+ /// Responses needed to consider a read succesfull
fn read_quorum(&self) -> usize;
- // Which nodes to send writes to
+ /// Which nodes to send writes to
fn write_nodes(&self, hash: &Hash) -> Vec<UUID>;
+ /// Responses needed to consider a write succesfull
fn write_quorum(&self) -> usize;
fn max_write_errors(&self) -> usize;
// Accessing partitions, for Merkle tree & sync
+ /// Get partition for data with given hash
fn partition_of(&self, hash: &Hash) -> Partition;
+ /// List of existing partitions
fn partitions(&self) -> Vec<(Partition, Hash)>;
}
diff --git a/src/table/replication/sharded.rs b/src/table/replication/sharded.rs
index dce74b03..f2d89729 100644
--- a/src/table/replication/sharded.rs
+++ b/src/table/replication/sharded.rs
@@ -6,22 +6,25 @@ use garage_util::data::*;
use crate::replication::*;
+/// Sharded replication schema:
+/// - based on the ring of nodes, a certain set of neighbors
+/// store entries, given as a function of the position of the
+/// entry's hash in the ring
+/// - reads are done on all of the nodes that replicate the data
+/// - writes as well
#[derive(Clone)]
pub struct TableShardedReplication {
+ /// The membership manager of this node
pub system: Arc<System>,
+ /// How many time each data should be replicated
pub replication_factor: usize,
+ /// How many nodes to contact for a read, should be at most `replication_factor`
pub read_quorum: usize,
+ /// How many nodes to contact for a write, should be at most `replication_factor`
pub write_quorum: usize,
}
impl TableReplication for TableShardedReplication {
- // Sharded replication schema:
- // - based on the ring of nodes, a certain set of neighbors
- // store entries, given as a function of the position of the
- // entry's hash in the ring
- // - reads are done on all of the nodes that replicate the data
- // - writes as well
-
fn read_nodes(&self, hash: &Hash) -> Vec<UUID> {
let ring = self.system.ring.borrow().clone();
ring.walk_ring(&hash, self.replication_factor)
diff --git a/src/table/schema.rs b/src/table/schema.rs
index 4d754664..13517271 100644
--- a/src/table/schema.rs
+++ b/src/table/schema.rs
@@ -4,7 +4,9 @@ use garage_util::data::*;
use crate::crdt::CRDT;
+/// Trait for field used to partition data
pub trait PartitionKey {
+ /// Get the key used to partition
fn hash(&self) -> Hash;
}
@@ -20,7 +22,9 @@ impl PartitionKey for Hash {
}
}
+/// Trait for field used to sort data
pub trait SortKey {
+ /// Get the key used to sort
fn sort_key(&self) -> &[u8];
}
@@ -36,25 +40,34 @@ impl SortKey for Hash {
}
}
+/// Trait for an entry in a table. It must be sortable and partitionnable.
pub trait Entry<P: PartitionKey, S: SortKey>:
CRDT + PartialEq + Clone + Serialize + for<'de> Deserialize<'de> + Send + Sync
{
+ /// Get the key used to partition
fn partition_key(&self) -> &P;
+ /// Get the key used to sort
fn sort_key(&self) -> &S;
+ /// Is the entry a tombstone? Default implementation always return false
fn is_tombstone(&self) -> bool {
false
}
}
+/// Trait for the schema used in a table
pub trait TableSchema: Send + Sync {
+ /// The partition key used in that table
type P: PartitionKey + Clone + PartialEq + Serialize + for<'de> Deserialize<'de> + Send + Sync;
+ /// The sort key used int that table
type S: SortKey + Clone + Serialize + for<'de> Deserialize<'de> + Send + Sync;
+ /// They type for an entry in that table
type E: Entry<Self::P, Self::S>;
type Filter: Clone + Serialize + for<'de> Deserialize<'de> + Send + Sync;
// Action to take if not able to decode current version:
// try loading from an older version
+ /// Try migrating an entry from an older version
fn try_migrate(_bytes: &[u8]) -> Option<Self::E> {
None
}
@@ -65,7 +78,5 @@ pub trait TableSchema: Send + Sync {
// to stderr.
fn updated(&self, _old: Option<Self::E>, _new: Option<Self::E>) {}
- fn matches_filter(_entry: &Self::E, _filter: &Self::Filter) -> bool {
- true
- }
+ fn matches_filter(entry: &Self::E, filter: &Self::Filter) -> bool;
}