aboutsummaryrefslogblamecommitdiff
path: root/src/table/replication/fullcopy.rs
blob: cb5471afd0906323190733186aee2edda83abf16 (plain) (tree)
1
2
3
4
5
6
7
8
9
                            

                   
                          
                               

                         
                          
 






                                                                           
                                                       

                                                          
                                                                    

                                 
                                               
                                
                                                                   
                              

 
                                                




                                                            
                                                         
                                    




                                        

                                                             
         
                                         
                                                                                        




                                                  

                                             
                               

         


                                                           









                                                              
                                                                                                     

                           

         
use std::iter::FromIterator;
use std::sync::Arc;

use garage_rpc::layout::*;
use garage_rpc::system::System;
use garage_util::data::*;

use crate::replication::*;

// TODO: find a way to track layout changes for this as well
// The hard thing is that this data is stored also on gateway nodes,
// whereas sharded data is stored only on non-Gateway nodes (storage nodes)
// Also we want to be more tolerant to failures of gateways so we don't
// want to do too much holding back of data when progress of gateway
// nodes is not reported in the layout history's ack/sync/sync_ack maps.

/// Full replication schema: all nodes store everything
/// Advantage: do all reads locally, extremely fast
/// Inconvenient: only suitable to reasonably small tables
/// Inconvenient: if some writes fail, nodes will read outdated data
#[derive(Clone)]
pub struct TableFullReplication {
	/// The membership manager of this node
	pub system: Arc<System>,
	/// Max number of faults allowed while replicating a record
	pub max_faults: usize,
}

impl TableReplication for TableFullReplication {
	fn storage_nodes(&self, _hash: &Hash) -> Vec<Uuid> {
		let layout = self.system.cluster_layout();
		layout.current().all_nodes().to_vec()
	}

	fn read_nodes(&self, _hash: &Hash) -> Vec<Uuid> {
		vec![self.system.id]
	}
	fn read_quorum(&self) -> usize {
		1
	}

	fn write_sets(&self, hash: &Hash) -> Vec<Vec<Uuid>> {
		vec![self.storage_nodes(hash)]
	}
	fn write_quorum(&self) -> usize {
		let nmembers = self.system.cluster_layout().current().all_nodes().len();
		if nmembers > self.max_faults {
			nmembers - self.max_faults
		} else {
			1
		}
	}
	fn max_write_errors(&self) -> usize {
		self.max_faults
	}

	fn partition_of(&self, _hash: &Hash) -> Partition {
		0u16
	}

	fn sync_partitions(&self) -> SyncPartitions {
		let layout = self.system.cluster_layout();
		let layout_version = layout.current().version;
		SyncPartitions {
			layout_version,
			partitions: vec![SyncPartition {
				partition: 0u16,
				first_hash: [0u8; 32].into(),
				last_hash: [0xff; 32].into(),
				storage_nodes: Vec::from_iter(layout.current().all_nodes().to_vec()),
			}],
		}
	}
}