use serde::{Deserialize, Serialize}; use std::collections::BTreeMap; use std::sync::Arc; use garage_db as db; use garage_util::background::BackgroundRunner; use garage_util::data::*; use garage_table::crdt::*; use garage_table::replication::TableShardedReplication; use garage_table::*; use crate::index_counter::*; use crate::s3::version_table::*; use garage_model_050::object_table as old; pub const OBJECTS: &str = "objects"; pub const UNFINISHED_UPLOADS: &str = "unfinished_uploads"; pub const BYTES: &str = "bytes"; /// An object #[derive(PartialEq, Clone, Debug, Serialize, Deserialize)] pub struct Object { /// The bucket in which the object is stored, used as partition key pub bucket_id: Uuid, /// The key at which the object is stored in its bucket, used as sorting key pub key: String, /// The list of currenty stored versions of the object versions: Vec, } impl Object { /// Initialize an Object struct from parts pub fn new(bucket_id: Uuid, key: String, versions: Vec) -> Self { let mut ret = Self { bucket_id, key, versions: vec![], }; for v in versions { ret.add_version(v) .expect("Twice the same ObjectVersion in Object constructor"); } ret } /// Adds a version if it wasn't already present #[allow(clippy::result_unit_err)] pub fn add_version(&mut self, new: ObjectVersion) -> Result<(), ()> { match self .versions .binary_search_by(|v| v.cmp_key().cmp(&new.cmp_key())) { Err(i) => { self.versions.insert(i, new); Ok(()) } Ok(_) => Err(()), } } /// Get a list of currently stored versions of `Object` pub fn versions(&self) -> &[ObjectVersion] { &self.versions[..] } } /// Informations about a version of an object #[derive(PartialEq, Clone, Debug, Serialize, Deserialize)] pub struct ObjectVersion { /// Id of the version pub uuid: Uuid, /// Timestamp of when the object was created pub timestamp: u64, /// State of the version pub state: ObjectVersionState, } /// State of an object version #[derive(PartialEq, Clone, Debug, Serialize, Deserialize)] pub enum ObjectVersionState { /// The version is being received Uploading(ObjectVersionHeaders), /// The version is fully received Complete(ObjectVersionData), /// The version uploaded containded errors or the upload was explicitly aborted Aborted, } impl Crdt for ObjectVersionState { fn merge(&mut self, other: &Self) { use ObjectVersionState::*; match other { Aborted => { *self = Aborted; } Complete(b) => match self { Aborted => {} Complete(a) => { a.merge(b); } Uploading(_) => { *self = Complete(b.clone()); } }, Uploading(_) => {} } } } /// Data stored in object version #[derive(PartialEq, Eq, PartialOrd, Ord, Clone, Debug, Serialize, Deserialize)] pub enum ObjectVersionData { /// The object was deleted, this Version is a tombstone to mark it as such DeleteMarker, /// The object is short, it's stored inlined Inline(ObjectVersionMeta, #[serde(with = "serde_bytes")] Vec), /// The object is not short, Hash of first block is stored here, next segments hashes are /// stored in the version table FirstBlock(ObjectVersionMeta, Hash), } impl AutoCrdt for ObjectVersionData { const WARN_IF_DIFFERENT: bool = true; } /// Metadata about the object version #[derive(PartialEq, Eq, PartialOrd, Ord, Clone, Debug, Serialize, Deserialize)] pub struct ObjectVersionMeta { /// Headers to send to the client pub headers: ObjectVersionHeaders, /// Size of the object pub size: u64, /// etag of the object pub etag: String, } /// Additional headers for an object #[derive(PartialEq, Eq, PartialOrd, Ord, Clone, Debug, Serialize, Deserialize)] pub struct ObjectVersionHeaders { /// Content type of the object pub content_type: String, /// Any other http headers to send pub other: BTreeMap, } impl ObjectVersion { fn cmp_key(&self) -> (u64, Uuid) { (self.timestamp, self.uuid) } /// Is the object version currently being uploaded pub fn is_uploading(&self) -> bool { matches!(self.state, ObjectVersionState::Uploading(_)) } /// Is the object version completely received pub fn is_complete(&self) -> bool { matches!(self.state, ObjectVersionState::Complete(_)) } /// Is the object version available (received and not a tombstone) pub fn is_data(&self) -> bool { match self.state { ObjectVersionState::Complete(ObjectVersionData::DeleteMarker) => false, ObjectVersionState::Complete(_) => true, _ => false, } } } impl Entry for Object { fn partition_key(&self) -> &Uuid { &self.bucket_id } fn sort_key(&self) -> &String { &self.key } fn is_tombstone(&self) -> bool { self.versions.len() == 1 && self.versions[0].state == ObjectVersionState::Complete(ObjectVersionData::DeleteMarker) } } impl Crdt for Object { fn merge(&mut self, other: &Self) { // Merge versions from other into here for other_v in other.versions.iter() { match self .versions .binary_search_by(|v| v.cmp_key().cmp(&other_v.cmp_key())) { Ok(i) => { self.versions[i].state.merge(&other_v.state); } Err(i) => { self.versions.insert(i, other_v.clone()); } } } // Remove versions which are obsolete, i.e. those that come // before the last version which .is_complete(). let last_complete = self .versions .iter() .enumerate() .rev() .find(|(_, v)| v.is_complete()) .map(|(vi, _)| vi); if let Some(last_vi) = last_complete { self.versions = self.versions.drain(last_vi..).collect::>(); } } } pub struct ObjectTable { pub background: Arc, pub version_table: Arc>, pub object_counter_table: Arc>, } #[derive(Clone, Copy, Debug, Serialize, Deserialize)] pub enum ObjectFilter { IsData, IsUploading, } impl TableSchema for ObjectTable { const TABLE_NAME: &'static str = "object"; type P = Uuid; type S = String; type E = Object; type Filter = ObjectFilter; fn updated( &self, tx: &mut db::Transaction, old: Option<&Self::E>, new: Option<&Self::E>, ) -> db::TxOpResult<()> { // 1. Count let counter_res = self.object_counter_table.count(tx, old, new); if let Err(e) = db::unabort(counter_res)? { error!( "Unable to update object counter: {}. Index values will be wrong!", e ); } // 2. Spawn threads that propagates deletions to version table let version_table = self.version_table.clone(); let old = old.cloned(); let new = new.cloned(); self.background.spawn(async move { if let (Some(old_v), Some(new_v)) = (old, new) { // Propagate deletion of old versions for v in old_v.versions.iter() { let newly_deleted = match new_v .versions .binary_search_by(|nv| nv.cmp_key().cmp(&v.cmp_key())) { Err(_) => true, Ok(i) => { new_v.versions[i].state == ObjectVersionState::Aborted && v.state != ObjectVersionState::Aborted } }; if newly_deleted { let deleted_version = Version::new(v.uuid, old_v.bucket_id, old_v.key.clone(), true); version_table.insert(&deleted_version).await?; } } } Ok(()) }); Ok(()) } fn matches_filter(entry: &Self::E, filter: &Self::Filter) -> bool { match filter { ObjectFilter::IsData => entry.versions.iter().any(|v| v.is_data()), ObjectFilter::IsUploading => entry.versions.iter().any(|v| v.is_uploading()), } } fn try_migrate(bytes: &[u8]) -> Option { let old_obj = rmp_serde::decode::from_read_ref::<_, old::Object>(bytes).ok()?; Some(migrate_object(old_obj)) } } impl CountedItem for Object { const COUNTER_TABLE_NAME: &'static str = "bucket_object_counter"; // Partition key = nothing type CP = EmptyKey; // Sort key = bucket id type CS = Uuid; fn counter_partition_key(&self) -> &EmptyKey { &EmptyKey } fn counter_sort_key(&self) -> &Uuid { &self.bucket_id } fn counts(&self) -> Vec<(&'static str, i64)> { let versions = self.versions(); let n_objects = if versions.iter().any(|v| v.is_data()) { 0 } else { 1 }; let n_unfinished_uploads = versions .iter() .filter(|v| matches!(v.state, ObjectVersionState::Uploading(_))) .count(); let n_bytes = versions .iter() .map(|v| match &v.state { ObjectVersionState::Complete(ObjectVersionData::Inline(meta, _)) | ObjectVersionState::Complete(ObjectVersionData::FirstBlock(meta, _)) => meta.size, _ => 0, }) .sum::(); vec![ (OBJECTS, n_objects), (UNFINISHED_UPLOADS, n_unfinished_uploads as i64), (BYTES, n_bytes as i64), ] } } // vvvvvvvv migration code, stupid stuff vvvvvvvvvvvv // (we just want to change bucket into bucket_id by hashing it) fn migrate_object(o: old::Object) -> Object { let versions = o .versions() .iter() .cloned() .map(migrate_object_version) .collect(); Object { bucket_id: blake2sum(o.bucket.as_bytes()), key: o.key, versions, } } fn migrate_object_version(v: old::ObjectVersion) -> ObjectVersion { ObjectVersion { uuid: Uuid::try_from(v.uuid.as_slice()).unwrap(), timestamp: v.timestamp, state: match v.state { old::ObjectVersionState::Uploading(h) => { ObjectVersionState::Uploading(migrate_object_version_headers(h)) } old::ObjectVersionState::Complete(d) => { ObjectVersionState::Complete(migrate_object_version_data(d)) } old::ObjectVersionState::Aborted => ObjectVersionState::Aborted, }, } } fn migrate_object_version_headers(h: old::ObjectVersionHeaders) -> ObjectVersionHeaders { ObjectVersionHeaders { content_type: h.content_type, other: h.other, } } fn migrate_object_version_data(d: old::ObjectVersionData) -> ObjectVersionData { match d { old::ObjectVersionData::DeleteMarker => ObjectVersionData::DeleteMarker, old::ObjectVersionData::Inline(m, b) => { ObjectVersionData::Inline(migrate_object_version_meta(m), b) } old::ObjectVersionData::FirstBlock(m, h) => ObjectVersionData::FirstBlock( migrate_object_version_meta(m), Hash::try_from(h.as_slice()).unwrap(), ), } } fn migrate_object_version_meta(m: old::ObjectVersionMeta) -> ObjectVersionMeta { ObjectVersionMeta { headers: migrate_object_version_headers(m.headers), size: m.size, etag: m.etag, } }