aboutsummaryrefslogtreecommitdiff
path: root/src
diff options
context:
space:
mode:
authorAlex Auvolat <alex@adnab.me>2024-02-20 11:35:18 +0100
committerAlex Auvolat <alex@adnab.me>2024-02-20 13:50:45 +0100
commit00d479358d31b445bfbe6d7ee3c37520be7e6d85 (patch)
tree8bd34b29b869467a39a30c1a2115523d7a4bce31 /src
parent0b9859befa80f0524da2c2ad121105e6263f8a33 (diff)
downloadgarage-00d479358d31b445bfbe6d7ee3c37520be7e6d85.tar.gz
garage-00d479358d31b445bfbe6d7ee3c37520be7e6d85.zip
[peer-metrics] refactor/simplify SystemMetrics
Diffstat (limited to 'src')
-rw-r--r--src/rpc/system.rs77
-rw-r--r--src/rpc/system_metrics.rs52
2 files changed, 41 insertions, 88 deletions
diff --git a/src/rpc/system.rs b/src/rpc/system.rs
index 147ec4d6..4a505f58 100644
--- a/src/rpc/system.rs
+++ b/src/rpc/system.rs
@@ -3,11 +3,9 @@ use std::collections::HashMap;
use std::io::{Read, Write};
use std::net::{IpAddr, SocketAddr};
use std::path::{Path, PathBuf};
-use std::sync::atomic::Ordering;
use std::sync::{Arc, RwLock};
use std::time::{Duration, Instant};
-use arc_swap::ArcSwap;
use async_trait::async_trait;
use futures::join;
use serde::{Deserialize, Serialize};
@@ -88,7 +86,7 @@ pub struct System {
persist_cluster_layout: Persister<ClusterLayout>,
persist_peer_list: Persister<PeerList>,
- local_status: ArcSwap<NodeStatus>,
+ local_status: Arc<RwLock<NodeStatus>>,
node_status: RwLock<HashMap<Uuid, (u64, NodeStatus)>>,
pub netapp: Arc<NetApp>,
@@ -106,7 +104,7 @@ pub struct System {
#[cfg(feature = "kubernetes-discovery")]
kubernetes_discovery: Option<KubernetesDiscoveryConfig>,
- metrics: SystemMetrics,
+ _metrics: SystemMetrics,
replication_mode: ReplicationMode,
replication_factor: usize,
@@ -280,10 +278,11 @@ impl System {
}
};
- let metrics = SystemMetrics::new(replication_factor);
-
let mut local_status = NodeStatus::initial(replication_factor, &cluster_layout);
- local_status.update_disk_usage(&config.metadata_dir, &config.data_dir, &metrics);
+ local_status.update_disk_usage(&config.metadata_dir, &config.data_dir);
+ let local_status = Arc::new(RwLock::new(local_status));
+
+ let metrics = SystemMetrics::new(replication_factor, local_status.clone());
let ring = Ring::new(cluster_layout, replication_factor);
let (update_ring, ring) = watch::channel(Arc::new(ring));
@@ -357,7 +356,7 @@ impl System {
id: netapp.id.into(),
persist_cluster_layout,
persist_peer_list,
- local_status: ArcSwap::new(Arc::new(local_status)),
+ local_status,
node_status: RwLock::new(HashMap::new()),
netapp: netapp.clone(),
peering: peering.clone(),
@@ -377,7 +376,7 @@ impl System {
consul_discovery,
#[cfg(feature = "kubernetes-discovery")]
kubernetes_discovery: config.kubernetes_discovery.clone(),
- metrics,
+ _metrics: metrics,
ring,
update_ring: Mutex::new(update_ring),
@@ -546,12 +545,9 @@ impl System {
}
};
+ let hostname = self.local_status.read().unwrap().hostname.clone();
if let Err(e) = c
- .publish_consul_service(
- self.netapp.id,
- &self.local_status.load_full().hostname,
- rpc_public_addr,
- )
+ .publish_consul_service(self.netapp.id, &hostname, rpc_public_addr)
.await
{
error!("Error while publishing Consul service: {}", e);
@@ -573,13 +569,8 @@ impl System {
}
};
- if let Err(e) = publish_kubernetes_node(
- k,
- self.netapp.id,
- &self.local_status.load_full().hostname,
- rpc_public_addr,
- )
- .await
+ let hostname = self.local_status.read().unwrap().hostname.clone();
+ if let Err(e) = publish_kubernetes_node(k, self.netapp.id, &hostname, rpc_public_addr).await
{
error!("Error while publishing node to Kubernetes: {}", e);
}
@@ -596,15 +587,13 @@ impl System {
}
fn update_local_status(&self) {
- let mut new_si: NodeStatus = self.local_status.load().as_ref().clone();
+ let mut local_status = self.local_status.write().unwrap();
let ring = self.ring.borrow();
- new_si.cluster_layout_version = ring.layout.version;
- new_si.cluster_layout_staging_hash = ring.layout.staging_hash;
-
- new_si.update_disk_usage(&self.metadata_dir, &self.data_dir, &self.metrics);
+ local_status.cluster_layout_version = ring.layout.version;
+ local_status.cluster_layout_staging_hash = ring.layout.staging_hash;
- self.local_status.swap(Arc::new(new_si));
+ local_status.update_disk_usage(&self.metadata_dir, &self.data_dir);
}
// --- RPC HANDLERS ---
@@ -629,7 +618,7 @@ impl System {
from: Uuid,
info: &NodeStatus,
) -> Result<SystemRpc, Error> {
- let local_info = self.local_status.load();
+ let local_info = self.local_status.read().unwrap();
if local_info.replication_factor < info.replication_factor {
error!("Some node have a higher replication factor ({}) than this one ({}). This is not supported and will lead to data corruption. Shutting down for safety.",
@@ -644,6 +633,8 @@ impl System {
tokio::spawn(self.clone().pull_cluster_layout(from));
}
+ drop(local_info);
+
self.node_status
.write()
.unwrap()
@@ -708,7 +699,7 @@ impl System {
let restart_at = Instant::now() + STATUS_EXCHANGE_INTERVAL;
self.update_local_status();
- let local_status: NodeStatus = self.local_status.load().as_ref().clone();
+ let local_status: NodeStatus = self.local_status.read().unwrap().clone();
let _ = self
.rpc
.broadcast(
@@ -893,12 +884,7 @@ impl NodeStatus {
}
}
- fn update_disk_usage(
- &mut self,
- meta_dir: &Path,
- data_dir: &DataDirEnum,
- metrics: &SystemMetrics,
- ) {
+ fn update_disk_usage(&mut self, meta_dir: &Path, data_dir: &DataDirEnum) {
use nix::sys::statvfs::statvfs;
let mount_avail = |path: &Path| match statvfs(path) {
Ok(x) => {
@@ -934,27 +920,6 @@ impl NodeStatus {
)
})(),
};
-
- if let Some((avail, total)) = self.meta_disk_avail {
- metrics
- .values
- .meta_disk_avail
- .store(avail, Ordering::Relaxed);
- metrics
- .values
- .meta_disk_total
- .store(total, Ordering::Relaxed);
- }
- if let Some((avail, total)) = self.data_disk_avail {
- metrics
- .values
- .data_disk_avail
- .store(avail, Ordering::Relaxed);
- metrics
- .values
- .data_disk_total
- .store(total, Ordering::Relaxed);
- }
}
}
diff --git a/src/rpc/system_metrics.rs b/src/rpc/system_metrics.rs
index af81b71f..7b390c25 100644
--- a/src/rpc/system_metrics.rs
+++ b/src/rpc/system_metrics.rs
@@ -1,31 +1,22 @@
-use std::sync::atomic::{AtomicU64, Ordering};
-use std::sync::Arc;
+use std::sync::{Arc, RwLock};
use opentelemetry::{global, metrics::*, KeyValue};
+use crate::system::NodeStatus;
+
/// TableMetrics reference all counter used for metrics
pub struct SystemMetrics {
pub(crate) _garage_build_info: ValueObserver<u64>,
pub(crate) _replication_factor: ValueObserver<u64>,
pub(crate) _disk_avail: ValueObserver<u64>,
pub(crate) _disk_total: ValueObserver<u64>,
- pub(crate) values: Arc<SystemMetricsValues>,
-}
-
-#[derive(Default)]
-pub struct SystemMetricsValues {
- pub(crate) data_disk_total: AtomicU64,
- pub(crate) data_disk_avail: AtomicU64,
- pub(crate) meta_disk_total: AtomicU64,
- pub(crate) meta_disk_avail: AtomicU64,
}
impl SystemMetrics {
- pub fn new(replication_factor: usize) -> Self {
+ pub fn new(replication_factor: usize, local_status: Arc<RwLock<NodeStatus>>) -> Self {
let meter = global::meter("garage_system");
- let values = Arc::new(SystemMetricsValues::default());
- let values1 = values.clone();
- let values2 = values.clone();
+ let st1 = local_status.clone();
+ let st2 = local_status.clone();
Self {
_garage_build_info: meter
.u64_value_observer("garage_build_info", move |observer| {
@@ -47,31 +38,28 @@ impl SystemMetrics {
.init(),
_disk_avail: meter
.u64_value_observer("garage_local_disk_avail", move |observer| {
- match values1.data_disk_avail.load(Ordering::Relaxed) {
- 0 => (),
- x => observer.observe(x, &[KeyValue::new("volume", "data")]),
- };
- match values1.meta_disk_avail.load(Ordering::Relaxed) {
- 0 => (),
- x => observer.observe(x, &[KeyValue::new("volume", "metadata")]),
- };
+ let st = st1.read().unwrap();
+ if let Some((avail, _total)) = st.data_disk_avail {
+ observer.observe(avail, &[KeyValue::new("volume", "data")]);
+ }
+ if let Some((avail, _total)) = st.meta_disk_avail {
+ observer.observe(avail, &[KeyValue::new("volume", "metadata")]);
+ }
})
.with_description("Garage available disk space on each node")
.init(),
_disk_total: meter
.u64_value_observer("garage_local_disk_total", move |observer| {
- match values2.data_disk_total.load(Ordering::Relaxed) {
- 0 => (),
- x => observer.observe(x, &[KeyValue::new("volume", "data")]),
- };
- match values2.meta_disk_total.load(Ordering::Relaxed) {
- 0 => (),
- x => observer.observe(x, &[KeyValue::new("volume", "metadata")]),
- };
+ let st = st2.read().unwrap();
+ if let Some((_avail, total)) = st.data_disk_avail {
+ observer.observe(total, &[KeyValue::new("volume", "data")]);
+ }
+ if let Some((_avail, total)) = st.meta_disk_avail {
+ observer.observe(total, &[KeyValue::new("volume", "metadata")]);
+ }
})
.with_description("Garage total disk space on each node")
.init(),
- values,
}
}
}