diff options
author | Alex Auvolat <alex@adnab.me> | 2021-10-26 10:20:05 +0200 |
---|---|---|
committer | Alex Auvolat <alex@adnab.me> | 2021-10-26 10:20:05 +0200 |
commit | ada7899b24cacb4f95ee67ae92ddb73c04ee4d66 (patch) | |
tree | e42c200cd5eb9e9190232d4340f5be27f302342c | |
parent | b2c51844a1dccf229473ce0c27fdb02a433ac243 (diff) | |
download | garage-ada7899b24cacb4f95ee67ae92ddb73c04ee4d66.tar.gz garage-ada7899b24cacb4f95ee67ae92ddb73c04ee4d66.zip |
Fix clippy lints (fix #121)
-rw-r--r-- | .drone.yml | 2 | ||||
-rw-r--r-- | src/api/api_server.rs | 35 | ||||
-rw-r--r-- | src/api/s3_delete.rs | 5 | ||||
-rw-r--r-- | src/api/s3_get.rs | 10 | ||||
-rw-r--r-- | src/api/s3_put.rs | 10 | ||||
-rw-r--r-- | src/api/signature.rs | 4 | ||||
-rw-r--r-- | src/garage/cli/util.rs | 2 | ||||
-rw-r--r-- | src/model/block.rs | 12 | ||||
-rw-r--r-- | src/rpc/system.rs | 2 | ||||
-rw-r--r-- | src/table/crdt/lww_map.rs | 6 | ||||
-rw-r--r-- | src/table/crdt/map.rs | 6 | ||||
-rw-r--r-- | src/table/merkle.rs | 4 | ||||
-rw-r--r-- | src/table/replication/sharded.rs | 4 | ||||
-rw-r--r-- | src/table/sync.rs | 4 | ||||
-rw-r--r-- | src/web/web_server.rs | 8 |
15 files changed, 56 insertions, 58 deletions
@@ -36,7 +36,7 @@ steps: path: /etc/nix commands: - nix-shell --arg release false --run "cargo fmt -- --check" - - nix-shell --arg release false --run "cargo clippy -- --allow clippy::needless_borrow --allow clippy::needless-return --deny warnings" + - nix-shell --arg release false --run "cargo clippy -- --deny warnings" - name: build image: nixpkgs/nix:nixos-21.05 diff --git a/src/api/api_server.rs b/src/api/api_server.rs index 52031ae0..d51b5a28 100644 --- a/src/api/api_server.rs +++ b/src/api/api_server.rs @@ -39,7 +39,7 @@ pub async fn run_api_server( } }); - let server = Server::bind(&addr).serve(service); + let server = Server::bind(addr).serve(service); let graceful = server.with_graceful_shutdown(shutdown_signal); info!("API server listening on http://{}", addr); @@ -88,8 +88,8 @@ async fn handler_inner(garage: Arc<Garage>, req: Request<Body>) -> Result<Respon let (bucket, key) = parse_bucket_key(&path)?; let allowed = match req.method() { - &Method::HEAD | &Method::GET => api_key.allow_read(&bucket), - _ => api_key.allow_write(&bucket), + &Method::HEAD | &Method::GET => api_key.allow_read(bucket), + _ => api_key.allow_write(bucket), }; if !allowed { return Err(Error::Forbidden( @@ -109,11 +109,11 @@ async fn handler_inner(garage: Arc<Garage>, req: Request<Body>) -> Result<Respon match *req.method() { Method::HEAD => { // HeadObject query - Ok(handle_head(garage, &req, &bucket, &key).await?) + Ok(handle_head(garage, &req, bucket, key).await?) } Method::GET => { // GetObject query - Ok(handle_get(garage, &req, &bucket, &key).await?) + Ok(handle_get(garage, &req, bucket, key).await?) } Method::PUT => { if params.contains_key(&"partnumber".to_string()) @@ -125,8 +125,8 @@ async fn handler_inner(garage: Arc<Garage>, req: Request<Body>) -> Result<Respon Ok(handle_put_part( garage, req, - &bucket, - &key, + bucket, + key, part_number, upload_id, content_sha256, @@ -136,46 +136,43 @@ async fn handler_inner(garage: Arc<Garage>, req: Request<Body>) -> Result<Respon // CopyObject query let copy_source = req.headers().get("x-amz-copy-source").unwrap().to_str()?; let copy_source = - percent_encoding::percent_decode_str(©_source).decode_utf8()?; + percent_encoding::percent_decode_str(copy_source).decode_utf8()?; let (source_bucket, source_key) = parse_bucket_key(©_source)?; - if !api_key.allow_read(&source_bucket) { + if !api_key.allow_read(source_bucket) { return Err(Error::Forbidden(format!( "Reading from bucket {} not allowed for this key", source_bucket ))); } let source_key = source_key.ok_or_bad_request("No source key specified")?; - Ok( - handle_copy(garage, &req, &bucket, &key, &source_bucket, &source_key) - .await?, - ) + Ok(handle_copy(garage, &req, bucket, key, source_bucket, source_key).await?) } else { // PutObject query - Ok(handle_put(garage, req, &bucket, &key, content_sha256).await?) + Ok(handle_put(garage, req, bucket, key, content_sha256).await?) } } Method::DELETE => { if params.contains_key(&"uploadid".to_string()) { // AbortMultipartUpload query let upload_id = params.get("uploadid").unwrap(); - Ok(handle_abort_multipart_upload(garage, &bucket, &key, upload_id).await?) + Ok(handle_abort_multipart_upload(garage, bucket, key, upload_id).await?) } else { // DeleteObject query - Ok(handle_delete(garage, &bucket, &key).await?) + Ok(handle_delete(garage, bucket, key).await?) } } Method::POST => { if params.contains_key(&"uploads".to_string()) { // CreateMultipartUpload call - Ok(handle_create_multipart_upload(garage, &req, &bucket, &key).await?) + Ok(handle_create_multipart_upload(garage, &req, bucket, key).await?) } else if params.contains_key(&"uploadid".to_string()) { // CompleteMultipartUpload call let upload_id = params.get("uploadid").unwrap(); Ok(handle_complete_multipart_upload( garage, req, - &bucket, - &key, + bucket, + key, upload_id, content_sha256, ) diff --git a/src/api/s3_delete.rs b/src/api/s3_delete.rs index efe325ae..425f86d7 100644 --- a/src/api/s3_delete.rs +++ b/src/api/s3_delete.rs @@ -55,7 +55,8 @@ async fn handle_delete_internal( ); garage.object_table.insert(&object).await?; - return Ok((deleted_version, version_uuid)); + + Ok((deleted_version, version_uuid)) } pub async fn handle_delete( @@ -82,7 +83,7 @@ pub async fn handle_delete_objects( let body = hyper::body::to_bytes(req.into_body()).await?; verify_signed_content(content_sha256, &body[..])?; - let cmd_xml = roxmltree::Document::parse(&std::str::from_utf8(&body)?)?; + let cmd_xml = roxmltree::Document::parse(std::str::from_utf8(&body)?)?; let cmd = parse_delete_objects_xml(&cmd_xml).ok_or_bad_request("Invalid delete XML query")?; let mut ret_deleted = Vec::new(); diff --git a/src/api/s3_get.rs b/src/api/s3_get.rs index b804d8ee..10d6591f 100644 --- a/src/api/s3_get.rs +++ b/src/api/s3_get.rs @@ -106,12 +106,12 @@ pub async fn handle_head( _ => unreachable!(), }; - if let Some(cached) = try_answer_cached(&version, version_meta, req) { + if let Some(cached) = try_answer_cached(version, version_meta, req) { return Ok(cached); } let body: Body = Body::empty(); - let response = object_headers(&version, version_meta) + let response = object_headers(version, version_meta) .header("Content-Length", format!("{}", version_meta.size)) .status(StatusCode::OK) .body(body) @@ -149,7 +149,7 @@ pub async fn handle_get( ObjectVersionData::FirstBlock(meta, _) => meta, }; - if let Some(cached) = try_answer_cached(&last_v, last_v_meta, req) { + if let Some(cached) = try_answer_cached(last_v, last_v_meta, req) { return Ok(cached); } @@ -179,7 +179,7 @@ pub async fn handle_get( .await; } - let resp_builder = object_headers(&last_v, last_v_meta) + let resp_builder = object_headers(last_v, last_v_meta) .header("Content-Length", format!("{}", last_v_meta.size)) .status(StatusCode::OK); @@ -190,7 +190,7 @@ pub async fn handle_get( Ok(resp_builder.body(body)?) } ObjectVersionData::FirstBlock(_, first_block_hash) => { - let read_first_block = garage.block_manager.rpc_get_block(&first_block_hash); + let read_first_block = garage.block_manager.rpc_get_block(first_block_hash); let get_next_blocks = garage.version_table.get(&last_v.uuid, &EmptyKey); let (first_block, version) = futures::try_join!(read_first_block, get_next_blocks)?; diff --git a/src/api/s3_put.rs b/src/api/s3_put.rs index aa285232..5eae3bf5 100644 --- a/src/api/s3_put.rs +++ b/src/api/s3_put.rs @@ -193,8 +193,8 @@ async fn read_and_put_blocks( let mut next_offset = first_block.len(); let mut put_curr_version_block = put_block_meta( - &garage, - &version, + garage, + version, part_number, 0, first_block_hash, @@ -213,8 +213,8 @@ async fn read_and_put_blocks( let block_hash = blake2sum(&block[..]); let block_len = block.len(); put_curr_version_block = put_block_meta( - &garage, - &version, + garage, + version, part_number, next_offset as u64, block_hash, @@ -437,7 +437,7 @@ pub async fn handle_complete_multipart_upload( let body = hyper::body::to_bytes(req.into_body()).await?; verify_signed_content(content_sha256, &body[..])?; - let body_xml = roxmltree::Document::parse(&std::str::from_utf8(&body)?)?; + let body_xml = roxmltree::Document::parse(std::str::from_utf8(&body)?)?; let body_list_of_parts = parse_complete_multpart_upload_body(&body_xml) .ok_or_bad_request("Invalid CompleteMultipartUpload XML")?; debug!( diff --git a/src/api/signature.rs b/src/api/signature.rs index fd7ad33b..53ca2ce5 100644 --- a/src/api/signature.rs +++ b/src/api/signature.rs @@ -70,7 +70,7 @@ pub async fn check_signature( let canonical_request = canonical_request( request.method(), &request.uri().path().to_string(), - &canonical_query_string(&request.uri()), + &canonical_query_string(request.uri()), &headers, &authorization.signed_headers, &authorization.content_sha256, @@ -252,7 +252,7 @@ fn canonical_request( method.as_str(), url_path, canonical_query_string, - &canonical_header_string(&headers, signed_headers), + &canonical_header_string(headers, signed_headers), "", signed_headers, content_sha256, diff --git a/src/garage/cli/util.rs b/src/garage/cli/util.rs index 28b4d8ea..647a2449 100644 --- a/src/garage/cli/util.rs +++ b/src/garage/cli/util.rs @@ -54,7 +54,7 @@ pub fn format_table(data: Vec<String>) { out.push_str(col); (0..col_len - col.chars().count() + 2).for_each(|_| out.push(' ')); } - out.push_str(&row[row.len() - 1]); + out.push_str(row[row.len() - 1]); out.push('\n'); } diff --git a/src/model/block.rs b/src/model/block.rs index 85468258..d1ea1512 100644 --- a/src/model/block.rs +++ b/src/model/block.rs @@ -129,7 +129,7 @@ impl BlockManager { /// Ask nodes that might have a block for it pub async fn rpc_get_block(&self, hash: &Hash) -> Result<Vec<u8>, Error> { - let who = self.replication.read_nodes(&hash); + let who = self.replication.read_nodes(hash); let resps = self .system .rpc @@ -224,7 +224,7 @@ impl BlockManager { })?; let old_rc = old_rc.map(u64_from_be_bytes).unwrap_or(0); if old_rc == 0 { - self.put_to_resync(&hash, BLOCK_RW_TIMEOUT)?; + self.put_to_resync(hash, BLOCK_RW_TIMEOUT)?; } Ok(()) } @@ -240,7 +240,7 @@ impl BlockManager { } })?; if new_rc.is_none() { - self.put_to_resync(&hash, BLOCK_GC_TIMEOUT)?; + self.put_to_resync(hash, BLOCK_GC_TIMEOUT)?; } Ok(()) } @@ -406,7 +406,7 @@ impl BlockManager { if exists && !needed { trace!("Offloading block {:?}", hash); - let mut who = self.replication.write_nodes(&hash); + let mut who = self.replication.write_nodes(hash); if who.len() < self.replication.write_quorum() { return Err(Error::Message("Not trying to offload block because we don't have a quorum of nodes to write to".to_string())); } @@ -478,7 +478,7 @@ impl BlockManager { // TODO find a way to not do this if they are sending it to us // Let's suppose this isn't an issue for now with the BLOCK_RW_TIMEOUT delay // between the RC being incremented and this part being called. - let block_data = self.rpc_get_block(&hash).await?; + let block_data = self.rpc_get_block(hash).await?; self.write_block(hash, &block_data[..]).await?; } @@ -600,7 +600,7 @@ impl BlockManagerLocked { let mut path2 = path.clone(); path2.set_extension(".corrupted"); fs::rename(path, path2).await?; - mgr.put_to_resync(&hash, Duration::from_millis(0))?; + mgr.put_to_resync(hash, Duration::from_millis(0))?; Ok(()) } diff --git a/src/rpc/system.rs b/src/rpc/system.rs index f5ce817f..ed18f657 100644 --- a/src/rpc/system.rs +++ b/src/rpc/system.rs @@ -543,7 +543,7 @@ impl EndpointHandler<SystemRpc> for System { SystemRpc::Connect(node) => self.handle_connect(node).await, SystemRpc::PullConfig => Ok(self.handle_pull_config()), SystemRpc::AdvertiseStatus(adv) => self.handle_advertise_status(from.into(), adv).await, - SystemRpc::AdvertiseConfig(adv) => self.clone().handle_advertise_config(&adv).await, + SystemRpc::AdvertiseConfig(adv) => self.clone().handle_advertise_config(adv).await, SystemRpc::GetKnownNodes => Ok(self.handle_get_known_nodes()), _ => Err(Error::BadRpc("Unexpected RPC message".to_string())), } diff --git a/src/table/crdt/lww_map.rs b/src/table/crdt/lww_map.rs index 36bbf667..fb25fd46 100644 --- a/src/table/crdt/lww_map.rs +++ b/src/table/crdt/lww_map.rs @@ -103,7 +103,7 @@ where } /// Get a reference to the value assigned to a key pub fn get(&self, k: &K) -> Option<&V> { - match self.vals.binary_search_by(|(k2, _, _)| k2.cmp(&k)) { + match self.vals.binary_search_by(|(k2, _, _)| k2.cmp(k)) { Ok(i) => Some(&self.vals[i].2), Err(_) => None, } @@ -132,14 +132,14 @@ where { fn merge(&mut self, other: &Self) { for (k, ts2, v2) in other.vals.iter() { - match self.vals.binary_search_by(|(k2, _, _)| k2.cmp(&k)) { + match self.vals.binary_search_by(|(k2, _, _)| k2.cmp(k)) { Ok(i) => { let (_, ts1, _v1) = &self.vals[i]; if ts2 > ts1 { self.vals[i].1 = *ts2; self.vals[i].2 = v2.clone(); } else if ts1 == ts2 { - self.vals[i].2.merge(&v2); + self.vals[i].2.merge(v2); } } Err(i) => { diff --git a/src/table/crdt/map.rs b/src/table/crdt/map.rs index e2aee40a..7553cd50 100644 --- a/src/table/crdt/map.rs +++ b/src/table/crdt/map.rs @@ -49,7 +49,7 @@ where /// Get a reference to the value assigned to a key pub fn get(&self, k: &K) -> Option<&V> { - match self.vals.binary_search_by(|(k2, _)| k2.cmp(&k)) { + match self.vals.binary_search_by(|(k2, _)| k2.cmp(k)) { Ok(i) => Some(&self.vals[i].1), Err(_) => None, } @@ -76,9 +76,9 @@ where { fn merge(&mut self, other: &Self) { for (k, v2) in other.vals.iter() { - match self.vals.binary_search_by(|(k2, _)| k2.cmp(&k)) { + match self.vals.binary_search_by(|(k2, _)| k2.cmp(k)) { Ok(i) => { - self.vals[i].1.merge(&v2); + self.vals[i].1.merge(v2); } Err(i) => { self.vals.insert(i, (k.clone(), v2.clone())); diff --git a/src/table/merkle.rs b/src/table/merkle.rs index 5c5cbec7..56f307d3 100644 --- a/src/table/merkle.rs +++ b/src/table/merkle.rs @@ -167,7 +167,7 @@ where // Calculate an update to apply to this node // This update is an Option<_>, so that it is None if the update is a no-op // and we can thus skip recalculating and re-storing everything - let mutate = match self.read_node_txn(tx, &key)? { + let mutate = match self.read_node_txn(tx, key)? { MerkleNode::Empty => new_vhash.map(|vhv| MerkleNode::Leaf(k.to_vec(), vhv)), MerkleNode::Intermediate(mut children) => { let key2 = key.next_key(khash); @@ -270,7 +270,7 @@ where }; if let Some(new_node) = mutate { - let hash = self.put_node_txn(tx, &key, &new_node)?; + let hash = self.put_node_txn(tx, key, &new_node)?; Ok(Some(hash)) } else { Ok(None) diff --git a/src/table/replication/sharded.rs b/src/table/replication/sharded.rs index 75043a17..1cf964af 100644 --- a/src/table/replication/sharded.rs +++ b/src/table/replication/sharded.rs @@ -27,7 +27,7 @@ pub struct TableShardedReplication { impl TableReplication for TableShardedReplication { fn read_nodes(&self, hash: &Hash) -> Vec<Uuid> { let ring = self.system.ring.borrow(); - ring.get_nodes(&hash, self.replication_factor) + ring.get_nodes(hash, self.replication_factor) } fn read_quorum(&self) -> usize { self.read_quorum @@ -35,7 +35,7 @@ impl TableReplication for TableShardedReplication { fn write_nodes(&self, hash: &Hash) -> Vec<Uuid> { let ring = self.system.ring.borrow(); - ring.get_nodes(&hash, self.replication_factor) + ring.get_nodes(hash, self.replication_factor) } fn write_quorum(&self) -> usize { self.write_quorum diff --git a/src/table/sync.rs b/src/table/sync.rs index 4fcdc528..c5795f65 100644 --- a/src/table/sync.rs +++ b/src/table/sync.rs @@ -266,7 +266,7 @@ where let nodes = self .data .replication - .write_nodes(&begin) + .write_nodes(begin) .into_iter() .collect::<Vec<_>>(); if nodes.contains(&self.system.id) { @@ -530,7 +530,7 @@ where Ok(SyncRpc::RootCkDifferent(hash != *h)) } SyncRpc::GetNode(k) => { - let node = self.merkle.read_node(&k)?; + let node = self.merkle.read_node(k)?; Ok(SyncRpc::Node(k.clone(), node)) } SyncRpc::Items(items) => { diff --git a/src/web/web_server.rs b/src/web/web_server.rs index babde62a..bff9e71c 100644 --- a/src/web/web_server.rs +++ b/src/web/web_server.rs @@ -36,7 +36,7 @@ pub async fn run_web_server( } }); - let server = Server::bind(&addr).serve(service); + let server = Server::bind(addr).serve(service); let graceful = server.with_graceful_shutdown(shutdown_signal); info!("Web server listening on http://{}", addr); @@ -95,12 +95,12 @@ async fn serve_file(garage: Arc<Garage>, req: Request<Body>) -> Result<Response< // Get path let path = req.uri().path().to_string(); let index = &garage.config.s3_web.index; - let key = path_to_key(&path, &index)?; + let key = path_to_key(&path, index)?; info!("Selected bucket: \"{}\", selected key: \"{}\"", bucket, key); let res = match *req.method() { - Method::HEAD => handle_head(garage, &req, &bucket, &key).await?, + Method::HEAD => handle_head(garage, &req, bucket, &key).await?, Method::GET => handle_get(garage, &req, bucket, &key).await?, _ => return Err(Error::BadRequest("HTTP method not supported".to_string())), }; @@ -173,7 +173,7 @@ fn host_to_bucket<'a>(host: &'a str, root: &str) -> &'a str { /// When a path ends with "/", we append the index name to match traditional web server behavior /// which is also AWS S3 behavior. fn path_to_key<'a>(path: &'a str, index: &str) -> Result<Cow<'a, str>, Error> { - let path_utf8 = percent_encoding::percent_decode_str(&path).decode_utf8()?; + let path_utf8 = percent_encoding::percent_decode_str(path).decode_utf8()?; if !path_utf8.starts_with('/') { return Err(Error::BadRequest( |