aboutsummaryrefslogtreecommitdiff
path: root/src/garage/tests/list.rs
diff options
context:
space:
mode:
Diffstat (limited to 'src/garage/tests/list.rs')
-rw-r--r--src/garage/tests/list.rs615
1 files changed, 0 insertions, 615 deletions
diff --git a/src/garage/tests/list.rs b/src/garage/tests/list.rs
deleted file mode 100644
index bb03f250..00000000
--- a/src/garage/tests/list.rs
+++ /dev/null
@@ -1,615 +0,0 @@
-use crate::common;
-
-const KEYS: [&str; 8] = ["a", "a/a", "a/b", "a/c", "a/d/a", "a/é", "b", "c"];
-const KEYS_MULTIPART: [&str; 5] = ["a", "a", "c", "c/a", "c/b"];
-
-#[tokio::test]
-async fn test_listobjectsv2() {
- let ctx = common::context();
- let bucket = ctx.create_bucket("listobjectsv2");
-
- for k in KEYS {
- ctx.client
- .put_object()
- .bucket(&bucket)
- .key(k)
- .send()
- .await
- .unwrap();
- }
-
- {
- // Scoping the variable to avoid reusing it
- // in a following assert due to copy paste
- let r = ctx
- .client
- .list_objects_v2()
- .bucket(&bucket)
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.contents.unwrap().len(), 8);
- assert!(r.common_prefixes.is_none());
- }
-
- //@FIXME aws-sdk-s3 automatically checks max-key values.
- // If we set it to zero, it drops it, and it is probably
- // the same behavior on values bigger than 1000.
- // Boto and awscli do not perform these tests, we should write
- // our own minimal library to bypass AWS SDK's tests and be
- // sure that we behave correctly.
-
- {
- // With 2 elements
- let r = ctx
- .client
- .list_objects_v2()
- .bucket(&bucket)
- .max_keys(2)
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.contents.unwrap().len(), 2);
- assert!(r.common_prefixes.is_none());
- assert!(r.next_continuation_token.is_some());
- }
-
- {
- // With pagination
- let mut cnt = 0;
- let mut next = None;
- let last_idx = KEYS.len() - 1;
-
- for i in 0..KEYS.len() {
- let r = ctx
- .client
- .list_objects_v2()
- .bucket(&bucket)
- .set_continuation_token(next)
- .max_keys(1)
- .send()
- .await
- .unwrap();
-
- cnt += 1;
- next = r.next_continuation_token;
-
- assert_eq!(r.contents.unwrap().len(), 1);
- assert!(r.common_prefixes.is_none());
- if i != last_idx {
- assert!(next.is_some());
- }
- }
- assert_eq!(cnt, KEYS.len());
- }
-
- {
- // With a delimiter
- let r = ctx
- .client
- .list_objects_v2()
- .bucket(&bucket)
- .delimiter("/")
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.contents.unwrap().len(), 3);
- assert_eq!(r.common_prefixes.unwrap().len(), 1);
- }
-
- {
- // With a delimiter and pagination
- let mut cnt_pfx = 0;
- let mut cnt_key = 0;
- let mut next = None;
-
- for _i in 0..KEYS.len() {
- let r = ctx
- .client
- .list_objects_v2()
- .bucket(&bucket)
- .set_continuation_token(next)
- .delimiter("/")
- .max_keys(1)
- .send()
- .await
- .unwrap();
-
- next = r.next_continuation_token;
- match (r.contents, r.common_prefixes) {
- (Some(k), None) if k.len() == 1 => cnt_key += 1,
- (None, Some(pfx)) if pfx.len() == 1 => cnt_pfx += 1,
- _ => unreachable!("logic error"),
- };
- if next.is_none() {
- break;
- }
- }
- assert_eq!(cnt_key, 3);
- assert_eq!(cnt_pfx, 1);
- }
-
- {
- // With a prefix
- let r = ctx
- .client
- .list_objects_v2()
- .bucket(&bucket)
- .prefix("a/")
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.contents.unwrap().len(), 5);
- assert!(r.common_prefixes.is_none());
- }
-
- {
- // With a prefix and a delimiter
- let r = ctx
- .client
- .list_objects_v2()
- .bucket(&bucket)
- .prefix("a/")
- .delimiter("/")
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.contents.unwrap().len(), 4);
- assert_eq!(r.common_prefixes.unwrap().len(), 1);
- }
-
- {
- // With a prefix, a delimiter and max_key
- let r = ctx
- .client
- .list_objects_v2()
- .bucket(&bucket)
- .prefix("a/")
- .delimiter("/")
- .max_keys(1)
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.contents.as_ref().unwrap().len(), 1);
- assert_eq!(
- r.contents
- .unwrap()
- .first()
- .unwrap()
- .key
- .as_ref()
- .unwrap()
- .as_str(),
- "a/a"
- );
- assert!(r.common_prefixes.is_none());
- }
- {
- // With start_after before all keys
- let r = ctx
- .client
- .list_objects_v2()
- .bucket(&bucket)
- .start_after("Z")
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.contents.unwrap().len(), 8);
- assert!(r.common_prefixes.is_none());
- }
- {
- // With start_after after all keys
- let r = ctx
- .client
- .list_objects_v2()
- .bucket(&bucket)
- .start_after("c")
- .send()
- .await
- .unwrap();
-
- assert!(r.contents.is_none());
- assert!(r.common_prefixes.is_none());
- }
-}
-
-#[tokio::test]
-async fn test_listobjectsv1() {
- let ctx = common::context();
- let bucket = ctx.create_bucket("listobjects");
-
- for k in KEYS {
- ctx.client
- .put_object()
- .bucket(&bucket)
- .key(k)
- .send()
- .await
- .unwrap();
- }
-
- {
- let r = ctx
- .client
- .list_objects()
- .bucket(&bucket)
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.contents.unwrap().len(), 8);
- assert!(r.common_prefixes.is_none());
- }
-
- {
- // With 2 elements
- let r = ctx
- .client
- .list_objects()
- .bucket(&bucket)
- .max_keys(2)
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.contents.unwrap().len(), 2);
- assert!(r.common_prefixes.is_none());
- assert!(r.next_marker.is_some());
- }
-
- {
- // With pagination
- let mut cnt = 0;
- let mut next = None;
- let last_idx = KEYS.len() - 1;
-
- for i in 0..KEYS.len() {
- let r = ctx
- .client
- .list_objects()
- .bucket(&bucket)
- .set_marker(next)
- .max_keys(1)
- .send()
- .await
- .unwrap();
-
- cnt += 1;
- next = r.next_marker;
-
- assert_eq!(r.contents.unwrap().len(), 1);
- assert!(r.common_prefixes.is_none());
- if i != last_idx {
- assert!(next.is_some());
- }
- }
- assert_eq!(cnt, KEYS.len());
- }
-
- {
- // With a delimiter
- let r = ctx
- .client
- .list_objects()
- .bucket(&bucket)
- .delimiter("/")
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.contents.unwrap().len(), 3);
- assert_eq!(r.common_prefixes.unwrap().len(), 1);
- }
-
- {
- // With a delimiter and pagination
- let mut cnt_pfx = 0;
- let mut cnt_key = 0;
- let mut next = None;
-
- for _i in 0..KEYS.len() {
- let r = ctx
- .client
- .list_objects()
- .bucket(&bucket)
- .delimiter("/")
- .set_marker(next)
- .max_keys(1)
- .send()
- .await
- .unwrap();
-
- next = r.next_marker;
- match (r.contents, r.common_prefixes) {
- (Some(k), None) if k.len() == 1 => cnt_key += 1,
- (None, Some(pfx)) if pfx.len() == 1 => cnt_pfx += 1,
- _ => unreachable!("logic error"),
- };
- if next.is_none() {
- break;
- }
- }
- assert_eq!(cnt_key, 3);
- // We have no optimization to skip the whole prefix
- // on listobjectsv1 so we return the same one 5 times,
- // for each element. It is up to the client to merge its result.
- // This is compliant with AWS spec.
- assert_eq!(cnt_pfx, 5);
- }
-
- {
- // With a prefix
- let r = ctx
- .client
- .list_objects()
- .bucket(&bucket)
- .prefix("a/")
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.contents.unwrap().len(), 5);
- assert!(r.common_prefixes.is_none());
- }
-
- {
- // With a prefix and a delimiter
- let r = ctx
- .client
- .list_objects()
- .bucket(&bucket)
- .prefix("a/")
- .delimiter("/")
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.contents.unwrap().len(), 4);
- assert_eq!(r.common_prefixes.unwrap().len(), 1);
- }
-
- {
- // With a prefix, a delimiter and max_key
- let r = ctx
- .client
- .list_objects()
- .bucket(&bucket)
- .prefix("a/")
- .delimiter("/")
- .max_keys(1)
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.contents.as_ref().unwrap().len(), 1);
- assert_eq!(
- r.contents
- .unwrap()
- .first()
- .unwrap()
- .key
- .as_ref()
- .unwrap()
- .as_str(),
- "a/a"
- );
- assert!(r.common_prefixes.is_none());
- }
- {
- // With marker before all keys
- let r = ctx
- .client
- .list_objects()
- .bucket(&bucket)
- .marker("Z")
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.contents.unwrap().len(), 8);
- assert!(r.common_prefixes.is_none());
- }
- {
- // With start_after after all keys
- let r = ctx
- .client
- .list_objects()
- .bucket(&bucket)
- .marker("c")
- .send()
- .await
- .unwrap();
-
- assert!(r.contents.is_none());
- assert!(r.common_prefixes.is_none());
- }
-}
-
-#[tokio::test]
-async fn test_listmultipart() {
- let ctx = common::context();
- let bucket = ctx.create_bucket("listmultipartuploads");
-
- for k in KEYS_MULTIPART {
- ctx.client
- .create_multipart_upload()
- .bucket(&bucket)
- .key(k)
- .send()
- .await
- .unwrap();
- }
-
- {
- // Default
- let r = ctx
- .client
- .list_multipart_uploads()
- .bucket(&bucket)
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.uploads.unwrap().len(), 5);
- assert!(r.common_prefixes.is_none());
- }
- {
- // With pagination
- let mut next = None;
- let mut upnext = None;
- let last_idx = KEYS_MULTIPART.len() - 1;
-
- for i in 0..KEYS_MULTIPART.len() {
- let r = ctx
- .client
- .list_multipart_uploads()
- .bucket(&bucket)
- .set_key_marker(next)
- .set_upload_id_marker(upnext)
- .max_uploads(1)
- .send()
- .await
- .unwrap();
-
- next = r.next_key_marker;
- upnext = r.next_upload_id_marker;
-
- assert_eq!(r.uploads.unwrap().len(), 1);
- assert!(r.common_prefixes.is_none());
- if i != last_idx {
- assert!(next.is_some());
- }
- }
- }
- {
- // With delimiter
- let r = ctx
- .client
- .list_multipart_uploads()
- .bucket(&bucket)
- .delimiter("/")
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.uploads.unwrap().len(), 3);
- assert_eq!(r.common_prefixes.unwrap().len(), 1);
- }
- {
- // With delimiter and pagination
- let mut next = None;
- let mut upnext = None;
- let mut upcnt = 0;
- let mut pfxcnt = 0;
- let mut loopcnt = 0;
-
- while loopcnt < KEYS_MULTIPART.len() {
- let r = ctx
- .client
- .list_multipart_uploads()
- .bucket(&bucket)
- .delimiter("/")
- .max_uploads(1)
- .set_key_marker(next)
- .set_upload_id_marker(upnext)
- .send()
- .await
- .unwrap();
-
- next = r.next_key_marker;
- upnext = r.next_upload_id_marker;
-
- loopcnt += 1;
- upcnt += r.uploads.unwrap_or_default().len();
- pfxcnt += r.common_prefixes.unwrap_or_default().len();
-
- if next.is_none() {
- break;
- }
- }
-
- assert_eq!(upcnt + pfxcnt, loopcnt);
- assert_eq!(upcnt, 3);
- assert_eq!(pfxcnt, 1);
- }
- {
- // With prefix
- let r = ctx
- .client
- .list_multipart_uploads()
- .bucket(&bucket)
- .prefix("c")
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.uploads.unwrap().len(), 3);
- assert!(r.common_prefixes.is_none());
- }
- {
- // With prefix and delimiter
- let r = ctx
- .client
- .list_multipart_uploads()
- .bucket(&bucket)
- .prefix("c")
- .delimiter("/")
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.uploads.unwrap().len(), 1);
- assert_eq!(r.common_prefixes.unwrap().len(), 1);
- }
- {
- // With prefix, delimiter and max keys
- let r = ctx
- .client
- .list_multipart_uploads()
- .bucket(&bucket)
- .prefix("c")
- .delimiter("/")
- .max_uploads(1)
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.uploads.unwrap().len(), 1);
- assert!(r.common_prefixes.is_none());
- }
- {
- // With starting token before the first element
- let r = ctx
- .client
- .list_multipart_uploads()
- .bucket(&bucket)
- .key_marker("ZZZZZ")
- .send()
- .await
- .unwrap();
-
- assert_eq!(r.uploads.unwrap().len(), 5);
- assert!(r.common_prefixes.is_none());
- }
- {
- // With starting token after the last element
- let r = ctx
- .client
- .list_multipart_uploads()
- .bucket(&bucket)
- .key_marker("d")
- .send()
- .await
- .unwrap();
-
- assert!(r.uploads.is_none());
- assert!(r.common_prefixes.is_none());
- }
-}