@@ -5092,8 +5092,6 @@ def test_copy_obj_between_zonegroups(zonegroup):
50925092 source_bucket = source_zone .create_bucket (gen_bucket_name ())
50935093
50945094 objname = 'dummy'
5095- k = new_key (source_zone , source_bucket .name , objname )
5096- k .set_contents_from_string ('foo' )
50975095
50985096 for zg in realm .current_period .zonegroups :
50995097 if zg .name == zonegroup .name :
@@ -5103,16 +5101,24 @@ def test_copy_obj_between_zonegroups(zonegroup):
51035101 dest_bucket = dest_zone .create_bucket (gen_bucket_name ())
51045102 realm_meta_checkpoint (realm )
51055103
5106- # copy object
5107- dest_zone .s3_client .copy_object (
5108- Bucket = dest_bucket .name ,
5109- CopySource = f'{ source_bucket .name } /{ objname } ' ,
5110- Key = objname
5111- )
5112-
5113- # check that object exists in destination bucket
5114- k = get_key (dest_zone , dest_bucket , objname )
5115- assert_equal (k .get_contents_as_string ().decode ('utf-8' ), 'foo' )
5104+ # try object sizes of 4K and 8MiB
5105+ # 4K to test no progress case
5106+ # 8MiB to test progress case
5107+ obj_sizes = [4096 , 8 * 1024 * 1024 ]
5108+ for size in obj_sizes :
5109+ k = new_key (source_zone , source_bucket .name , objname )
5110+ k .set_contents_from_string ('x' * size )
5111+
5112+ # copy object
5113+ dest_zone .s3_client .copy_object (
5114+ Bucket = dest_bucket .name ,
5115+ CopySource = f'{ source_bucket .name } /{ objname } ' ,
5116+ Key = objname
5117+ )
5118+
5119+ # check that object exists in destination bucket
5120+ k = get_key (dest_zone , dest_bucket , objname )
5121+ assert_equal (k .get_contents_as_string ().decode ('utf-8' ), 'x' * size )
51165122
51175123@allow_bucket_replication
51185124def test_bucket_replication_alt_user_delete_forbidden ():
0 commit comments