@@ -3201,7 +3201,7 @@ async def test_bug_5221(df_factory):
3201
3201
3202
3202
@pytest .mark .parametrize ("proactors" , [1 , 4 , 6 ])
3203
3203
@pytest .mark .parametrize ("backlog_len" , [1 , 256 , 1024 , 1300 ])
3204
- async def test_partial_sync (df_factory , df_seeder_factory , proactors , backlog_len ):
3204
+ async def test_partial_sync (df_factory , proactors , backlog_len ):
3205
3205
keys = 5_000
3206
3206
if proactors > 1 :
3207
3207
keys = 10_000
@@ -3335,8 +3335,11 @@ async def state_transitioned_stable(
3335
3335
3336
3336
3337
3337
@pytest .mark .slow
3338
- async def test_replicaiton_onmove_flow (df_factory ):
3339
- master = df_factory .create (proactor_threads = 2 , cache_mode = True )
3338
+ @pytest .mark .parametrize ("serialization_max_size" , [1 , 64000 ])
3339
+ async def test_replication_onmove_flow (df_factory , serialization_max_size ):
3340
+ master = df_factory .create (
3341
+ proactor_threads = 2 , cache_mode = True , serialization_max_chunk_size = serialization_max_size
3342
+ )
3340
3343
replica = df_factory .create (proactor_threads = 2 )
3341
3344
3342
3345
df_factory .start_all ([master , replica ])
@@ -3345,7 +3348,7 @@ async def test_replicaiton_onmove_flow(df_factory):
3345
3348
3346
3349
key_target = 100000
3347
3350
# Fill master with test data
3348
- await c_master .execute_command (f"DEBUG POPULATE { key_target } key 1048 RAND" )
3351
+ await c_master .execute_command (f"DEBUG POPULATE { key_target } key 32 RAND TYPE hash ELEMENTS 10 " )
3349
3352
logging .debug ("finished populate" )
3350
3353
3351
3354
stop_event = asyncio .Event ()
@@ -3355,7 +3358,7 @@ async def get_keys():
3355
3358
pipe = c_master .pipeline (transaction = False )
3356
3359
for _ in range (50 ):
3357
3360
id = random .randint (0 , key_target )
3358
- pipe .get (f"key:{ id } " )
3361
+ pipe .hlen (f"key:{ id } " )
3359
3362
await pipe .execute ()
3360
3363
3361
3364
get_task = asyncio .create_task (get_keys ())
0 commit comments