@@ -140,7 +140,7 @@ def get_this_proc(v1: bool):
140
140
141
141
142
142
@pytest .mark .parametrize ("v1" , [True , False ])
143
- @pytest .mark .timeout (60 )
143
+ @pytest .mark .timeout (360 )
144
144
async def test_choose (v1 : bool ):
145
145
proc = spawn_procs_on_fake_host (v1 , {"gpus" : 2 })
146
146
v = proc .spawn ("counter" , Counter , 3 )
@@ -160,7 +160,7 @@ async def test_choose(v1: bool):
160
160
161
161
162
162
@pytest .mark .parametrize ("v1" , [True , False ])
163
- @pytest .mark .timeout (60 )
163
+ @pytest .mark .timeout (360 )
164
164
async def test_stream (v1 : bool ):
165
165
proc = spawn_procs_on_fake_host (v1 , {"gpus" : 2 })
166
166
v = proc .spawn ("counter2" , Counter , 3 )
@@ -182,7 +182,7 @@ async def fetch(self, to: To):
182
182
183
183
184
184
@pytest .mark .parametrize ("v1" , [True , False ])
185
- @pytest .mark .timeout (60 )
185
+ @pytest .mark .timeout (360 )
186
186
async def test_mesh_passed_to_mesh (v1 : bool ):
187
187
proc = spawn_procs_on_fake_host (v1 , {"gpus" : 2 })
188
188
f = proc .spawn ("from" , From )
@@ -193,7 +193,7 @@ async def test_mesh_passed_to_mesh(v1: bool):
193
193
194
194
195
195
@pytest .mark .parametrize ("v1" , [True , False ])
196
- @pytest .mark .timeout (60 )
196
+ @pytest .mark .timeout (360 )
197
197
async def test_mesh_passed_to_mesh_on_different_proc_mesh (v1 : bool ):
198
198
proc = spawn_procs_on_fake_host (v1 , {"gpus" : 2 })
199
199
proc2 = spawn_procs_on_fake_host (v1 , {"gpus" : 2 })
@@ -205,7 +205,7 @@ async def test_mesh_passed_to_mesh_on_different_proc_mesh(v1: bool):
205
205
206
206
207
207
@pytest .mark .parametrize ("v1" , [True , False ])
208
- @pytest .mark .timeout (60 )
208
+ @pytest .mark .timeout (360 )
209
209
def test_actor_slicing (v1 : bool ):
210
210
proc = spawn_procs_on_fake_host (v1 , {"gpus" : 2 })
211
211
proc2 = spawn_procs_on_fake_host (v1 , {"gpus" : 2 })
@@ -222,7 +222,7 @@ def test_actor_slicing(v1: bool):
222
222
223
223
224
224
@pytest .mark .parametrize ("v1" , [True , False ])
225
- @pytest .mark .timeout (60 )
225
+ @pytest .mark .timeout (360 )
226
226
async def test_aggregate (v1 : bool ):
227
227
proc = spawn_procs_on_fake_host (v1 , {"gpus" : 2 })
228
228
counter = proc .spawn ("counter" , Counter , 1 )
@@ -243,7 +243,7 @@ async def return_current_rank_str(self):
243
243
244
244
245
245
@pytest .mark .parametrize ("v1" , [True , False ])
246
- @pytest .mark .timeout (60 )
246
+ @pytest .mark .timeout (360 )
247
247
async def test_rank_size (v1 : bool ):
248
248
proc = spawn_procs_on_fake_host (v1 , {"gpus" : 2 })
249
249
r = proc .spawn ("runit" , RunIt )
@@ -255,7 +255,7 @@ async def test_rank_size(v1: bool):
255
255
256
256
257
257
@pytest .mark .parametrize ("v1" , [True , False ])
258
- @pytest .mark .timeout (60 )
258
+ @pytest .mark .timeout (360 )
259
259
async def test_rank_string (v1 : bool ):
260
260
if v1 :
261
261
per_host = {"gpus" : 2 }
@@ -277,7 +277,7 @@ def sync_endpoint(self, a_counter: Counter):
277
277
278
278
279
279
@pytest .mark .parametrize ("v1" , [True , False ])
280
- @pytest .mark .timeout (60 )
280
+ @pytest .mark .timeout (360 )
281
281
async def test_sync_actor (v1 : bool ):
282
282
proc = spawn_procs_on_fake_host (v1 , {"gpus" : 2 })
283
283
a = proc .spawn ("actor" , SyncActor )
@@ -287,7 +287,7 @@ async def test_sync_actor(v1: bool):
287
287
288
288
289
289
@pytest .mark .parametrize ("v1" , [True , False ])
290
- @pytest .mark .timeout (60 )
290
+ @pytest .mark .timeout (360 )
291
291
def test_sync_actor_sync_client (v1 : bool ) -> None :
292
292
proc = spawn_procs_on_fake_host (v1 , {"gpus" : 2 })
293
293
a = proc .spawn ("actor" , SyncActor )
@@ -297,14 +297,14 @@ def test_sync_actor_sync_client(v1: bool) -> None:
297
297
298
298
299
299
@pytest .mark .parametrize ("v1" , [True , False ])
300
- @pytest .mark .timeout (60 )
300
+ @pytest .mark .timeout (360 )
301
301
def test_proc_mesh_size (v1 : bool ) -> None :
302
302
proc = spawn_procs_on_fake_host (v1 , {"gpus" : 2 })
303
303
assert 2 == proc .size ("gpus" )
304
304
305
305
306
306
@pytest .mark .parametrize ("v1" , [True , False ])
307
- @pytest .mark .timeout (60 )
307
+ @pytest .mark .timeout (360 )
308
308
def test_rank_size_sync (v1 : bool ) -> None :
309
309
proc = spawn_procs_on_fake_host (v1 , {"gpus" : 2 })
310
310
r = proc .spawn ("runit" , RunIt )
@@ -315,7 +315,7 @@ def test_rank_size_sync(v1: bool) -> None:
315
315
316
316
317
317
@pytest .mark .parametrize ("v1" , [True , False ])
318
- @pytest .mark .timeout (60 )
318
+ @pytest .mark .timeout (360 )
319
319
def test_accumulate_sync (v1 : bool ) -> None :
320
320
proc = spawn_procs_on_fake_host (v1 , {"gpus" : 2 })
321
321
counter = proc .spawn ("counter" , Counter , 1 )
@@ -332,7 +332,7 @@ def doit(self, c: Counter):
332
332
333
333
334
334
@pytest .mark .parametrize ("v1" , [True , False ])
335
- @pytest .mark .timeout (60 )
335
+ @pytest .mark .timeout (360 )
336
336
def test_value_mesh (v1 : bool ) -> None :
337
337
if v1 :
338
338
per_host = {"gpus" : 2 }
@@ -349,7 +349,7 @@ def test_value_mesh(v1: bool) -> None:
349
349
assert list (x ) == n .slice (gpus = 0 ).doit .call_one (counter ).get ()
350
350
351
351
352
- @pytest .mark .timeout (60 )
352
+ @pytest .mark .timeout (360 )
353
353
def test_rust_binding_modules_correct () -> None :
354
354
"""
355
355
This tests that rust bindings will survive pickling correctly.
@@ -378,7 +378,7 @@ def check(module, path):
378
378
379
379
380
380
@pytest .mark .parametrize ("v1" , [True , False ])
381
- @pytest .mark .timeout (60 )
381
+ @pytest .mark .timeout (360 )
382
382
def test_proc_mesh_liveness (v1 : bool ) -> None :
383
383
mesh = spawn_procs_on_this_host (v1 , {"gpus" : 2 })
384
384
counter = mesh .spawn ("counter" , Counter , 1 )
@@ -414,7 +414,7 @@ async def get_async(self):
414
414
415
415
416
416
@pytest .mark .parametrize ("v1" , [True , False ])
417
- @pytest .mark .timeout (60 )
417
+ @pytest .mark .timeout (360 )
418
418
async def test_actor_tls (v1 : bool ) -> None :
419
419
"""Test that thread-local state is respected."""
420
420
pm = spawn_procs_on_this_host (v1 , {"gpus" : 1 })
@@ -445,7 +445,7 @@ def get_value(self):
445
445
446
446
447
447
@pytest .mark .parametrize ("v1" , [True , False ])
448
- @pytest .mark .timeout (60 )
448
+ @pytest .mark .timeout (360 )
449
449
async def test_actor_tls_full_sync (v1 : bool ) -> None :
450
450
"""Test that thread-local state is respected."""
451
451
pm = spawn_procs_on_this_host (v1 , {"gpus" : 1 })
@@ -473,7 +473,7 @@ async def no_more(self) -> None:
473
473
474
474
475
475
@pytest .mark .parametrize ("v1" , [True , False ])
476
- @pytest .mark .timeout (60 )
476
+ @pytest .mark .timeout (360 )
477
477
async def test_async_concurrency (v1 : bool ):
478
478
"""Test that async endpoints will be processed concurrently."""
479
479
pm = spawn_procs_on_this_host (v1 , {})
@@ -611,7 +611,7 @@ def _handle_undeliverable_message(
611
611
612
612
613
613
@pytest .mark .parametrize ("v1" , [True , False ])
614
- @pytest .mark .timeout (60 )
614
+ @pytest .mark .timeout (360 )
615
615
async def test_actor_log_streaming (v1 : bool ) -> None :
616
616
# Save original file descriptors
617
617
original_stdout_fd = os .dup (1 ) # stdout
@@ -876,7 +876,7 @@ def _stream_logs(self) -> bool:
876
876
877
877
878
878
@pytest .mark .parametrize ("v1" , [True , False ])
879
- @pytest .mark .timeout (60 )
879
+ @pytest .mark .timeout (360 )
880
880
async def test_logging_option_defaults (v1 : bool ) -> None :
881
881
# Save original file descriptors
882
882
original_stdout_fd = os .dup (1 ) # stdout
@@ -1155,7 +1155,7 @@ async def test_flush_logs_fast_exit() -> None:
1155
1155
1156
1156
1157
1157
@pytest .mark .parametrize ("v1" , [True , False ])
1158
- @pytest .mark .timeout (60 )
1158
+ @pytest .mark .timeout (360 )
1159
1159
async def test_flush_on_disable_aggregation (v1 : bool ) -> None :
1160
1160
"""Test that logs are flushed when disabling aggregation.
1161
1161
@@ -1280,7 +1280,7 @@ async def test_multiple_ongoing_flushes_no_deadlock(v1: bool) -> None:
1280
1280
1281
1281
1282
1282
@pytest .mark .parametrize ("v1" , [True , False ])
1283
- @pytest .mark .timeout (60 )
1283
+ @pytest .mark .timeout (360 )
1284
1284
async def test_adjust_aggregation_window (v1 : bool ) -> None :
1285
1285
"""Test that the flush deadline is updated when the aggregation window is adjusted.
1286
1286
@@ -1371,7 +1371,7 @@ async def send(self, port: Port[int]):
1371
1371
1372
1372
1373
1373
@pytest .mark .parametrize ("v1" , [True , False ])
1374
- @pytest .mark .timeout (60 )
1374
+ @pytest .mark .timeout (360 )
1375
1375
def test_port_as_argument (v1 : bool ) -> None :
1376
1376
proc_mesh = spawn_procs_on_fake_host (v1 , {"gpus" : 1 })
1377
1377
s = proc_mesh .spawn ("send_alot" , SendAlot )
@@ -1483,7 +1483,7 @@ def add(self, port: "Port[int]", b: int) -> None:
1483
1483
1484
1484
1485
1485
@pytest .mark .parametrize ("v1" , [True , False ])
1486
- @pytest .mark .timeout (60 )
1486
+ @pytest .mark .timeout (360 )
1487
1487
def test_ported_actor (v1 : bool ):
1488
1488
proc_mesh = spawn_procs_on_fake_host (v1 , {"gpus" : 1 })
1489
1489
a = proc_mesh .spawn ("port_actor" , PortedActor )
@@ -1499,7 +1499,7 @@ async def consume():
1499
1499
assert r == (7 , 2 , 3 )
1500
1500
1501
1501
1502
- @pytest .mark .timeout (60 )
1502
+ @pytest .mark .timeout (360 )
1503
1503
def test_python_task_tuple () -> None :
1504
1504
PythonTask .from_coroutine (consume ()).block_on ()
1505
1505
@@ -1568,7 +1568,7 @@ def _handle_undeliverable_message(
1568
1568
1569
1569
1570
1570
@pytest .mark .parametrize ("v1" , [True , False ])
1571
- @pytest .mark .timeout (60 )
1571
+ @pytest .mark .timeout (360 )
1572
1572
async def test_undeliverable_message_with_override (v1 : bool ) -> None :
1573
1573
pm = spawn_procs_on_this_host (v1 , {"gpus" : 1 })
1574
1574
receiver = pm .spawn ("undeliverable_receiver" , UndeliverableMessageReceiver )
@@ -1584,7 +1584,7 @@ async def test_undeliverable_message_with_override(v1: bool) -> None:
1584
1584
1585
1585
1586
1586
@pytest .mark .parametrize ("v1" , [True , False ])
1587
- @pytest .mark .timeout (60 )
1587
+ @pytest .mark .timeout (360 )
1588
1588
async def test_undeliverable_message_without_override (v1 : bool ) -> None :
1589
1589
pm = spawn_procs_on_this_host (v1 , {"gpus" : 1 })
1590
1590
sender = pm .spawn ("undeliverable_sender" , UndeliverableMessageSender )
@@ -1659,7 +1659,7 @@ async def spawning_from_endpoint(self, name, root) -> None:
1659
1659
await get_or_spawn_controller (name , SpawningActorFromEndpointActor , root = root )
1660
1660
1661
1661
1662
- @pytest .mark .timeout (60 )
1662
+ @pytest .mark .timeout (360 )
1663
1663
def test_get_or_spawn_controller_inside_actor_endpoint ():
1664
1664
actor_1 = get_or_spawn_controller ("actor_1" , SpawningActorFromEndpointActor ).get ()
1665
1665
actor_1 .spawning_from_endpoint .call_one ("actor_2" , root = "actor_1" ).get ()
0 commit comments