|
52 | 52 | retry_on_connect_failures, |
53 | 53 | run_tests, |
54 | 54 | skip_but_pass_in_sandcastle, |
| 55 | + skipIfRocm, |
55 | 56 | TestCase, |
56 | 57 | ) |
57 | 58 |
|
@@ -385,6 +386,7 @@ def test_broadcast_stress(self): |
385 | 386 | inputs = [torch.tensor([i * self.world_size + self.rank]) for i in range(1000)] |
386 | 387 | self._test_broadcast_stress(inputs) |
387 | 388 |
|
| 389 | + @skipIfRocm |
388 | 390 | @skip_if_lt_x_gpu(2) |
389 | 391 | @requires_gloo() |
390 | 392 | def test_broadcast_stress_cuda(self): |
@@ -490,6 +492,7 @@ def test_allreduce_stress(self): |
490 | 492 | inputs = [torch.tensor([i + self.rank]) for i in range(1000)] |
491 | 493 | self._test_allreduce_stress(inputs) |
492 | 494 |
|
| 495 | + @skipIfRocm |
493 | 496 | @skip_if_lt_x_gpu(2) |
494 | 497 | @requires_gloo() |
495 | 498 | def test_allreduce_stress_cuda(self): |
@@ -922,6 +925,8 @@ def test_scatter_stress(self): |
922 | 925 | @skip_but_pass_in_sandcastle( |
923 | 926 | "Test is flaky, see https://github.com/pytorch/pytorch/issues/15963" |
924 | 927 | ) |
| 928 | + |
| 929 | + @skipIfRocm |
925 | 930 | @skip_if_lt_x_gpu(2) |
926 | 931 | @requires_gloo() |
927 | 932 | def test_scatter_stress_cuda(self): |
@@ -1096,6 +1101,7 @@ def test_gather_stress(self): |
1096 | 1101 | inputs = [torch.tensor([i + self.rank]) for i in range(1000)] |
1097 | 1102 | self._test_gather_stress(inputs, lambda t: t.clone()) |
1098 | 1103 |
|
| 1104 | + @skipIfRocm |
1099 | 1105 | @skip_if_lt_x_gpu(2) |
1100 | 1106 | @requires_gloo() |
1101 | 1107 | def test_gather_stress_cuda(self): |
@@ -1231,6 +1237,7 @@ def test_allgather_stress(self): |
1231 | 1237 | inputs = [torch.tensor([i + self.rank]) for i in range(1000)] |
1232 | 1238 | self._test_allgather_stress(inputs, lambda t: t.clone()) |
1233 | 1239 |
|
| 1240 | + @skipIfRocm |
1234 | 1241 | @skip_if_lt_x_gpu(2) |
1235 | 1242 | @requires_gloo() |
1236 | 1243 | def test_allgather_stress_cuda(self): |
@@ -1417,6 +1424,7 @@ def test_reduce_stress(self): |
1417 | 1424 | inputs = [torch.tensor([i + self.rank]) for i in range(1000)] |
1418 | 1425 | self._test_reduce_stress(inputs) |
1419 | 1426 |
|
| 1427 | + @skipIfRocm |
1420 | 1428 | @skip_if_lt_x_gpu(2) |
1421 | 1429 | @requires_gloo() |
1422 | 1430 | def test_reduce_stress_cuda(self): |
|
0 commit comments