Skip to content

Commit 2005b5f

Browse files
IvanKobzarevpytorchmergebot
authored andcommitted
[inductor] Use runtime estimations in iterative reorder collectives pass (pytorch#167080)
Split of pytorch#162469 to be under 2K reorder iterative part Pull Request resolved: pytorch#167080 Approved by: https://github.com/eellison
1 parent b2d72a4 commit 2005b5f

File tree

5 files changed

+851
-357
lines changed

5 files changed

+851
-357
lines changed

test/distributed/test_inductor_collectives.py

Lines changed: 1 addition & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1985,6 +1985,7 @@ def _reorder_communication_preserving_peak_memory(
19851985
"bucket_reduce_scatters_fx_bucket_size_determinator": lambda _: 2,
19861986
"reorder_for_compute_comm_overlap": True,
19871987
"reorder_for_compute_comm_overlap_passes": [
1988+
_reorder_communication_preserving_peak_memory,
19881989
sink_waits_iterative,
19891990
_reorder_communication_preserving_peak_memory,
19901991
],
@@ -2046,11 +2047,6 @@ def _reorder_communication_preserving_peak_memory(
20462047
assert node_stats is not None
20472048
self.assertTrue(isinstance(node_stats, dict))
20482049
self.assertEqual(len(node_stats), 4)
2049-
it = iter(node_stats.values())
2050-
node_stat0 = next(it)
2051-
self.assertTrue(node_stat0.limiting_factor == "None")
2052-
node_stat1 = next(it)
2053-
self.assertTrue("collective ordering" in node_stat1.limiting_factor)
20542050

20552051
@skipIfXpu # https://github.com/intel/torch-xpu-ops/issues/1581
20562052
@unittest.skipIf(not HAS_GPU, "Inductor+gpu needs triton and recent GPU arch")

0 commit comments

Comments
 (0)