You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This PR removes the CUDA specific code from the `benchmarks` directory.
This is in line with the CUDA deprecation that started on release 2.8.
**Key Changes:**
- Removed the `keep_model_data_on_cuda` parameter
- Used in combination with zero-overhead CUDA to XLA:CUDA data movement,
removed in [#9598][1] and [#9603][2]
- Deleted `llama.py`, `nightly.sh`, `run_benchmark.sh`,
`run_single_graph_bm.sh`, and `run_top_tier_bm.sh`
- All of them ran benchmarks comparing PyTorch Inductor with XLA:CUDA,
specifically
[1]: #9598
[2]: #9603
Copy file name to clipboardExpand all lines: benchmarks/experiment_runner.py
-5Lines changed: 0 additions & 5 deletions
Original file line number
Diff line number
Diff line change
@@ -936,11 +936,6 @@ def __str__(self):
936
936
help="""Collect CUDA and CPU times per operation. This will also gather
937
937
CPU fallbacks.""",
938
938
)
939
-
parser.add_argument(
940
-
"--keep-model-data-on-cuda",
941
-
action="store_true",
942
-
help="""Whether to keep the model and data on CUDA and not to move to an XLA device. This is to be used with PyTorch/XLA dynamo. When set, PyTorch/XLA dynamo bridge move the model and data to the XLA device.""",
0 commit comments