You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@@ -52,21 +54,10 @@ For hardware, we used a 96GB 700W H100 GPU. Some of the optimizations applied (B
52
54
53
55
## Run the optimized pipeline
54
56
55
-
```
56
-
python gen_image.py --prompt "An astronaut standing next to a giant lemon" --output-file output.png --use-cached-model
57
-
```
58
-
59
-
This will include all optimizations and will attempt to use pre-cached binary models
60
-
generated via `torch.export` + AOTI. To generate these binaries for subsequent runs, run
61
-
the above command without the `--use-cached-model` flag.
57
+
TODO
62
58
63
59
> [!IMPORTANT]
64
-
> The binaries won't work for hardware that is sufficiently different from the hardware they were
65
-
> obtained on. For example, if the binaries were obtained on an H100, they won't work on A100.
66
-
> Further, the binaries are currently Linux-only and include dependencies on specific versions
67
-
> of system libs such as libstdc++; they will not work if they were generated in a sufficiently
68
-
> different environment than the one present at runtime. The PyTorch Compiler team is working on
69
-
> solutions for more portable binaries / artifact caching.
60
+
> The binaries won't work for hardware that are different from the ones they were obtained on. For example, if the binaries were obtained on an H100, they won't work on A100.
70
61
71
62
## Benchmarking
72
63
[`run_benchmark.py`](./run_benchmark.py) is the main script for benchmarking the different optimization techniques.
0 commit comments