You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
# Note: if --compiler is omitted, PyTorch’s built-in compiler is used by default
99
106
```
100
107
101
-
After executing, ```graph_net.torch.test_compiler``` will:
108
+
After executing, `graph_net.torch.test_compiler` will:
102
109
1. Running the original model in eager mode to record a baseline.
103
110
2. Compiling the model with the specified backend (e.g., CINN, TVM, Inductor, TensorRT, XLA, BladeDISC).
104
111
3. Executing the compiled model and collecting its runtime and outputs.
105
112
4. Conduct speedup by comparing the compiled results against the baseline.
106
113
107
114
**Step 2: Analysis**
108
115
109
-
After processing, we provide ```graph_net/analysis.py``` to generate [violin plot](https://en.m.wikipedia.org/wiki/Violin_plot) based on the JSON results.
116
+
After processing, we provide `graph_net/analysis.py` to generate [violin plot](https://en.m.wikipedia.org/wiki/Violin_plot) based on the JSON results.
After executing, one summary plot of results on all compilers, as well as multiple sub-plots of results in categories (model tasks, Library...) on a single compiler will be exported.
118
125
119
-
The script is designed to process a file structure as ```/benchmark_path/compiler_name/category_name/``` (for example ```/benchmark_logs/paddle/nlp/```), and items on x-axis are identified by name of the folders. So you can modify ```read_all_speedups``` function to fit the benchmark settings on your demand.
126
+
The script is designed to process a file structure as `/benchmark_path/compiler_name/category_name/` (for example `/benchmark_logs/paddle/nlp/`), and items on x-axis are identified by name of the folders. So you can modify `read_all_speedups` function to fit the benchmark settings on your demand.
120
127
121
128
## 📌 Roadmap
122
129
@@ -125,9 +132,7 @@ The script is designed to process a file structure as ```/benchmark_path/compile
125
132
3. Extract samples from multi-GPU scenarios to support benchmarking and optimization for large-scale, distributed computing.
126
133
4. Enable splitting full graphs into independently optimized subgraphs and operator sequences.
127
134
128
-
**Vision**: GraphNet aims to lay the foundation for [ai4c](https://github.com/PaddlePaddle/ai4c) by enabling large-scale, systematic evaluation of tensor compiler optimizations.
129
-
130
-
We aim to achieve cross-hardware portability of compiler optimizations by allowing models to learn and transfer optimization strategies. It will significantly reduce the manual effort required to develop efficient operator implementations.
135
+
**Vision**: GraphNet aims to lay the foundation for [ai4c](https://github.com/PaddlePaddle/ai4c) by **enabling large-scale, systematic evaluation** of tensor compiler optimizations, and providing a dataset for **models to learn and transfer optimization strategies**.
0 commit comments