You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+6-6Lines changed: 6 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -11,11 +11,11 @@
11
11
12
12
**GraphNet** is a large-scale dataset of deep learning **computation graphs**, built as a standard benchmark for **tensor compiler** optimization. It provides over 2.7K computation graphs extracted from state-of-the-art deep learning models spanning diverse tasks and ML frameworks. With standardized formats and rich metadata, GraphNet enables fair comparison and reproducible evaluation of the general optimization capabilities of tensor compilers, thereby supporting advanced research such as AI for System on compilers.
13
13
14
-
## News
14
+
## 📣 News
15
15
-[2025-10-14] ✨ Our technical report is out: a detailed study of dataset construction and compiler benchmarking, introducing the novel performance metrics Speedup Score S(t) and Error-aware Speedup Score ES(t). [📘 GraphNet: A Large-Scale Computational Graph Dataset for Tensor Compiler Research](./GraphNet_technical_report.pdf)
16
16
-[2025-8-20] 🚀 The second round of [open contribution tasks](https://github.com/PaddlePaddle/Paddle/issues/74773) was released. (completed ✅)
17
17
-[2025-7-30] 🚀 The first round of [open contribution tasks](https://github.com/PaddlePaddle/GraphNet/issues/44) was released. (completed ✅)
18
-
## Benchmark Results
18
+
## 📊 Benchmark Results
19
19
We evaluate two representative tensor compiler backends, CINN (PaddlePaddle) and TorchInductor (PyTorch), on GraphNet's NLP and CV subsets. The evaluation adopts two quantitative metrics proposed in the [Technical Report](./GraphNet_technical_report.pdf):
The scripts are designed to process a file structure as `/benchmark_path/category_name/`, and items on x-axis are identified by name of the sub-directories. After executing, several summary plots of result in categories (model tasks, libraries...) will be exported to `$GRAPH_NET_BENCHMARK_PATH`.
93
93
94
-
## 🧱 Construction & Contribution Guide
94
+
###🧱 Construction & Contribution Guide
95
95
Want to understand how GraphNet is built or contribute new samples?
96
96
Check out the [Construction Guide](./docs/README_contribute.md) for details on the extraction and validation workflow.
97
97
98
98
99
-
## Future Roadmap
99
+
## 🚀 Future Roadmap
100
100
101
101
1. Scale GraphNet to 10K+ graphs.
102
102
2. Further annotate GraphNet samples into more granular sub-categories
@@ -130,7 +130,7 @@ GraphNet is released under the [MIT License](./LICENSE).
130
130
If you find this project helpful, please cite:
131
131
132
132
```bibtex
133
-
@article{li2025graphnet,
133
+
@misc{li2025graphnet,
134
134
title = {GraphNet: A Large-Scale Computational Graph Dataset for Tensor Compiler Research},
135
135
author = {Xinqi Li and Yiqun Liu and Shan Jiang and Enrong Zheng and Huaijin Zheng and Wenhao Dai and Haodong Deng and Dianhai Yu and Yanjun Ma},
All the **construction constraints** will be examined automatically. After passing validation, a unique `graph_hash.txt` will be generated and later checked in CI procedure to avoid redundant.
68
+
All the **construction constraints** will be examined automatically. After passing validation, a unique `graph_hash.txt` will be generated and later checked in CI procedure to avoid redundant.
69
+
70
+
## 📁 Repository Structure
71
+
This repository is organized as follows:
72
+
73
+
| Directory | Description |
74
+
|------------|--------------|
75
+
|**graph_net/**| Core module for graph extraction, validation, and benchmarking |
76
+
|**paddle_samples/**| Computation graph samples extracted from PaddlePaddle |
77
+
|**samples/**| Computation graph samples extracted from PyTorch |
78
+
|**docs/**| Technical documents and contributor guides|
0 commit comments