You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+6-6Lines changed: 6 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -11,11 +11,11 @@
11
11
12
12
**GraphNet** is a large-scale dataset of deep learning **computation graphs**, built as a standard benchmark for **tensor compiler** optimization. It provides over 2.7K computation graphs extracted from state-of-the-art deep learning models spanning diverse tasks and ML frameworks. With standardized formats and rich metadata, GraphNet enables fair comparison and reproducible evaluation of the general optimization capabilities of tensor compilers, thereby supporting advanced research such as AI for System on compilers.
13
13
14
-
## News
14
+
## 📣 News
15
15
-[2025-10-14] ✨ Our technical report is out: a detailed study of dataset construction and compiler benchmarking, introducing the novel performance metrics Speedup Score S(t) and Error-aware Speedup Score ES(t). [📘 GraphNet: A Large-Scale Computational Graph Dataset for Tensor Compiler Research](./GraphNet_technical_report.pdf)
16
16
-[2025-8-20] 🚀 The second round of [open contribution tasks](https://github.com/PaddlePaddle/Paddle/issues/74773) was released. (completed ✅)
17
17
-[2025-7-30] 🚀 The first round of [open contribution tasks](https://github.com/PaddlePaddle/GraphNet/issues/44) was released. (completed ✅)
18
-
## Benchmark Results
18
+
## 📊 Benchmark Results
19
19
We evaluate two representative tensor compiler backends, CINN (PaddlePaddle) and TorchInductor (PyTorch), on GraphNet's NLP and CV subsets. The evaluation adopts two quantitative metrics proposed in the [Technical Report](./GraphNet_technical_report.pdf):
The scripts are designed to process a file structure as `/benchmark_path/category_name/`, and items on x-axis are identified by name of the sub-directories. After executing, several summary plots of result in categories (model tasks, libraries...) will be exported to `$GRAPH_NET_BENCHMARK_PATH`.
99
99
100
-
## 🧱 Construction & Contribution Guide
100
+
###🧱 Construction & Contribution Guide
101
101
Want to understand how GraphNet is built or contribute new samples?
102
102
Check out the [Construction Guide](./docs/README_contribute.md) for details on the extraction and validation workflow.
103
103
104
104
105
-
## Future Roadmap
105
+
## 🚀 Future Roadmap
106
106
107
107
1. Scale GraphNet to 10K+ graphs.
108
108
2. Further annotate GraphNet samples into more granular sub-categories
@@ -136,7 +136,7 @@ GraphNet is released under the [MIT License](./LICENSE).
136
136
If you find this project helpful, please cite:
137
137
138
138
```bibtex
139
-
@article{li2025graphnet,
139
+
@misc{li2025graphnet,
140
140
title = {GraphNet: A Large-Scale Computational Graph Dataset for Tensor Compiler Research},
141
141
author = {Xinqi Li and Yiqun Liu and Shan Jiang and Enrong Zheng and Huaijin Zheng and Wenhao Dai and Haodong Deng and Dianhai Yu and Yanjun Ma},
All the **construction constraints** will be examined automatically. After passing validation, a unique `graph_hash.txt` will be generated and later checked in CI procedure to avoid redundant.
68
+
All the **construction constraints** will be examined automatically. After passing validation, a unique `graph_hash.txt` will be generated and later checked in CI procedure to avoid redundant.
69
+
70
+
## 📁 Repository Structure
71
+
This repository is organized as follows:
72
+
73
+
| Directory | Description |
74
+
|------------|--------------|
75
+
|**graph_net/**| Core module for graph extraction, validation, and benchmarking |
76
+
|**paddle_samples/**| Computation graph samples extracted from PaddlePaddle |
77
+
|**samples/**| Computation graph samples extracted from PyTorch |
78
+
|**docs/**| Technical documents and contributor guides|
0 commit comments