Skip to content

Commit 03687e7

Browse files
authored
Add repo structure description (#309)
* Update * update * Update README.md * update
1 parent b4b4623 commit 03687e7

File tree

4 files changed

+26
-7
lines changed

4 files changed

+26
-7
lines changed

README.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -11,11 +11,11 @@
1111

1212
**GraphNet** is a large-scale dataset of deep learning **computation graphs**, built as a standard benchmark for **tensor compiler** optimization. It provides over 2.7K computation graphs extracted from state-of-the-art deep learning models spanning diverse tasks and ML frameworks. With standardized formats and rich metadata, GraphNet enables fair comparison and reproducible evaluation of the general optimization capabilities of tensor compilers, thereby supporting advanced research such as AI for System on compilers.
1313

14-
## News
14+
## 📣 News
1515
- [2025-10-14] ✨ Our technical report is out: a detailed study of dataset construction and compiler benchmarking, introducing the novel performance metrics Speedup Score S(t) and Error-aware Speedup Score ES(t). [📘 GraphNet: A Large-Scale Computational Graph Dataset for Tensor Compiler Research](./GraphNet_technical_report.pdf)
1616
- [2025-8-20] 🚀 The second round of [open contribution tasks](https://github.com/PaddlePaddle/Paddle/issues/74773) was released. (completed ✅)
1717
- [2025-7-30] 🚀 The first round of [open contribution tasks](https://github.com/PaddlePaddle/GraphNet/issues/44) was released. (completed ✅)
18-
## Benchmark Results
18+
## 📊 Benchmark Results
1919
We evaluate two representative tensor compiler backends, CINN (PaddlePaddle) and TorchInductor (PyTorch), on GraphNet's NLP and CV subsets. The evaluation adopts two quantitative metrics proposed in the [Technical Report](./GraphNet_technical_report.pdf):
2020
- **Speedup Score** S(t) — evaluates compiler performance under varying numerical tolerance levels.
2121
<div align="center">
@@ -28,7 +28,7 @@ We evaluate two representative tensor compiler backends, CINN (PaddlePaddle) and
2828

2929
</div>
3030

31-
## Quick Start
31+
## Quick Start
3232
This section shows how to evaluate tensor compilers and reproduce benchmark results (for compiler users and developers),
3333
as well as how to contribute new computation graphs (for GraphNet contributors).
3434

@@ -97,12 +97,12 @@ python -m graph_net.plot_violin \
9797

9898
The scripts are designed to process a file structure as `/benchmark_path/category_name/`, and items on x-axis are identified by name of the sub-directories. After executing, several summary plots of result in categories (model tasks, libraries...) will be exported to `$GRAPH_NET_BENCHMARK_PATH`.
9999

100-
## 🧱 Construction & Contribution Guide
100+
### 🧱 Construction & Contribution Guide
101101
Want to understand how GraphNet is built or contribute new samples?
102102
Check out the [Construction Guide](./docs/README_contribute.md) for details on the extraction and validation workflow.
103103

104104

105-
## Future Roadmap
105+
## 🚀 Future Roadmap
106106

107107
1. Scale GraphNet to 10K+ graphs.
108108
2. Further annotate GraphNet samples into more granular sub-categories
@@ -136,7 +136,7 @@ GraphNet is released under the [MIT License](./LICENSE).
136136
If you find this project helpful, please cite:
137137

138138
```bibtex
139-
@article{li2025graphnet,
139+
@misc{li2025graphnet,
140140
title = {GraphNet: A Large-Scale Computational Graph Dataset for Tensor Compiler Research},
141141
author = {Xinqi Li and Yiqun Liu and Shan Jiang and Enrong Zheng and Huaijin Zheng and Wenhao Dai and Haodong Deng and Dianhai Yu and Yanjun Ma},
142142
year = {2025},
File renamed without changes.
File renamed without changes.

docs/README_contribute.md

Lines changed: 20 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -65,4 +65,23 @@ python -m graph_net.torch.validate \
6565
--model-path $GRAPH_NET_EXTRACT_WORKSPACE/model_name
6666
```
6767

68-
All the **construction constraints** will be examined automatically. After passing validation, a unique `graph_hash.txt` will be generated and later checked in CI procedure to avoid redundant.
68+
All the **construction constraints** will be examined automatically. After passing validation, a unique `graph_hash.txt` will be generated and later checked in CI procedure to avoid redundant.
69+
70+
## 📁 Repository Structure
71+
This repository is organized as follows:
72+
73+
| Directory | Description |
74+
|------------|--------------|
75+
| **graph_net/** | Core module for graph extraction, validation, and benchmarking |
76+
| **paddle_samples/** | Computation graph samples extracted from PaddlePaddle |
77+
| **samples/** | Computation graph samples extracted from PyTorch |
78+
| **docs/** | Technical documents and contributor guides|
79+
80+
Below is the structure of the **graph_net/**:
81+
```text
82+
graph_net/
83+
├─ config/ # Config files, params
84+
├─ paddle/ # PaddlePaddle graph extraction & validation
85+
├─ torch/ # PyTorch graph extraction & validation
86+
├─ test/ # Unit tests and example scripts
87+
└─ *.py # Benchmark & analysis scripts

0 commit comments

Comments
 (0)