You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
- Multiple [strategies](https://github.com/NUS-HPC-AI-Lab/DD-Ranking/tree/main/dd_ranking/loss) of using soft labels;
92
-
- Data augmentation, reconsidered as [optional tricks](https://github.com/NUS-HPC-AI-Lab/DD-Ranking/tree/main/dd_ranking/aug) in DD;
93
-
- Commonly used [model architectures](https://github.com/NUS-HPC-AI-Lab/DD-Ranking/blob/main/dd_ranking/utils/networks.py) in DD.
94
-
A new ranking on representative DD methods.
95
-
96
-
DD-Ranking is flexible and easy to use, supported by:
97
-
<!-- Defualt configs: Customized configs -->
98
-
<!-- Integrated classes: 1) Optimizer and etc.; 2) random selection tests (additionally, w/ or w/o hard labels)-->
99
-
- Extensive configs provided;
100
-
- Cutomized configs;
101
-
- Testing and training framework with integrated metrics.
91
+
- Multiple [strategies](https://github.com/NUS-HPC-AI-Lab/DD-Ranking/tree/main/dd_ranking/loss) of using soft labels in existing works;
92
+
- Commonly used [data augmentation](https://github.com/NUS-HPC-AI-Lab/DD-Ranking/tree/main/dd_ranking/aug) methods in existing works;
93
+
- Commonly used [model architectures](https://github.com/NUS-HPC-AI-Lab/DD-Ranking/blob/main/dd_ranking/utils/networks.py) in existing works.
94
+
95
+
DD-Ranking has the following features:
96
+
-**Fair Evaluation**: DD-Ranking provides a fair evaluation scheme for DD methods that can decouple the impacts from knowledge distillation and data augmentation to reflect the real informativeness of the distilled data.
97
+
-**Easy-to-use**: DD-Ranking provides a unified interface for dataset distillation evaluation.
98
+
-**Extensible**: DD-Ranking supports various datasets and models.
99
+
-**Customizable**: DD-Ranking supports various data augmentations and soft label strategies.
102
100
103
101
</details>
104
102
105
103
## Overview
106
-
Included datasets and methods (hard/soft label).
107
-
|Dataset|Hard Label|Soft Label|
104
+
Included datasets and methods (categorized by hard/soft label).
105
+
|Supported Dataset|Evaluated Hard Label Methods|Evaluated Soft Label Methods|
108
106
|:-|:-|:-|
109
107
|CIFAR-10|DC|DATM|
110
108
|CIFAR-100|DSA|SRe2L|
111
109
|TinyImageNet|DM|RDED|
112
110
||MTT|D4M|
113
111
114
-
## Coming Soon
115
-
Rank on different data augmentation methods.
116
-
Rank on different data augmentation methods.
112
+
Evaluation results can be found in the [leaderboard](https://huggingface.co/spaces/Soptq/DD-Ranking).
113
+
117
114
## Tutorial
118
115
119
116
Install DD-Ranking with `pip` or from [source](https://github.com/NUS-HPC-AI-Lab/DD-Ranking/tree/main):
@@ -221,6 +218,10 @@ The following results will be returned to you:
221
218
- [Quickstart]()
222
219
- [Supported Models]() -->
223
220
221
+
## Coming Soon
222
+
-[ ] DD-Ranking scores that decouple the impacts from data augmentation.
223
+
-[ ] Evaluation results on ImageNet subsets.
224
+
224
225
## Contributing
225
226
226
227
<!-- Only PR for the 1st version of DD-Ranking -->
Copy file name to clipboardExpand all lines: doc/introduction.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -8,7 +8,7 @@ Dataset Distillation (DD) aims to condense a large dataset into a much smaller o
8
8
9
9

10
10
11
-
Notebaly, more and more methods are transitting from "hard label" to "soft label" in dataset distillation, especially during evaluation. **Hard labels** are categorical, having the same format of the real dataset. **Soft labels** are distributions, typically generated by a pre-trained teacher model.
11
+
Notebaly, more and more methods are transitting from "hard label" to "soft label" in dataset distillation, especially during evaluation. **Hard labels** are categorical, having the same format of the real dataset. **Soft labels** are outputs of a pre-trained teacher model.
12
12
Recently, Deng et al., pointed out that "a label is worth a thousand images". They showed analytically that soft labels are exetremely useful for accuracy improvement.
13
13
14
14
However, since the essence of soft labels is **knowledge distillation**, we find that when applying the same evaluation method to randomly selected data, the test accuracy also improves significantly (see the figure above).
0 commit comments