You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+12-10Lines changed: 12 additions & 10 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -65,13 +65,13 @@ DD-Ranking (DD, *i.e.*, Dataset Distillation) is an integrated and easy-to-use b
65
65
66
66
<!-- Hard label is tested -->
67
67
<!-- Keep the same compression ratio, comparing with random selection -->
68
-
**Performance benchmark**
68
+
### Benchmark
69
69
70
70
Revisit the original goal of dataset distillation:
71
-
> The idea is to synthesize a small number of data points that do not need to come from the correct data distribution, but will, when given to the learning algorithm as training data, approximate the model trained on the original data.
71
+
> The idea is to synthesize a small number of data points that do not need to come from the correct data distribution, but will, when given to the learning algorithm as training data, approximate the model trained on the original data. (Wang et al., 2020)
72
72
>
73
73
74
-
The evaluation method for DD-Ranking is grounded in the essence of dataset distillation, aiming to better reflect the information content of the synthesized data by assessing the following two aspects:
74
+
The evaluation method for DD-Ranking is grounded in the essence of dataset distillation, aiming to better reflect the informativeness of the synthesized data by assessing the following two aspects:
75
75
1. The degree to which the original dataset is recovered under hard labels (hard label recovery): $\text{HLR}=\text{Acc.}{\text{full-hard}}-\text{Acc.}{\text{syn-hard}}$.
76
76
77
77
2. The improvement over random selection when using personalized evaluation methods (improvement over random): $\text{IOR}=\text{Acc.}{\text{syn-any}}-\text{Acc.}{\text{rdm-any}}$.
@@ -81,9 +81,13 @@ $\text{Acc.}$ is the accuracy of models trained on different samples. Samples' m
81
81
- $\text{syn-any}$: Synthetic dataset with personalized evaluation methods (hard or soft labels);
82
82
- $\text{rdm-any}$: Randomly selected dataset (under the same compression ratio) with the same personalized evaluation methods.
83
83
84
-
To rank different methods, we combine the above two metrics as follows:
84
+
<!--To rank different methods, we combine the above two metrics as follows:
@@ -98,18 +102,15 @@ DD-Ranking has the following features:
98
102
-**Extensible**: DD-Ranking supports various datasets and models.
99
103
-**Customizable**: DD-Ranking supports various data augmentations and soft label strategies.
100
104
101
-
</details>
102
-
103
-
## Overview
104
-
Included datasets and methods (categorized by hard/soft label).
105
+
DD-Ranking currently includes the following datasets and methods (categorized by hard/soft label). Evaluation results can be found in the [leaderboard](https://huggingface.co/spaces/Soptq/DD-Ranking).
105
106
|Supported Dataset|Evaluated Hard Label Methods|Evaluated Soft Label Methods|
106
107
|:-|:-|:-|
107
108
|CIFAR-10|DC|DATM|
108
109
|CIFAR-100|DSA|SRe2L|
109
110
|TinyImageNet|DM|RDED|
110
111
||MTT|D4M|
111
112
112
-
Evaluation results can be found in the [leaderboard](https://huggingface.co/spaces/Soptq/DD-Ranking).
113
+
113
114
114
115
## Tutorial
115
116
@@ -221,6 +222,7 @@ The following results will be returned to you:
221
222
## Coming Soon
222
223
-[ ] DD-Ranking scores that decouple the impacts from data augmentation.
Copy file name to clipboardExpand all lines: doc/metrics/hard-label.md
+1-3Lines changed: 1 addition & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -71,7 +71,7 @@ This method computes the HLR, IOR, and DD-Ranking scores for the given image and
71
71
1. Compute the test accuracy of the surrogate model on the synthetic dataset under hard labels. We tune the learning rate for the best performance if `syn_lr` is not provided.
72
72
2. Compute the test accuracy of the surrogate model on the real dataset under the same setting as step 1.
73
73
3. Compute the test accuracy of the surrogate model on the randomly selected dataset under the same setting as step 1.
74
-
4. Compute the HLR, IOR, and DD-Ranking scores.
74
+
4. Compute the HLRand IOR scores.
75
75
76
76
The final scores are the average of the scores from `num_eval` rounds.
77
77
@@ -90,8 +90,6 @@ A dictionary with the following keys:
90
90
-**hard_label_recovery_std**: Standard deviation of HLR scores from `num_eval` rounds.
91
91
-**improvement_over_random_mean**: Mean of improvement over random scores from `num_eval` rounds.
92
92
-**improvement_over_random_std**: Standard deviation of improvement over random scores from `num_eval` rounds.
93
-
-**dd_ranking_mean**: Mean of DD-Ranking scores from `num_eval` rounds.
94
-
-**dd_ranking_std**: Standard deviation of DD-Ranking scores from `num_eval` rounds.
Copy file name to clipboardExpand all lines: doc/metrics/soft-label.md
+4-3Lines changed: 4 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -50,6 +50,7 @@ A class for evaluating the performance of a dataset distillation method with sof
50
50
-**temperature**(<spanstyle="color:#FF6B00;">float</span>): Temperature for knowledge distillation.
51
51
-**data_aug_func**(<spanstyle="color:#FF6B00;">str</span>): Data augmentation function used during training. Currently supports `dsa`, `cutmix`, `mixup`. See [augmentations](../augmentations/overview.md) for more details.
52
52
-**aug_params**(<spanstyle="color:#FF6B00;">dict</span>): Parameters for the data augmentation function.
53
+
-**use_aug_for_hard**(<spanstyle="color:#FF6B00;">bool</span>): Whether to use the data augmentation specified in `data_aug_func` for hard label evaluation.
53
54
-**optimizer**(<spanstyle="color:#FF6B00;">str</span>): Name of the optimizer. Currently supports torch-based optimizers - `sgd`, `adam`, and `adamw`.
54
55
-**lr_scheduler**(<spanstyle="color:#FF6B00;">str</span>): Name of the learning rate scheduler. Currently supports torch-based schedulers - `step`, `cosine`, `lambda_step`, and `lambda_cos`.
55
56
-**weight_decay**(<spanstyle="color:#FF6B00;">float</span>): Weight decay for the optimizer.
@@ -83,7 +84,7 @@ This method computes the HLR, IOR, and DD-Ranking scores for the given image and
83
84
2. Compute the test accuracy of the surrogate model on the real dataset under the same setting as step 1.
84
85
3. Compute the test accuracy of the surrogate model on the synthetic dataset under soft labels.
85
86
4. Compute the test accuracy of the surrogate model on the randomly selected dataset under the same setting as step 3.
86
-
5. Compute the HLR, IOR, and DD-Ranking scores.
87
+
5. Compute the HLRand IOR scores.
87
88
88
89
The final scores are the average of the scores from `num_eval` rounds.
89
90
@@ -102,8 +103,8 @@ A dictionary with the following keys:
102
103
-**hard_label_recovery_std**: Standard deviation of HLR scores from `num_eval` rounds.
103
104
-**improvement_over_random_mean**: Mean of improvement over random scores from `num_eval` rounds.
104
105
-**improvement_over_random_std**: Standard deviation of improvement over random scores from `num_eval` rounds.
105
-
-**dd_ranking_mean**: Mean of DD-Ranking scores from `num_eval` rounds.
106
-
-**dd_ranking_std**: Standard deviation of DD-Ranking scores from `num_eval` rounds.
106
+
<!--- **dd_ranking_mean**: Mean of DD-Ranking scores from `num_eval` rounds.
107
+
- **dd_ranking_std**: Standard deviation of DD-Ranking scores from `num_eval` rounds.-->
0 commit comments