Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Are Large-scale Soft Labels Necessary for Large-scale Dataset Distillation?

About

In ImageNet-condensation, the storage for auxiliary soft labels exceeds that of the condensed dataset by over 30 times. However, are large-scale soft labels necessary for large-scale dataset distillation? In this paper, we first discover that the high within-class similarity in condensed datasets necessitates the use of large-scale soft labels. This high within-class similarity can be attributed to the fact that previous methods use samples from different classes to construct a single batch for batch normalization (BN) matching. To reduce the within-class similarity, we introduce class-wise supervision during the image synthesizing process by batching the samples within classes, instead of across classes. As a result, we can increase within-class diversity and reduce the size of required soft labels. A key benefit of improved image diversity is that soft label compression can be achieved through simple random pruning, eliminating the need for complex rule-based strategies. Experiments validate our discoveries. For example, when condensing ImageNet-1K to 200 images per class, our approach compresses the required soft labels from 113 GB to 2.8 GB (40x compression) with a 2.6% performance gain. Code is available at: https://github.com/he-y/soft-label-pruning-for-dataset-distillation

Lingao Xiao, Yang He• 2024

Related benchmarks

TaskDatasetResultRank
Image ClassificationTinyImageNet (val)
Accuracy55.4
240
Image ClassificationTiny-ImageNet
Accuracy55.4
227
Image ClassificationImageNet-1k (val)
Accuracy39
189
Dataset DistillationImageNet-1k (val)
Accuracy57.7
64
Image ClassificationImageNet 1k (train)
Top-1 Accuracy67.8
58
Image ClassificationTiny-ImageNet 64x64, C=200 (val)
Top-1 Acc34.3
50
Dataset DistillationImageNet-21K-P
Accuracy31.3
10
Image ClassificationImageNet-1k (val)
Accuracy67.4
7
Showing 8 of 8 rows

Other info

Code

Follow for update