Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Distilling Long-tailed Datasets

About

Dataset distillation aims to synthesize a small, information-rich dataset from a large one for efficient model training. However, existing dataset distillation methods struggle with long-tailed datasets, which are prevalent in real-world scenarios. By investigating the reasons behind this unexpected result, we identified two main causes: 1) The distillation process on imbalanced datasets develops biased gradients, leading to the synthesis of similarly imbalanced distilled datasets. 2) The experts trained on such datasets perform suboptimally on tail classes, resulting in misguided distillation supervision and poor-quality soft-label initialization. To address these issues, we first propose Distribution-agnostic Matching to avoid directly matching the biased expert trajectories. It reduces the distance between the student and the biased expert trajectories and prevents the tail class bias from being distilled to the synthetic dataset. Moreover, we improve the distillation guidance with Expert Decoupling, which jointly matches the decoupled backbone and classifier to improve the tail class performance and initialize reliable soft labels. This work pioneers the field of long-tailed dataset distillation, marking the first effective effort to distill long-tailed datasets.

Zhenghao Zhao, Haoxuan Wang, Yuzhang Shang, Kai Wang, Yan Yan• 2024

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-10 Long Tailed Imbalance Ratio 50 (test)
Top-1 Accuracy65.8
57
Long-Tailed Image ClassificationCIFAR10-LT imbalance factor 100 (test)
Top-1 Accuracy64
46
Image ClassificationCIFAR-10 Imbalance Factor 10 Long-Tailed (test)
Accuracy70.5
30
Image ClassificationCIFAR-10 Imbalance Factor 200 Long-Tailed (test)
Accuracy62.3
28
Showing 4 of 4 rows

Other info

Code

Follow for update