Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Distilling Virtual Examples for Long-tailed Recognition

About

We tackle the long-tailed visual recognition problem from the knowledge distillation perspective by proposing a Distill the Virtual Examples (DiVE) method. Specifically, by treating the predictions of a teacher model as virtual examples, we prove that distilling from these virtual examples is equivalent to label distribution learning under certain constraints. We show that when the virtual example distribution becomes flatter than the original input distribution, the under-represented tail classes will receive significant improvements, which is crucial in long-tailed recognition. The proposed DiVE method can explicitly tune the virtual example distribution to become flat. Extensive experiments on three benchmark datasets, including the large-scale iNaturalist ones, justify that the proposed DiVE method can significantly outperform state-of-the-art methods. Furthermore, additional analyses and experiments verify the virtual example interpretation, and demonstrate the effectiveness of tailored designs in DiVE for long-tailed problems.

Yin-Yin He, Jianxin Wu, Xiu-Shen Wei• 2021

Related benchmarks

TaskDatasetResultRank
Image ClassificationImageNet LT
Top-1 Accuracy53.1
251
Long-Tailed Image ClassificationImageNet-LT (test)
Top-1 Acc (Overall)53.1
220
Image ClassificationiNaturalist 2018 (test)
Top-1 Accuracy71.71
192
Image ClassificationImageNet-LT (test)
Top-1 Acc (All)53.1
159
Image ClassificationiNaturalist 2018 (val)--
116
Long-tailed Visual RecognitionImageNet LT
Overall Accuracy53.1
89
Image ClassificationCIFAR-100-LT IF 100 (test)
Top-1 Acc45.4
77
Image ClassificationImageNet-LT (val)
Top-1 Acc (Total)57.12
72
Image ClassificationCIFAR-100 Imbalance Ratio LT-50 (test)
Accuracy51.1
62
Image ClassificationCIFAR-100-LT Imbalance Ratio 100 (test)
Accuracy45.4
62
Showing 10 of 33 rows

Other info

Follow for update