Long-Tailed Classification with Gradual Balanced Loss and Adaptive Feature Generation
About
The real-world data distribution is essentially long-tailed, which poses great challenge to the deep model. In this work, we propose a new method, Gradual Balanced Loss and Adaptive Feature Generator (GLAG) to alleviate imbalance. GLAG first learns a balanced and robust feature model with Gradual Balanced Loss, then fixes the feature model and augments the under-represented tail classes on the feature level with the knowledge from well-represented head classes. And the generated samples are mixed up with real training samples during training epochs. Gradual Balanced Loss is a general loss and it can combine with different decoupled training methods to improve the original performance. State-of-the-art results have been achieved on long-tail datasets such as CIFAR100-LT, ImageNetLT, and iNaturalist, which demonstrates the effectiveness of GLAG for long-tailed visual recognition.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | iNaturalist 2018 | Top-1 Accuracy69.2 | 287 | |
| Image Classification | ImageNet LT | Top-1 Accuracy51.5 | 251 | |
| Image Classification | CIFAR100 LT-100 1.0 (test) | Top-1 Acc (All)51.7 | 35 | |
| Image Classification | CIFAR100 (IM=10) LT (test) | Top-1 Acc64.5 | 20 | |
| Image Classification | CIFAR100 (IM=50) LT (test) | Top-1 Acc55.3 | 12 | |
| Image Classification | CIFAR100 IM=200 LT (test) | Top-1 Acc46.9 | 7 |