Learning Imbalanced Datasets with Label-Distribution-Aware Margin Loss
About
Deep learning algorithms can fare poorly when the training dataset suffers from heavy class-imbalance but the testing criterion requires good generalization on less frequent classes. We design two novel methods to improve performance in such scenarios. First, we propose a theoretically-principled label-distribution-aware margin (LDAM) loss motivated by minimizing a margin-based generalization bound. This loss replaces the standard cross-entropy objective during training and can be applied with prior strategies for training with class-imbalance such as re-weighting or re-sampling. Second, we propose a simple, yet effective, training schedule that defers re-weighting until after the initial stage, allowing the model to learn an initial representation while avoiding some of the complications associated with re-weighting or re-sampling. We test our methods on several benchmark vision tasks including the real-world imbalanced dataset iNaturalist 2018. Our experiments show that either of these methods alone can already improve over existing techniques and their combination achieves even better performance gains.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | CIFAR-100 (test) | Accuracy54.01 | 3518 | |
| Image Classification | CIFAR-10 (test) | Accuracy85.94 | 3381 | |
| Fine-grained Image Classification | CUB200 2011 (test) | Accuracy88.2 | 536 | |
| Fine-grained Image Classification | Stanford Cars (test) | Accuracy93.6 | 348 | |
| Fine-grained visual classification | FGVC-Aircraft (test) | Top-1 Acc92.9 | 287 | |
| Image Classification | iNaturalist 2018 | Top-1 Accuracy68 | 287 | |
| Image Classification | ImageNet LT | Top-1 Accuracy51.1 | 251 | |
| Image Classification | PACS | Overall Average Accuracy66.6 | 230 | |
| Long-Tailed Image Classification | ImageNet-LT (test) | Top-1 Acc (Overall)49.8 | 220 | |
| Image Classification | CIFAR-10 long-tailed (test) | Top-1 Acc90.3 | 201 |