Long-tail learning via logit adjustment
About
Real-world classification problems typically exhibit an imbalanced or long-tailed label distribution, wherein many labels are associated with only a few samples. This poses a challenge for generalisation on such labels, and also makes na\"ive learning biased towards dominant labels. In this paper, we present two simple modifications of standard softmax cross-entropy training to cope with these challenges. Our techniques revisit the classic idea of logit adjustment based on the label frequencies, either applied post-hoc to a trained model, or enforced in the loss during training. Such adjustment encourages a large relative margin between logits of rare versus dominant labels. These techniques unify and generalise several recent proposals in the literature, while possessing firmer statistical grounding and empirical performance.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | iNaturalist 2018 | Top-1 Accuracy68.4 | 287 | |
| Image Classification | ImageNet LT | Top-1 Accuracy56.5 | 251 | |
| Long-Tailed Image Classification | ImageNet-LT (test) | -- | 220 | |
| Image Classification | CIFAR-10 long-tailed (test) | Top-1 Acc75.3 | 201 | |
| Image Classification | iNaturalist 2018 (test) | Top-1 Accuracy66.4 | 192 | |
| Text Classification | SST-2 (test) | Accuracy86.61 | 185 | |
| Image Classification | ImageNet-LT (test) | Top-1 Acc (All)51.1 | 159 | |
| Image Classification | CIFAR100 long-tailed (test) | Accuracy58.6 | 155 | |
| Classification | CIFAR100-LT (test) | Accuracy62.4 | 136 | |
| Long-tailed Visual Recognition | ImageNet LT | Overall Accuracy52.1 | 89 |