Supervised Contrastive Learning on Blended Images for Long-tailed Recognition
About
Real-world data often have a long-tailed distribution, where the number of samples per class is not equal over training classes. The imbalanced data form a biased feature space, which deteriorates the performance of the recognition model. In this paper, we propose a novel long-tailed recognition method to balance the latent feature space. First, we introduce a MixUp-based data augmentation technique to reduce the bias of the long-tailed data. Furthermore, we propose a new supervised contrastive learning method, named Supervised contrastive learning on Mixed Classes (SMC), for blended images. SMC creates a set of positives based on the class labels of the original images. The combination ratio of positives weights the positives in the training loss. SMC with the class-mixture-based loss explores more diverse data space, enhancing the generalization capability of the model. Extensive experiments on various benchmarks show the effectiveness of our one-stage training method.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | iNaturalist 2018 (test) | Top-1 Accuracy70.6 | 192 | |
| Image Classification | ImageNet-LT (test) | Top-1 Acc (All)56.6 | 159 | |
| Image Classification | CIFAR-100 Long-Tailed (test) | Top-1 Accuracy62.5 | 149 | |
| Image Classification | CIFAR-100 LT (val) | Top-1 Accuracy62.5 | 69 | |
| Image Classification | ImageNet-LT 1.0 (test) | Top-1 Accuracy56.6 | 37 | |
| Image Classification | CIFAR-100 LT IF 50 (test) | Top-1 Acc52.3 | 35 | |
| Image Classification | CIFAR-100-LT ($ ho=100$) (test) | Accuracy48.9 | 13 |