Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

An Embarrassingly Simple Baseline for Imbalanced Semi-Supervised Learning

About

Semi-supervised learning (SSL) has shown great promise in leveraging unlabeled data to improve model performance. While standard SSL assumes uniform data distribution, we consider a more realistic and challenging setting called imbalanced SSL, where imbalanced class distributions occur in both labeled and unlabeled data. Although there are existing endeavors to tackle this challenge, their performance degenerates when facing severe imbalance since they can not reduce the class imbalance sufficiently and effectively. In this paper, we study a simple yet overlooked baseline -- SimiS -- which tackles data imbalance by simply supplementing labeled data with pseudo-labels, according to the difference in class distribution from the most frequent class. Such a simple baseline turns out to be highly effective in reducing class imbalance. It outperforms existing methods by a significant margin, e.g., 12.8%, 13.6%, and 16.7% over previous SOTA on CIFAR100-LT, FOOD101-LT, and ImageNet127 respectively. The reduced imbalance results in faster convergence and better pseudo-label accuracy of SimiS. The simplicity of our method also makes it possible to be combined with other re-balancing techniques to improve the performance further. Moreover, our method shows great robustness to a wide range of data distributions, which holds enormous potential in practice. Code will be publicly available.

Hao Chen, Yue Fan, Yidong Wang, Jindong Wang, Bernt Schiele, Xing Xie, Marios Savvides, Bhiksha Raj• 2022

Related benchmarks

TaskDatasetResultRank
Medical Image SegmentationAMOS 5% labeled
Mean Dice47.27
29
Semi-supervised medical image segmentationSynapse (20% labeled)
Average Dice Score40.07
27
Multi-organ SegmentationSynapse 20% labeled data (test)
Avg. Dice40.07
16
Showing 3 of 3 rows

Other info

Follow for update