Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Diffusion Models and Semi-Supervised Learners Benefit Mutually with Few Labels

About

In an effort to further advance semi-supervised generative and classification tasks, we propose a simple yet effective training strategy called dual pseudo training (DPT), built upon strong semi-supervised learners and diffusion models. DPT operates in three stages: training a classifier on partially labeled data to predict pseudo-labels; training a conditional generative model using these pseudo-labels to generate pseudo images; and retraining the classifier with a mix of real and pseudo images. Empirically, DPT consistently achieves SOTA performance of semi-supervised generation and classification across various settings. In particular, with one or two labels per class, DPT achieves a Fr\'echet Inception Distance (FID) score of 3.08 or 2.52 on ImageNet 256x256. Besides, DPT outperforms competitive semi-supervised baselines substantially on ImageNet classification tasks, achieving top-1 accuracies of 59.0 (+2.8), 69.5 (+3.0), and 74.4 (+2.0) with one, two, or five labels per class, respectively. Notably, our results demonstrate that diffusion can generate realistic images with only a few labels (e.g., <0.1%) and generative augmentation remains viable for semi-supervised classification. Our code is available at https://github.com/ML-GSAI/DPT.

Zebin You, Yong Zhong, Fan Bao, Jiacheng Sun, Chongxuan Li, Jun Zhu• 2023

Related benchmarks

TaskDatasetResultRank
Image ClassificationImageNet 1k (test)--
798
Image GenerationImageNet 512x512 (val)
FID-50K4.05
184
Image GenerationImageNet 256x256 (train)
FID2.29
91
Image ClassificationCIFAR-10 40 labels
Error Rate4.68
81
Image GenerationImageNet 128x128
FID4.58
51
Image ClassificationImageNet-1K 1.0 (1% labels)
Top-1 Acc80.2
28
Image ClassificationCIFAR-10 25 labels per class 32x32--
15
Image ClassificationImageNet-1k (val)
Top-1 Acc (1% labels)80.2
9
Showing 8 of 8 rows

Other info

Code

Follow for update