Good Semi-supervised Learning that Requires a Bad GAN
About
Semi-supervised learning methods based on generative adversarial networks (GANs) obtained strong empirical results, but it is not clear 1) how the discriminator benefits from joint training with a generator, and 2) why good semi-supervised classification performance and a good generator cannot be obtained at the same time. Theoretically, we show that given the discriminator objective, good semisupervised learning indeed requires a bad generator, and propose the definition of a preferred generator. Empirically, we derive a novel formulation based on our analysis that substantially improves over feature matching GANs, obtaining state-of-the-art results on multiple benchmark datasets.
Zihang Dai, Zhilin Yang, Fan Yang, William W. Cohen, Ruslan Salakhutdinov• 2017
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Classification | SVHN (test) | -- | 182 | |
| Image Classification | CIFAR-10 standard (test) | -- | 68 | |
| Image Classification | MNIST standard (test) | -- | 40 | |
| Image Classification | CIFAR-10 400 labels per class (test) | Accuracy85.59 | 22 | |
| Image Classification | SVHN 1,000 labels (train) | Error Rate (%)4.25 | 15 | |
| Image Classification | CIFAR10 4,000 labels (train) | Error Rate14.41 | 15 | |
| Semi-supervised classification | MNIST 100 labels statically binarized (test) | Error Rate (%)0.8 | 10 | |
| Image Classification | SVHN 1k labeled 72k unlabeled (test) | Accuracy95.75 | 8 |
Showing 8 of 8 rows