Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning

About

We propose a new regularization method based on virtual adversarial loss: a new measure of local smoothness of the conditional label distribution given input. Virtual adversarial loss is defined as the robustness of the conditional label distribution around each input data point against local perturbation. Unlike adversarial training, our method defines the adversarial direction without label information and is hence applicable to semi-supervised learning. Because the directions in which we smooth the model are only "virtually" adversarial, we call our method virtual adversarial training (VAT). The computational cost of VAT is relatively low. For neural networks, the approximated gradient of virtual adversarial loss can be computed with no more than two pairs of forward- and back-propagations. In our experiments, we applied VAT to supervised and semi-supervised learning tasks on multiple benchmark datasets. With a simple enhancement of the algorithm based on the entropy minimization principle, our VAT achieves state-of-the-art performance for semi-supervised learning tasks on SVHN and CIFAR-10.

Takeru Miyato, Shin-ichi Maeda, Masanori Koyama, Shin Ishii• 2017

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy63.76
3518
Image ClassificationCIFAR-10 (test)
Accuracy84.73
3381
Semantic segmentationPASCAL VOC 2012 (val)
Mean IoU63.34
2142
Image ClassificationCIFAR-10 (test)
Accuracy92.25
906
Image ClassificationMNIST (test)
Accuracy98.4
894
Image ClassificationCIFAR-100
Accuracy36.3
691
Image ClassificationFashion MNIST (test)
Accuracy87.1
592
Action RecognitionNTU RGB+D 60 (Cross-View)
Accuracy78.6
588
Image ClassificationCIFAR10 (test)--
585
Image ClassificationClothing1M (test)
Accuracy69.57
574
Showing 10 of 148 rows
...

Other info

Code

Follow for update