Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Regularization With Stochastic Transformations and Perturbations for Deep Semi-Supervised Learning

About

Effective convolutional neural networks are trained on large sets of labeled data. However, creating large labeled datasets is a very costly and time-consuming task. Semi-supervised learning uses unlabeled data to train a model with higher accuracy when there is a limited set of labeled data available. In this paper, we consider the problem of semi-supervised learning with convolutional neural networks. Techniques such as randomized data augmentation, dropout and random max-pooling provide better generalization and stability for classifiers that are trained using gradient descent. Multiple passes of an individual sample through the network might lead to different predictions due to the non-deterministic behavior of these techniques. We propose an unsupervised loss function that takes advantage of the stochastic nature of these methods and minimizes the difference between the predictions of multiple passes of a training sample through the network. We evaluate the proposed method on several benchmark datasets.

Mehdi Sajjadi, Mehran Javanmardi, Tolga Tasdizen• 2016

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy63.55
3518
Image ClassificationCIFAR-10 (test)
Accuracy84.19
3381
Image ClassificationSTL-10 (test)
Accuracy69.73
357
Image ClassificationSVHN (test)
Accuracy87.83
199
ClassificationSVHN (test)
Error Rate6.03
182
Image ClassificationILSVRC 2012 (val)--
156
Digit ClassificationMNIST (test)
Error Rate0.27
94
Image ClassificationCIFAR-10 1.0 (test)
Accuracy85.99
54
Image ClassificationCIFAR-100 10000 labels (test)
Error Rate39.19
23
Image ClassificationSVHN 1.0 (test)
Accuracy92.46
22
Showing 10 of 11 rows

Other info

Follow for update