Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Training independent subnetworks for robust prediction

About

Recent approaches to efficiently ensemble neural networks have shown that strong robustness and uncertainty performance can be achieved with a negligible gain in parameters over the original network. However, these methods still require multiple forward passes for prediction, leading to a significant computational cost. In this work, we show a surprising result: the benefits of using multiple predictions can be achieved `for free' under a single model's forward pass. In particular, we show that, using a multi-input multi-output (MIMO) configuration, one can utilize a single model's capacity to train multiple subnetworks that independently learn the task at hand. By ensembling the predictions made by the subnetworks, we improve model robustness without increasing compute. We observe a significant improvement in negative log-likelihood, accuracy, and calibration error on CIFAR10, CIFAR100, ImageNet, and their out-of-distribution variants compared to previous methods.

Marton Havasi, Rodolphe Jenatton, Stanislav Fort, Jeremiah Zhe Liu, Jasper Snoek, Balaji Lakshminarayanan, Andrew M. Dai, Dustin Tran• 2020

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-10 (test)
Accuracy95.4
3381
Image ClassificationCIFAR-100--
622
Image ClassificationCIFAR-100--
302
ClassificationCIFAR-100 (test)
Accuracy56.2
129
Image ClassificationCIFAR-10-C
Accuracy69.99
127
Image ClassificationCIFAR-100-C
Accuracy (Corruption)47.35
44
Out-of-Distribution DetectionCIFAR-10 vs SVHN
AUC0.9387
30
Image ClassificationCIFAR-100 WRN-28-10 (test)
Top-1 Accuracy82.74
28
Glaucoma Classificationretinal Glaucoma dataset (test)
Accuracy0.724
28
OOD DetectionRetinal Glaucoma images REFUGE (test)
AUROC61.74
28
Showing 10 of 17 rows

Other info

Follow for update