Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Self-Distribution Distillation: Efficient Uncertainty Estimation

About

Deep learning is increasingly being applied in safety-critical domains. For these scenarios it is important to know the level of uncertainty in a model's prediction to ensure appropriate decisions are made by the system. Deep ensembles are the de-facto standard approach to obtaining various measures of uncertainty. However, ensembles often significantly increase the resources required in the training and/or deployment phases. Approaches have been developed that typically address the costs in one of these phases. In this work we propose a novel training approach, self-distribution distillation (S2D), which is able to efficiently train a single model that can estimate uncertainties. Furthermore it is possible to build ensembles of these models and apply hierarchical ensemble distillation approaches. Experiments on CIFAR-100 showed that S2D models outperformed standard models and Monte-Carlo dropout. Additional out-of-distribution detection experiments on LSUN, Tiny ImageNet, SVHN showed that even a standard deep ensemble can be outperformed using S2D based ensembles and novel distilled models.

Yassir Fathullah, Mark J. F. Gales• 2022

Related benchmarks

TaskDatasetResultRank
OOD DetectionCIFAR-10 (IND) SVHN (OOD)
AUROC0.933
91
OOD DetectionCIFAR-100 IND SVHN OOD
AUROC (%)62.1
74
OOD DetectionCIFAR10 ID FMNIST OOD
AUROC0.9
54
OOD DetectionCIFAR-10 OOD (test)
AUROC93.9
36
Selective ClassificationCIFAR-100 (test)
AUC0.844
32
OOD DetectionCIFAR100 ID TImageNet OOD
AUROC0.715
31
OOD DetectionTinyImageNet (In-distribution) / CIFAR10 (OOD)
AUPR82.4
24
Selective ClassificationCIFAR-10 (test)
AUC0.894
21
OOD DetectionCIFAR-10 IND ImageNet R OOD
AUROC87
20
OOD DetectionCIFAR-10 vs SVHN (test)--
19
Showing 10 of 18 rows

Other info

Follow for update