Self-Distribution Distillation: Efficient Uncertainty Estimation
About
Deep learning is increasingly being applied in safety-critical domains. For these scenarios it is important to know the level of uncertainty in a model's prediction to ensure appropriate decisions are made by the system. Deep ensembles are the de-facto standard approach to obtaining various measures of uncertainty. However, ensembles often significantly increase the resources required in the training and/or deployment phases. Approaches have been developed that typically address the costs in one of these phases. In this work we propose a novel training approach, self-distribution distillation (S2D), which is able to efficiently train a single model that can estimate uncertainties. Furthermore it is possible to build ensembles of these models and apply hierarchical ensemble distillation approaches. Experiments on CIFAR-100 showed that S2D models outperformed standard models and Monte-Carlo dropout. Additional out-of-distribution detection experiments on LSUN, Tiny ImageNet, SVHN showed that even a standard deep ensemble can be outperformed using S2D based ensembles and novel distilled models.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| OOD Detection | CIFAR-10 (IND) SVHN (OOD) | AUROC0.933 | 91 | |
| OOD Detection | CIFAR-100 IND SVHN OOD | AUROC (%)62.1 | 74 | |
| OOD Detection | CIFAR10 ID FMNIST OOD | AUROC0.9 | 54 | |
| OOD Detection | CIFAR-10 OOD (test) | AUROC93.9 | 36 | |
| Selective Classification | CIFAR-100 (test) | AUC0.844 | 32 | |
| OOD Detection | CIFAR100 ID TImageNet OOD | AUROC0.715 | 31 | |
| OOD Detection | TinyImageNet (In-distribution) / CIFAR10 (OOD) | AUPR82.4 | 24 | |
| Selective Classification | CIFAR-10 (test) | AUC0.894 | 21 | |
| OOD Detection | CIFAR-10 IND ImageNet R OOD | AUROC87 | 20 | |
| OOD Detection | CIFAR-10 vs SVHN (test) | -- | 19 |