Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Certifiably Adversarially Robust Detection of Out-of-Distribution Data

About

Deep neural networks are known to be overconfident when applied to out-of-distribution (OOD) inputs which clearly do not belong to any class. This is a problem in safety-critical applications since a reliable assessment of the uncertainty of a classifier is a key property, allowing the system to trigger human intervention or to transfer into a safe state. In this paper, we aim for certifiable worst case guarantees for OOD detection by enforcing not only low confidence at the OOD point but also in an $l_\infty$-ball around it. For this purpose, we use interval bound propagation (IBP) to upper bound the maximal confidence in the $l_\infty$-ball and minimize this upper bound during training time. We show that non-trivial bounds on the confidence for OOD data generalizing beyond the OOD dataset seen at training time are possible. Moreover, in contrast to certified adversarial robustness which typically comes with significant loss in prediction performance, certified guarantees for worst case OOD detection are possible without much loss in accuracy.

Julian Bitterwolf, Alexander Meinke, Matthias Hein• 2020

Related benchmarks

TaskDatasetResultRank
Out-of-Distribution DetectionCIFAR-10 In-distribution vs LSUN_CR OOD
AUC97.4
8
Out-of-Distribution DetectionCIFAR-10
AUC (LSUN)96.5
8
Out-of-Distribution DetectionCIFAR-10 (In-distribution) vs SVHN (OOD)
AUC90.8
8
Out-of-Distribution DetectionCIFAR-10 (In-distribution) vs Smooth (OOD)
AUC96.2
8
Out-of-Distribution DetectionCIFAR-10 ID CIFAR-100 OOD
AUC76.7
8
Showing 5 of 5 rows

Other info

Follow for update