Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

An Effective Baseline for Robustness to Distributional Shift

About

Refraining from confidently predicting when faced with categories of inputs different from those seen during training is an important requirement for the safe deployment of deep learning systems. While simple to state, this has been a particularly challenging problem in deep learning, where models often end up making overconfident predictions in such situations. In this work we present a simple, but highly effective approach to deal with out-of-distribution detection that uses the principle of abstention: when encountering a sample from an unseen class, the desired behavior is to abstain from predicting. Our approach uses a network with an extra abstention class and is trained on a dataset that is augmented with an uncurated set that consists of a large number of out-of-distribution (OoD) samples that are assigned the label of the abstention class; the model is then trained to learn an effective discriminator between in and out-of-distribution samples. We compare this relatively simple approach against a wide variety of more complex methods that have been proposed both for out-of-distribution detection as well as uncertainty modeling in deep learning, and empirically demonstrate its effectiveness on a wide variety of of benchmarks and deep architectures for image recognition and text classification, often outperforming existing approaches by significant margins. Given the simplicity and effectiveness of this method, we propose that this approach be used as a new additional baseline for future work in this domain.

Sunil Thulasidasan, Sushil Thapa, Sayera Dhaubhadel, Gopinath Chennupati, Tanmoy Bhattacharya, Jeff Bilmes• 2021

Related benchmarks

TaskDatasetResultRank
Out-of-Distribution DetectionCIFAR-10 vs SVHN (test)
AUROC0.9949
101
Out-of-Distribution DetectionCIFAR-100 SVHN in-distribution out-of-distribution (test)
AUROC86.85
90
Out-of-Distribution DetectionCIFAR-10 (in-distribution) TinyImageNet (out-of-distribution) (test)
AUROC99.88
71
Out-of-Distribution DetectionCIFAR-100 (in-distribution) / LSUN (out-of-distribution) (test)
AUROC98.45
67
Out-of-Distribution DetectionLSUN (Out-of-distribution) vs CIFAR-10 (In-distribution)
AUROC99.96
28
Out-of-Distribution DetectionCIFAR-10 Gaussian
AUROC99.98
11
Out-of-Distribution Detection20 Newsgroup (test)
FPR@950.76
10
Out-of-Distribution DetectionTiny ImageNet (Out-of-distribution) vs CIFAR-100 (In-distribution)--
10
Out-of-Distribution DetectionCIFAR-100 Gaussian
AUROC99.65
8
Out-of-Distribution DetectionPlaces365 (Out-of-distribution) vs CIFAR-100 (In-distribution)
FPR@9523.37
7
Showing 10 of 11 rows

Other info

Follow for update