Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

WAIC, but Why? Generative Ensembles for Robust Anomaly Detection

About

Machine learning models encounter Out-of-Distribution (OoD) errors when the data seen at test time are generated from a different stochastic generator than the one used to generate the training data. One proposal to scale OoD detection to high-dimensional data is to learn a tractable likelihood approximation of the training distribution, and use it to reject unlikely inputs. However, likelihood models on natural data are themselves susceptible to OoD errors, and even assign large likelihoods to samples from other datasets. To mitigate this problem, we propose Generative Ensembles, which robustify density-based OoD detection by way of estimating epistemic uncertainty of the likelihood model. We present a puzzling observation in need of an explanation -- although likelihood measures cannot account for the typical set of a distribution, and therefore should not be suitable on their own for OoD detection, WAIC performs surprisingly well in practice.

Hyunsun Choi, Eric Jang, Alexander A. Alemi• 2018

Related benchmarks

TaskDatasetResultRank
Out-of-Distribution DetectionCIFAR-10 vs SVHN (test)
AUROC14.3
101
Out-of-Distribution DetectionCIFAR-10 vs CIFAR-100 (test)
AUROC53.2
93
Out-of-Distribution DetectionCIFAR-10 (ID) vs SVHN (OOD) (test)
AUROC100
79
Out-of-Distribution DetectionFashionMNIST (ID) vs MNIST (OoD)
AUROC0.766
61
Out-of-Distribution DetectionCIFAR-10 (ID) vs Celeb-A (OOD)
AUROC99.7
55
Out-of-Distribution DetectionMNIST Out-of-Distribution (test)--
7
OOD DetectionFashionMNIST (In-Distribution) vs Omniglot (Out-of-Distribution) original (test)
AUROC0.796
4
Out-of-Distribution DetectionOMNIGLOT Out-of-Distribution (test)
AUROC56.8
3
Showing 8 of 8 rows

Other info

Follow for update