Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Evaluating Robustness of Predictive Uncertainty Estimation: Are Dirichlet-based Models Reliable?

About

Dirichlet-based uncertainty (DBU) models are a recent and promising class of uncertainty-aware models. DBU models predict the parameters of a Dirichlet distribution to provide fast, high-quality uncertainty estimates alongside with class predictions. In this work, we present the first large-scale, in-depth study of the robustness of DBU models under adversarial attacks. Our results suggest that uncertainty estimates of DBU models are not robust w.r.t. three important tasks: (1) indicating correctly and wrongly classified samples; (2) detecting adversarial examples; and (3) distinguishing between in-distribution (ID) and out-of-distribution (OOD) data. Additionally, we explore the first approaches to make DBU models more robust. While adversarial training has a minor effect, our median smoothing based approach significantly increases robustness of DBU models.

Anna-Kathrin Kopetzki, Bertrand Charpentier, Daniel Z\"ugner, Sandhya Giri, Stephan G\"unnemann• 2020

Related benchmarks

TaskDatasetResultRank
Image Classification, OOD Detection, and Adversarial Attack DetectionMNIST (ID) -> EMNIST (Near-OOD) (test)
ID Accuracy99.99
11
Image Classification, OOD Detection, and Adversarial Attack DetectionOxford Flowers low-shot (ID) -> Deep Weeds (OOD) (test)
ID Accuracy (%)99.09
11
Image Classification, OOD Detection, and Adversarial Attack DetectionMNIST (ID) -> KMNIST (OOD) (test)
ID Accuracy99.97
11
Image Classification, OOD Detection, and Adversarial Attack DetectionCIFAR10 (ID) -> CIFAR100 (Near-OOD) (test)
ID Accuracy96.69
11
Image Classification, OOD Detection, and Adversarial Attack DetectionCIFAR10 (ID) -> SVHN (OOD) (test)
ID Accuracy (%)96.07
11
Image Classification, OOD Detection, and Adversarial Attack DetectionMNIST (ID) -> FashionMNIST (OOD) (test)
ID Accuracy (%)99.95
11
Showing 6 of 6 rows

Other info

Follow for update