Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Are Uncertainty Quantification Capabilities of Evidential Deep Learning a Mirage?

About

This paper questions the effectiveness of a modern predictive uncertainty quantification approach, called \emph{evidential deep learning} (EDL), in which a single neural network model is trained to learn a meta distribution over the predictive distribution by minimizing a specific objective function. Despite their perceived strong empirical performance on downstream tasks, a line of recent studies by Bengs et al. identify limitations of the existing methods to conclude their learned epistemic uncertainties are unreliable, e.g., in that they are non-vanishing even with infinite data. Building on and sharpening such analysis, we 1) provide a sharper understanding of the asymptotic behavior of a wide class of EDL methods by unifying various objective functions; 2) reveal that the EDL methods can be better interpreted as an out-of-distribution detection algorithm based on energy-based-models; and 3) conduct extensive ablation studies to better assess their empirical effectiveness with real-world datasets. Through all these analyses, we conclude that even when EDL methods are empirically effective on downstream tasks, this occurs despite their poor uncertainty quantification capabilities. Our investigation suggests that incorporating model uncertainty can help EDL methods faithfully quantify uncertainties and further improve performance on representative downstream tasks, albeit at the cost of additional computational complexity.

Maohao Shen, J. Jon Ryu, Soumya Ghosh, Yuheng Bu, Prasanna Sattigeri, Subhro Das, Gregory W. Wornell• 2024

Related benchmarks

TaskDatasetResultRank
OOD DetectionCIFAR-10 (IND) SVHN (OOD)
AUROC0.978
91
OOD DetectionCIFAR-100 IND SVHN OOD
AUROC (%)85.6
74
OOD DetectionCIFAR10 ID FMNIST OOD
AUROC0.944
54
OOD DetectionCIFAR-10 OOD (test)
AUROC98.4
36
Selective ClassificationCIFAR-100 (test)
AUC0.857
32
OOD DetectionCIFAR100 ID TImageNet OOD
AUROC0.841
31
OOD DetectionTinyImageNet (In-distribution) / CIFAR10 (OOD)
AUPR85.6
24
Selective ClassificationCIFAR-10 (test)
AUC0.903
21
OOD DetectionCIFAR-10 IND ImageNet R OOD
AUROC89.6
20
OOD DetectionCIFAR-10 vs SVHN (test)--
19
Showing 10 of 18 rows

Other info

Follow for update