Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Evidential Deep Learning to Quantify Classification Uncertainty

About

Deterministic neural nets have been shown to learn effective predictors on a wide range of machine learning problems. However, as the standard approach is to train the network to minimize a prediction loss, the resultant model remains ignorant to its prediction confidence. Orthogonally to Bayesian neural nets that indirectly infer prediction uncertainty through weight uncertainties, we propose explicit modeling of the same using the theory of subjective logic. By placing a Dirichlet distribution on the class probabilities, we treat predictions of a neural net as subjective opinions and learn the function that collects the evidence leading to these opinions by a deterministic neural net from data. The resultant predictor for a multi-class classification problem is another Dirichlet distribution whose parameters are set by the continuous output of a neural net. We provide a preliminary analysis on how the peculiarities of our new loss function drive improved uncertainty estimation. We observe that our method achieves unprecedented success on detection of out-of-distribution queries and endurance against adversarial perturbations.

Murat Sensoy, Lance Kaplan, Melih Kandemir• 2018

Related benchmarks

TaskDatasetResultRank
ClassificationCIFAR10 (test)
Accuracy83.55
266
ClassificationCIFAR-100 (test)
Accuracy45.91
129
Image ClassificationCIFAR-10-C
Accuracy59.54
127
OOD DetectionCIFAR-10 (IND) SVHN (OOD)
AUROC0.965
91
OOD DetectionCIFAR-100 IND SVHN OOD
AUROC (%)68.7
74
Out-of-Distribution DetectionCIFAR-10 (in-distribution) TinyImageNet (out-of-distribution) (test)
AUROC51.64
71
OOD DetectionCIFAR10 ID FMNIST OOD
AUROC0.903
54
OOD DetectionCIFAR-10 (test)
AUROC66
40
OOD DetectionCIFAR-10 OOD (test)
AUROC98.8
36
Out-of-Distribution DetectionMNIST (In-distribution) vs Fashion-MNIST (OOD) (test)
AUPR0.8022
36
Showing 10 of 66 rows

Other info

Follow for update