Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

A Data-Driven Measure of Relative Uncertainty for Misclassification Detection

About

Misclassification detection is an important problem in machine learning, as it allows for the identification of instances where the model's predictions are unreliable. However, conventional uncertainty measures such as Shannon entropy do not provide an effective way to infer the real uncertainty associated with the model's predictions. In this paper, we introduce a novel data-driven measure of uncertainty relative to an observer for misclassification detection. By learning patterns in the distribution of soft-predictions, our uncertainty measure can identify misclassified samples based on the predicted class probabilities. Interestingly, according to the proposed measure, soft-predictions corresponding to misclassified instances can carry a large amount of uncertainty, even though they may have low Shannon entropy. We demonstrate empirical improvements over multiple image classification tasks, outperforming state-of-the-art misclassification detection methods.

Eduardo Dadalto, Marco Romanelli, Georg Pichler, Pablo Piantanida• 2023

Related benchmarks

TaskDatasetResultRank
Out-of-Distribution DetectionCIFAR100 (ID) vs SVHN (OOD) (test)
AUROC95
40
Misclassification DetectionCIFAR-100--
27
Out-of-Distribution DetectionPlaces365 (OOD) / CIFAR-100 (ID) (test)
AUC0.99
16
Attack DetectionCIFAR-100 (test)
BIM AUC86
16
Adversarial Attack DetectionCIFAR-100 Adversarial
BIM Detection Score98
14
General Robustness and Detection EvaluationAggregate (CIFAR-100, CIFAR-100C, SVHN, Places365, Attacks)
Mean Score96
14
Out-of-Distribution DetectionPlaces365
FPR0.11
14
Robustness to CorruptionsCIFAR-100-C
Acc (C0)97
14
Out-of-Distribution DetectionSVHN
FPR55
14
Showing 9 of 9 rows

Other info

Follow for update