Robustness quantification: a new method for assessing the reliability of the predictions of a classifier
About
Based on existing ideas in the field of imprecise probabilities, we present a new approach for assessing the reliability of the individual predictions of a generative probabilistic classifier. We call this approach robustness quantification, compare it to uncertainty quantification, and demonstrate that it continues to work well even for classifiers that are learned from small training sets that are sampled from a shifted distribution.
Adri\'an Detavernier, Jasper De Bock• 2025
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Selective Classification | Solar Flare big | AU-ARC89.26 | 15 | |
| Selective Classification | Student Port | AU-ARC92.76 | 15 | |
| Selective Classification | Solar Flare small | AU-ARC85.97 | 15 | |
| Selective Classification | SPECT Heart | AU-ARC89.15 | 15 | |
| Binary Classification | HEART DISEASE | AUC0.754 | 15 | |
| Classification | NPHA | AU-ARC0.5159 | 10 | |
| Classification | NURSERY | AU-ARC98.22 | 10 | |
| Classification | German Credit | AU-ARC0.838 | 10 | |
| Selective Classification | bank-marketing | AU-ARC94.52 | 10 | |
| Selective Classification | Lymphography | AU-ARC94.19 | 10 |
Showing 10 of 35 rows