Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning Credal Ensembles via Distributionally Robust Optimization

About

Credal predictors are models that are aware of epistemic uncertainty and produce a convex set of probabilistic predictions. They offer a principled way to quantify predictive epistemic uncertainty (EU) and have been shown to improve model robustness in various settings. However, most state-of-the-art methods mainly define EU as disagreement caused by random training initializations, which mostly reflects sensitivity to optimization randomness rather than uncertainty from deeper sources. To address this, we define EU as disagreement among models trained with varying relaxations of the i.i.d. assumption between training and test data. Based on this idea, we propose CreDRO, which learns an ensemble of plausible models through distributionally robust optimization. As a result, CreDRO captures EU not only from training randomness but also from meaningful disagreement due to potential distribution shifts between training and test data. Empirical results show that CreDRO consistently outperforms existing credal methods on tasks such as out-of-distribution detection across multiple benchmarks and selective classification in medical applications.

Kaizheng Wang, Ghifari Adam Faza, Fabio Cuzzolin, Siu Lun Chau, David Moens, Hans Hallez• 2026

Related benchmarks

TaskDatasetResultRank
OOD DetectionCIFAR-10 (IND) SVHN (OOD)
AUROC0.974
91
OOD DetectionCIFAR-10 (ID) vs Places 365 (OOD)
AUROC92.7
77
OOD DetectionCIFAR10 ID FMNIST OOD
AUROC0.964
54
OOD DetectionImageNet (OOD) with CIFAR10 (ID)
AUROC91.1
36
OOD DetectionCIFAR10 (ID) vs CIFAR100 (OOD) (test)
AUROC92.5
36
OOD DetectionCIFAR10 (ID) vs CIFAR100 (OOD)
AUROC92.5
8
Image ClassificationCIFAR10 (test)
Accuracy (Test)96.37
2
Showing 7 of 7 rows

Other info

Follow for update