Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Neural Ensemble Search for Uncertainty Estimation and Dataset Shift

About

Ensembles of neural networks achieve superior performance compared to stand-alone networks in terms of accuracy, uncertainty calibration and robustness to dataset shift. \emph{Deep ensembles}, a state-of-the-art method for uncertainty estimation, only ensemble random initializations of a \emph{fixed} architecture. Instead, we propose two methods for automatically constructing ensembles with \emph{varying} architectures, which implicitly trade-off individual architectures' strengths against the ensemble's diversity and exploit architectural variation as a source of diversity. On a variety of classification tasks and modern architecture search spaces, we show that the resulting ensembles outperform deep ensembles not only in terms of accuracy but also uncertainty calibration and robustness to dataset shift. Our further analysis and ablation studies provide evidence of higher ensemble diversity due to architectural variation, resulting in ensembles that can outperform deep ensembles, even when having weaker average base learners. To foster reproducibility, our code is available: \url{https://github.com/automl/nes}

Sheheryar Zaidi, Arber Zela, Thomas Elsken, Chris Holmes, Frank Hutter, Yee Whye Teh• 2020

Related benchmarks

TaskDatasetResultRank
CASHEEG (test)
Test Error0.0268
9
CASHbank32nh (test)
Test Error18.21
9
CASHSATIMAGE (test)
Test Error8.66
9
CASHquake (test)
Test Error46.42
9
CASHSpambase (test)
Test Error0.0623
9
CASH2dplanes (test)
Test Error7.11
9
CASHcpu_act (test)
Test Error5.59
9
CASHelevators (test)
Test Error9.75
9
CASHOpenML 15 datasets aggregate
Average Rank5
9
CASHcpu small (test)
Test Error7.28
9
Showing 10 of 16 rows

Other info

Follow for update