Bayesian Hyperparameter Optimization for Ensemble Learning
About
In this paper, we bridge the gap between hyperparameter optimization and ensemble learning by performing Bayesian optimization of an ensemble with regards to its hyperparameters. Our method consists in building a fixed-size ensemble, optimizing the configuration of one classifier of the ensemble at each iteration of the hyperparameter optimization algorithm, taking into consideration the interaction with the other models when evaluating potential performances. We also consider the case where the ensemble is to be reconstructed at the end of the hyperparameter optimization phase, through a greedy selection over the pool of models generated during the optimization. We study the performance of our proposed method on three different hyperparameter spaces, showing that our approach is better than both the best single model and a greedy ensemble construction over the models produced by a standard Bayesian optimization.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Classification | 50 classification tasks (test) | Average Test Rank9.12 | 19 | |
| Automated Machine Learning | 80 tasks all (test) | Average Test Rank10.36 | 19 | |
| Regression | 30 regression tasks (test) | Average Test Rank12.43 | 19 | |
| CASH | amazon_employee (test) | Test Error (%)5.19 | 9 | |
| CASH | pollen (test) | Test Error49.01 | 9 | |
| CASH | EEG (test) | Test Error0.0354 | 9 | |
| CASH | puma32H (test) | Test Error9.74 | 9 | |
| CASH | house_8L (test) | Test Error (%)11.14 | 9 | |
| CASH | Spambase (test) | Test Error0.0641 | 9 | |
| CASH | 2dplanes (test) | Test Error7.13 | 9 |