Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

PriorBand: Practical Hyperparameter Optimization in the Age of Deep Learning

About

Hyperparameters of Deep Learning (DL) pipelines are crucial for their downstream performance. While a large number of methods for Hyperparameter Optimization (HPO) have been developed, their incurred costs are often untenable for modern DL. Consequently, manual experimentation is still the most prevalent approach to optimize hyperparameters, relying on the researcher's intuition, domain knowledge, and cheap preliminary explorations. To resolve this misalignment between HPO algorithms and DL researchers, we propose PriorBand, an HPO algorithm tailored to DL, able to utilize both expert beliefs and cheap proxy tasks. Empirically, we demonstrate PriorBand's efficiency across a range of DL benchmarks and show its gains under informative expert input and robustness against poor expert beliefs

Neeratyoy Mallik, Edward Bergman, Carl Hvarfner, Danny Stoll, Maciej Janowski, Marius Lindauer, Luigi Nardi, Frank Hutter• 2023

Related benchmarks

TaskDatasetResultRank
Hyperparameter OptimizationJAHS-C10 Bench (val)
Validation Error8.236
52
Hyperparameter OptimizationJAHS-Bench CH (val)
Validation Error4.399
31
Hyperparameter OptimizationJAHS-Bench FM (val)
Validation Error4.709
28
Hyperparameter OptimizationCifar100 PD1 (val)
Validation Error22.1
24
Hyperparameter OptimizationImageNet PD1 (val)
Validation Error21.7
24
Hyperparameter OptimizationPD1-LM1B (val)
Validation Error0.628
24
Hyperparameter OptimizationPD1 WMT (val)
Validation Error34.7
24
Hyperparameter OptimizationLC-126026 (val)
Validation Error2.3
6
Hyperparameter OptimizationLC-167190 (val)
Validation Error13.3
6
Hyperparameter OptimizationLC-168330 (val)
Validation Error0.275
6
Showing 10 of 14 rows

Other info

Code

Follow for update