Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

High Dimensional Bayesian Optimization Using Dropout

About

Scaling Bayesian optimization to high dimensions is challenging task as the global optimization of high-dimensional acquisition function can be expensive and often infeasible. Existing methods depend either on limited active variables or the additive form of the objective function. We propose a new method for high-dimensional Bayesian optimization, that uses a dropout strategy to optimize only a subset of variables at each iteration. We derive theoretical bounds for the regret and show how it can inform the derivation of our algorithm. We demonstrate the efficacy of our algorithms for optimization on two benchmark functions and two real-world applications- training cascade classifiers and optimizing alloy composition.

Cheng Li, Sunil Gupta, Santu Rana, Vu Nguyen, Svetha Venkatesh, Alistair Shilton• 2018

Related benchmarks

TaskDatasetResultRank
Code GenerationHumanEval
Pass@126.83
850
Mathematical ReasoningMATH
Accuracy11.88
535
Black-box OptimizationHartmann-6D 300 evaluations
Wall Clock Time (s)3.237
10
Black-box OptimizationHartmann-6D 500 evaluations
Wall Clock Time (s)3.685
10
Black-box OptimizationLevy-10D 300 evaluations
Wall Clock Time (s)3.225
8
Black-box OptimizationLevy-10D 100 evaluations
Wall Clock Time (s)2.707
8
Showing 6 of 6 rows

Other info

Follow for update