An optimal algorithm for the Thresholding Bandit Problem
About
We study a specific \textit{combinatorial pure exploration stochastic bandit problem} where the learner aims at finding the set of arms whose means are above a given threshold, up to a given precision, and \textit{for a fixed time horizon}. We propose a parameter-free algorithm based on an original heuristic, and prove that it is optimal for this problem by deriving matching upper and lower bounds. To the best of our knowledge, this is the first non-trivial pure exploration setting with \textit{fixed budget} for which optimal strategies are constructed.
Andrea Locatelli, Maurilio Gutzeit, Alexandra Carpentier• 2016
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| e-good arm identification | Caption 853 | Probability of False Selection0.35 | 18 | |
| e-good arm identification | Example 2 | False Selection Probability25 | 16 | |
| e-good arm identification | Example 1 | False Selection Probability17 | 8 | |
| e-good arm identification | Example 3 | False Selection Probability0.49 | 8 | |
| e-good arm identification | Dose-finding | False Selection Probability53 | 8 | |
| e-good arm identification | Drug Selection | False Selection Probability70 | 8 | |
| e-good arm identification | Caption 854 | False Selection Probability48 | 8 |
Showing 7 of 7 rows