Improving the Knowledge Gradient Algorithm
About
The knowledge gradient (KG) algorithm is a popular policy for the best arm identification (BAI) problem. It is built on the simple idea of always choosing the measurement that yields the greatest expected one-step improvement in the estimate of the best mean of the arms. In this research, we show that this policy has limitations, causing the algorithm not asymptotically optimal. We next provide a remedy for it, by following the manner of one-step look ahead of KG, but instead choosing the measurement that yields the greatest one-step improvement in the probability of selecting the best arm. The new policy is called improved knowledge gradient (iKG). iKG can be shown to be asymptotically optimal. In addition, we show that compared to KG, it is easier to extend iKG to variant problems of BAI, with the $\epsilon$-good arm identification and feasible arm identification as two examples. The superior performances of iKG on these problems are further demonstrated using numerical examples.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| e-good arm identification | Caption 853 | Probability of False Selection0.00e+0 | 18 | |
| e-good arm identification | Example 2 | False Selection Probability0.00e+0 | 16 | |
| Best Arm Identification | Example 1 Synthetic | False Selection Probability3 | 10 | |
| Best Arm Identification | Example 2 Synthetic | False Selection Probability3 | 10 | |
| Best Arm Identification | Example 3 Synthetic | False Selection Rate1 | 10 | |
| Best Arm Identification | Dose-finding ACR50 | Probability of False Selection1 | 10 | |
| Best Arm Identification | Drug Review Dataset Selection | Probability of False Selection23 | 10 | |
| Best Arm Identification | New Yorker Cartoon Caption Contest Caption 854 | False Selection Probability0.04 | 10 | |
| e-good arm identification | Example 1 | False Selection Probability3 | 8 | |
| e-good arm identification | Example 3 | False Selection Probability0.03 | 8 |