Stochastic Batch Acquisition: A Simple Baseline for Deep Active Learning
About
We examine a simple stochastic strategy for adapting well-known single-point acquisition functions to allow batch active learning. Unlike acquiring the top-K points from the pool set, score- or rank-based sampling takes into account that acquisition scores change as new data are acquired. This simple strategy for adapting standard single-sample acquisition strategies can even perform just as well as compute-intensive state-of-the-art batch acquisition functions, like BatchBALD or BADGE, while using orders of magnitude less compute. In addition to providing a practical option for machine learning practitioners, the surprising success of the proposed method in a wide range of experimental settings raises a difficult question for the field: when are these expensive batch acquisition methods pulling their weight?
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | Food101 | Accuracy90.1 | 457 | |
| Image Classification | CIFAR100 | Mean Accuracy91.2 | 55 | |
| Image Classification | DomainNet Real | Mean Accuracy84.7 | 55 | |
| Image Classification | CIFAR10 (train test) | Execution Time3.56e+3 | 11 | |
| Active Learning | nnActive average | AUBC64.85 | 9 |