Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Greedy Policy Search: A Simple Baseline for Learnable Test-Time Augmentation

About

Test-time data augmentation$-$averaging the predictions of a machine learning model across multiple augmented samples of data$-$is a widely used technique that improves the predictive performance. While many advanced learnable data augmentation techniques have emerged in recent years, they are focused on the training phase. Such techniques are not necessarily optimal for test-time augmentation and can be outperformed by a policy consisting of simple crops and flips. The primary goal of this paper is to demonstrate that test-time augmentation policies can be successfully learned too. We introduce greedy policy search (GPS), a simple but high-performing method for learning a policy of test-time augmentation. We demonstrate that augmentation policies learned with GPS achieve superior predictive performance on image classification problems, provide better in-domain uncertainty estimation, and improve the robustness to domain shift.

Dmitry Molchanov, Alexander Lyzhov, Yuliya Molchanova, Arsenii Ashukha, Dmitry Vetrov• 2020

Related benchmarks

TaskDatasetResultRank
Out-of-Distribution DetectionCIFAR-100 SVHN in-distribution out-of-distribution (test)
AUROC81.6
90
Out-of-Distribution DetectionImageNet-O
AUROC0.305
74
Out-of-Distribution DetectionCIFAR-100 (in-distribution) / LSUN (out-of-distribution) (test)
AUROC85
67
Out-of-Distribution DetectionCIFAR100 (ID) vs SVHN (OOD) (test)
AUROC81.6
40
Out-of-Distribution DetectionCIFAR-100 In-distribution vs Smooth (OOD)
AUC73.2
22
Out-of-Distribution DetectionImageNet-R
ROC AUC0.858
9
OOD DetectionCIFAR-100 (in-distribution) and LSUN (out-of-distribution) (test)
ROC AUC85
6
Showing 7 of 7 rows

Other info

Follow for update