Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Deep AutoAugment

About

While recent automated data augmentation methods lead to state-of-the-art results, their design spaces and the derived data augmentation strategies still incorporate strong human priors. In this work, instead of fixing a set of hand-picked default augmentations alongside the searched data augmentations, we propose a fully automated approach for data augmentation search named Deep AutoAugment (DeepAA). DeepAA progressively builds a multi-layer data augmentation pipeline from scratch by stacking augmentation layers one at a time until reaching convergence. For each augmentation layer, the policy is optimized to maximize the cosine similarity between the gradients of the original and augmented data along the direction with low variance. Our experiments show that even without default augmentations, we can learn an augmentation policy that achieves strong performance with that of previous works. Extensive ablation studies show that the regularized gradient matching is an effective search method for data augmentation policies. Our code is available at: https://github.com/MSU-MLSys-Lab/DeepAA .

Yu Zheng, Zhi Zhang, Shen Yan, Mi Zhang• 2022

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)--
3518
Image ClassificationCIFAR-10 (test)--
3381
Image ClassificationCIFAR10 (test)
Accuracy97.57
585
Image ClassificationImageNet (test)
Top-1 Acc81.32
235
Image ClassificationSVHNcore (test)
Accuracy98.04
10
Showing 5 of 5 rows

Other info

Code

Follow for update