Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

How to train your ViT? Data, Augmentation, and Regularization in Vision Transformers

About

Vision Transformers (ViT) have been shown to attain highly competitive performance for a wide range of vision applications, such as image classification, object detection and semantic image segmentation. In comparison to convolutional neural networks, the Vision Transformer's weaker inductive bias is generally found to cause an increased reliance on model regularization or data augmentation ("AugReg" for short) when training on smaller training datasets. We conduct a systematic empirical study in order to better understand the interplay between the amount of training data, AugReg, model size and compute budget. As one result of this study we find that the combination of increased compute and AugReg can yield models with the same performance as models trained on an order of magnitude more training data: we train ViT models of various sizes on the public ImageNet-21k dataset which either match or outperform their counterparts trained on the larger, but not publicly available JFT-300M dataset.

Andreas Steiner, Alexander Kolesnikov, Xiaohua Zhai, Ross Wightman, Jakob Uszkoreit, Lucas Beyer• 2021

Related benchmarks

TaskDatasetResultRank
Semantic segmentationADE20K (val)--
2731
Object DetectionCOCO 2017 (val)--
2454
ClassificationImageNet-1K 1.0 (val)
Top-1 Accuracy (%)85.5
1155
Instance SegmentationCOCO 2017 (val)
APm0.415
1144
Image ClassificationImageNet-1k (val)
Top-1 Acc78.7
706
Image ClassificationCIFAR-100 (val)--
661
Image ClassificationCIFAR-100
Top-1 Accuracy87.8
622
Image ClassificationImageNet A
Top-1 Acc19.03
553
Image ClassificationImageNet-1K
Top-1 Acc86.8
524
Image ClassificationImageNet V2
Top-1 Acc67.9
487
Showing 10 of 36 rows

Other info

Follow for update