Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Rotary Position Embedding for Vision Transformer

About

Rotary Position Embedding (RoPE) performs remarkably on language models, especially for length extrapolation of Transformers. However, the impacts of RoPE on computer vision domains have been underexplored, even though RoPE appears capable of enhancing Vision Transformer (ViT) performance in a way similar to the language domain. This study provides a comprehensive analysis of RoPE when applied to ViTs, utilizing practical implementations of RoPE for 2D vision data. The analysis reveals that RoPE demonstrates impressive extrapolation performance, i.e., maintaining precision while increasing image resolution at inference. It eventually leads to performance improvement for ImageNet-1k, COCO detection, and ADE-20k segmentation. We believe this study provides thorough guidelines to apply RoPE into ViT, promising improved backbone performance with minimal extra computational overhead. Our code and pre-trained models are available at https://github.com/naver-ai/rope-vit

Byeongho Heo, Song Park, Dongyoon Han, Sangdoo Yun• 2024

Related benchmarks

TaskDatasetResultRank
Image ClassificationImageNet-1K
Top-1 Acc83.8
836
Image ClassificationImageNet-1k (val)
Top-1 Accuracy80.4
512
Action RecognitionUCF101
Accuracy41.2
365
Image ClassificationImageNet 1k (test)
Top-1 Accuracy80
359
Image ClassificationCIFAR-100
Accuracy75.6
302
Object DetectionCOCO
mAP38
107
3D Object ClassificationModelNet40--
62
Semantic segmentationScanNet
mIoU71.1
59
Image ClassificationImageNet-1k (val)
Top-1 Accuracy85.4
32
Image-to-Text RetrievalDOCCI
R@539.5
28
Showing 10 of 20 rows

Other info

Follow for update