Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ActBERT: Learning Global-Local Video-Text Representations

About

In this paper, we introduce ActBERT for self-supervised learning of joint video-text representations from unlabeled data. First, we leverage global action information to catalyze the mutual interactions between linguistic texts and local regional objects. It uncovers global and local visual clues from paired video sequences and text descriptions for detailed visual and text relation modeling. Second, we introduce an ENtangled Transformer block (ENT) to encode three sources of information, i.e., global actions, local regional objects, and linguistic descriptions. Global-local correspondences are discovered via judicious clues extraction from contextual information. It enforces the joint videotext representation to be aware of fine-grained objects as well as global human intention. We validate the generalization capability of ActBERT on downstream video-and language tasks, i.e., text-video clip retrieval, video captioning, video question answering, action segmentation, and action step localization. ActBERT significantly outperforms the state-of-the-arts, demonstrating its superiority in video-text representation learning.

Linchao Zhu, Yi Yang• 2020

Related benchmarks

TaskDatasetResultRank
Video Question AnsweringMSRVTT-QA (test)
Accuracy85.7
371
Text-to-Video RetrievalMSR-VTT
Recall@116.3
313
Text-to-Video RetrievalMSR-VTT (test)
R@116.3
234
Text-to-Video RetrievalMSR-VTT (1k-A)
R@1056.9
211
Text-to-Video RetrievalMSRVTT (test)
Recall@10.163
155
Text-to-Video RetrievalYouCook2
Recall@1038
117
Video CaptioningYouCook2
METEOR13.3
104
Video CaptioningYouCook II (val)
CIDEr65
98
Text-to-Video RetrievalMSRVTT
R@18.6
98
Text-to-Video RetrievalMSRVTT
R@116.3
75
Showing 10 of 37 rows

Other info

Follow for update