Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

An Empirical Study of End-to-End Video-Language Transformers with Masked Visual Modeling

About

Masked visual modeling (MVM) has been recently proven effective for visual pre-training. While similar reconstructive objectives on video inputs (e.g., masked frame modeling) have been explored in video-language (VidL) pre-training, previous studies fail to find a truly effective MVM strategy that can largely benefit the downstream performance. In this work, we systematically examine the potential of MVM in the context of VidL learning. Specifically, we base our study on a fully end-to-end VIdeO-LanguagE Transformer (VIOLET), where the supervision from MVM training can be backpropagated to the video pixel space. In total, eight different reconstructive targets of MVM are explored, from low-level pixel values and oriented gradients to high-level depth maps, optical flow, discrete visual tokens, and latent visual features. We conduct comprehensive experiments and provide insights into the factors leading to effective MVM training, resulting in an enhanced model VIOLETv2. Empirically, we show VIOLETv2 pre-trained with MVM objective achieves notable improvements on 13 VidL benchmarks, ranging from video question answering, video captioning, to text-to-video retrieval.

Tsu-Jui Fu, Linjie Li, Zhe Gan, Kevin Lin, William Yang Wang, Lijuan Wang, Zicheng Liu• 2022

Related benchmarks

TaskDatasetResultRank
Video Question AnsweringMSRVTT-QA
Accuracy44.5
481
Text-to-Video RetrievalDiDeMo (test)
R@147.9
376
Video Question AnsweringMSRVTT-QA (test)
Accuracy44.5
371
Text-to-Video RetrievalDiDeMo
R@10.479
360
Video Question AnsweringMSVD-QA
Accuracy54.7
340
Video Question AnsweringMSVD-QA (test)
Accuracy54.7
274
Text-to-Video RetrievalLSMDC (test)
R@124
225
Text-to-Video RetrievalActivityNet
R@10.181
197
Video Question AnsweringEgoSchema (Full)
Accuracy19.9
193
Text-to-Video RetrievalMSRVTT (test)
Recall@10.372
155
Showing 10 of 37 rows

Other info

Code

Follow for update