Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

VLM: Task-agnostic Video-Language Model Pre-training for Video Understanding

About

We present a simplified, task-agnostic multi-modal pre-training approach that can accept either video or text input, or both for a variety of end tasks. Existing pre-training are task-specific by adopting either a single cross-modal encoder that requires both modalities, limiting their use for retrieval-style end tasks or more complex multitask learning with two unimodal encoders, limiting early cross-modal fusion. We instead introduce new pretraining masking schemes that better mix across modalities (e.g. by forcing masks for text to predict the closest video embeddings) while also maintaining separability (e.g. unimodal predictions are sometimes required, without using all the input). Experimental results show strong performance across a wider range of tasks than any previous methods, often outperforming task-specific pre-training. Code is made available at https://github.com/pytorch/fairseq/tree/main/examples/MMPT.

Hu Xu, Gargi Ghosh, Po-Yao Huang, Prahal Arora, Masoumeh Aminzadeh, Christoph Feichtenhofer, Florian Metze, Luke Zettlemoyer• 2021

Related benchmarks

TaskDatasetResultRank
Text-to-Video RetrievalMSR-VTT
Recall@128.1
313
Text-to-Video RetrievalMSR-VTT (test)
R@128.1
234
Text-to-Video RetrievalYouCook2
Recall@1069.4
117
Video CaptioningYouCook2
METEOR18.22
104
Video CaptioningYouCook II (val)
CIDEr138.7
98
Text-to-Video RetrievalMSR-VTT 1k-A (test)
R@128.1
57
Video Question AnsweringMSR-VTT
Accuracy91.64
42
Action Step LocalizationCrossTask (test)
Recall46.5
32
Action SegmentationCOIN
Frame Accuracy68.39
29
Text-to-Video RetrievalMSR-VTT 1K videos (test)
Recall@1067.4
25
Showing 10 of 11 rows

Other info

Code

Follow for update