Video Understanding as Machine Translation
About
With the advent of large-scale multimodal video datasets, especially sequences with audio or transcribed speech, there has been a growing interest in self-supervised learning of video representations. Most prior work formulates the objective as a contrastive metric learning problem between the modalities. To enable effective learning, however, these strategies require a careful selection of positive and negative samples often combined with hand-designed curriculum policies. In this work we remove the need for negative sampling by taking a generative modeling approach that poses the objective as a translation problem between modalities. Such a formulation allows us to tackle a wide variety of downstream video understanding tasks by means of a single unified framework, without the need for large batches of negative samples common in contrastive metric learning. We experiment with the large-scale HowTo100M dataset for training, and report performance gains over the state-of-the-art on several downstream tasks including video classification (EPIC-Kitchens), question answering (TVQA), captioning (TVC, YouCook2, and MSR-VTT), and text-based clip retrieval (YouCook2 and MSR-VTT).
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Text-to-Video Retrieval | MSR-VTT | Recall@114.7 | 313 | |
| Text-to-Video Retrieval | MSR-VTT (1k-A) | R@1052.8 | 211 | |
| Video Captioning | MSR-VTT (test) | -- | 121 | |
| Text-to-Video Retrieval | YouCook2 | Recall@1043.9 | 117 | |
| Video Captioning | YouCook2 | METEOR13.4 | 104 | |
| Video Captioning | YouCook II (val) | -- | 98 | |
| Text-to-Video Retrieval | MSR-VTT 7K | Recall@1052.8 | 27 | |
| Text-to-Video Retrieval | MSRVTT 1K 1.0 (test) | R@114.7 | 23 |