Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning Video Representations from Large Language Models

About

We introduce LaViLa, a new approach to learning video-language representations by leveraging Large Language Models (LLMs). We repurpose pre-trained LLMs to be conditioned on visual input, and finetune them to create automatic video narrators. Our auto-generated narrations offer a number of advantages, including dense coverage of long videos, better temporal synchronization of the visual information and text, and much higher diversity of text. The video-text embedding learned contrastively with these additional auto-generated narrations outperforms the previous state-of-the-art on multiple first-person and third-person video tasks, both in zero-shot and finetuned setups. Most notably, LaViLa obtains an absolute gain of 10.1% on EGTEA classification and 5.9% Epic-Kitchens-100 multi-instance retrieval benchmarks. Furthermore, LaViLa trained with only half the narrations from the Ego4D dataset outperforms baseline models trained on the full set, and shows positive scaling behavior on increasing pre-training data and model size.

Yue Zhao, Ishan Misra, Philipp Kr\"ahenb\"uhl, Rohit Girdhar• 2022

Related benchmarks

TaskDatasetResultRank
Video Question AnsweringMSRVTT-QA (test)
Accuracy44.9
371
Text-to-Video RetrievalDiDeMo
R@10.566
360
Video Question AnsweringMSVD-QA (test)
Accuracy53.7
274
Text-to-Video RetrievalActivityNet
R@10.587
197
Video-to-Text retrievalDiDeMo
R@147.1
108
Video-to-Text retrievalActivityNet
R@10.503
99
Text-to-Video RetrievalMSRVTT
R@156
98
Action RecognitionEPIC-KITCHENS (val)
Verb Top-1 Acc72
36
Action RecognitionEpic Kitchens 100--
26
Video-to-Text retrievalMSRVTT
R@149
24
Showing 10 of 23 rows

Other info

Follow for update