Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Look Before you Speak: Visually Contextualized Utterances

About

While most conversational AI systems focus on textual dialogue only, conditioning utterances on visual context (when it's available) can lead to more realistic conversations. Unfortunately, a major challenge for incorporating visual context into conversational dialogue is the lack of large-scale labeled datasets. We provide a solution in the form of a new visually conditioned Future Utterance Prediction task. Our task involves predicting the next utterance in a video, using both visual frames and transcribed speech as context. By exploiting the large number of instructional videos online, we train a model to solve this task at scale, without the need for manual annotations. Leveraging recent advances in multimodal learning, our model consists of a novel co-attentional multimodal video transformer, and when trained on both textual and visual context, outperforms baselines that use textual inputs alone. Further, we demonstrate that our model trained for this task on unlabelled videos achieves state-of-the-art performance on a number of downstream VideoQA benchmarks such as MSRVTT-QA, MSVD-QA, ActivityNet-QA and How2QA.

Paul Hongsuck Seo, Arsha Nagrani, Cordelia Schmid• 2020

Related benchmarks

TaskDatasetResultRank
Video Question AnsweringMSRVTT-QA
Accuracy39.5
481
Video Question AnsweringMSRVTT-QA (test)
Accuracy39.5
371
Video Question AnsweringMSVD-QA
Accuracy42.6
340
Video Question AnsweringActivityNet-QA
Accuracy38.8
319
Video Question AnsweringActivityNet-QA (test)
Accuracy38.8
275
Video Question AnsweringMSVD-QA (test)--
274
Video Question AnsweringMSVD (test)
Accuracy42.6
30
Video Question AnsweringMSRVTT (test)
Accuracy39.5
26
Video Question AnsweringHow2QA public (val)
Top-1 Acc82.3
4
Showing 9 of 9 rows

Other info

Follow for update