Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Large-scale Pretraining for Visual Dialog: A Simple State-of-the-Art Baseline

About

Prior work in visual dialog has focused on training deep neural models on VisDial in isolation. Instead, we present an approach to leverage pretraining on related vision-language datasets before transferring to visual dialog. We adapt the recently proposed ViLBERT (Lu et al., 2019) model for multi-turn visually-grounded conversations. Our model is pretrained on the Conceptual Captions and Visual Question Answering datasets, and finetuned on VisDial. Our best single model outperforms prior published work (including model ensembles) by more than 1% absolute on NDCG and MRR. Next, we find that additional finetuning using "dense" annotations in VisDial leads to even higher NDCG -- more than 10% over our base model -- but hurts MRR -- more than 17% below our base model! This highlights a trade-off between the two primary metrics -- NDCG and MRR -- which we find is due to dense annotations not correlating well with the original ground-truth answers to questions.

Vishvak Murahari, Dhruv Batra, Devi Parikh, Abhishek Das• 2019

Related benchmarks

TaskDatasetResultRank
Visual DialogVisDial v1.0 (test-std)
NDCG74.47
77
Visual DialogVisDial 1.0 (val)
MRR0.691
65
Visual DialogueVisDial v1.0 (test)
NDCG63.87
26
Visual DialogVisDial v1.0 (val)
Y/N Accuracy (C)68.99
2
Visual DialogVisDial (val)
RAD (Y/N <- C)62.08
2
Showing 5 of 5 rows

Other info

Follow for update