Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Efficient Vision-Language Pretraining with Visual Concepts and Hierarchical Alignment

About

Vision and Language Pretraining has become the prevalent approach for tackling multimodal downstream tasks. The current trend is to move towards ever larger models and pretraining datasets. This computational headlong rush does not seem reasonable in the long term to move toward sustainable solutions, and de facto excludes academic laboratories with limited resources. In this work, we propose a new framework, dubbed ViCHA, that efficiently exploits the input data to boost the learning by: (a) a new hierarchical cross-modal alignment loss, (b) new self-supervised scheme based on masked image modeling, (c) leveraging image-level annotations, called Visual Concepts, obtained with existing foundation models such as CLIP to boost the performance of the image encoder. Although pretrained on four times less data, our ViCHA strategy outperforms other approaches on several downstream tasks such as Image-Text Retrieval, VQA, Visual Reasoning, Visual Entailment and Visual Grounding. The code will be made publicly available here: https://github.com/mshukor/ViCHA

Mustafa Shukor, Guillaume Couairon, Matthieu Cord• 2022

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVQA 2.0 (test-dev)
Accuracy73.55
337
Natural Language Visual ReasoningNLVR2 (test-p)
Accuracy77
327
Natural Language Visual ReasoningNLVR2 (dev)
Accuracy78.14
288
Visual EntailmentSNLI-VE (test)
Overall Accuracy78.65
197
Visual EntailmentSNLI-VE (val)
Overall Accuracy79.2
109
Visual ReasoningNLVR2 (test-dev)
Accuracy77.27
14
Showing 6 of 6 rows

Other info

Follow for update