ERNIE-ViL: Knowledge Enhanced Vision-Language Representations Through Scene Graph
About
We propose a knowledge-enhanced approach, ERNIE-ViL, which incorporates structured knowledge obtained from scene graphs to learn joint representations of vision-language. ERNIE-ViL tries to build the detailed semantic connections (objects, attributes of objects and relationships between objects) across vision and language, which are essential to vision-language cross-modal tasks. Utilizing scene graphs of visual scenes, ERNIE-ViL constructs Scene Graph Prediction tasks, i.e., Object Prediction, Attribute Prediction and Relationship Prediction tasks in the pre-training phase. Specifically, these prediction tasks are implemented by predicting nodes of different types in the scene graph parsed from the sentence. Thus, ERNIE-ViL can learn the joint representations characterizing the alignments of the detailed semantics across vision and language. After pre-training on large scale image-text aligned datasets, we validate the effectiveness of ERNIE-ViL on 5 cross-modal downstream tasks. ERNIE-ViL achieves state-of-the-art performances on all these tasks and ranks the first place on the VCR leaderboard with an absolute improvement of 3.7%.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Visual Question Answering | VQA v2 (test-dev) | Overall Accuracy72.62 | 664 | |
| Visual Question Answering | VQA v2 (test-std) | Accuracy74.93 | 466 | |
| Image-to-Text Retrieval | Flickr30K 1K (test) | R@188.7 | 439 | |
| Image-to-Text Retrieval | Flickr30K | R@188.7 | 379 | |
| Text-to-Image Retrieval | Flickr30K 1K (test) | R@176.7 | 375 | |
| Referring Expression Comprehension | RefCOCO+ (val) | Accuracy75.95 | 345 | |
| Visual Question Answering | VQA 2.0 (test-dev) | Accuracy74.75 | 337 | |
| Referring Expression Comprehension | RefCOCO+ (testB) | Accuracy66.91 | 235 | |
| Image Retrieval | MS-COCO 5K (test) | R@132.3 | 217 | |
| Referring Expression Comprehension | RefCOCO+ (testA) | Accuracy82.37 | 207 |