Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Explainable and Explicit Visual Reasoning over Scene Graphs

About

We aim to dismantle the prevalent black-box neural architectures used in complex visual reasoning tasks, into the proposed eXplainable and eXplicit Neural Modules (XNMs), which advance beyond existing neural module networks towards using scene graphs --- objects as nodes and the pairwise relationships as edges --- for explainable and explicit reasoning with structured knowledge. XNMs allow us to pay more attention to teach machines how to "think", regardless of what they "look". As we will show in the paper, by using scene graphs as an inductive bias, 1) we can design XNMs in a concise and flexible fashion, i.e., XNMs merely consist of 4 meta-types, which significantly reduce the number of parameters by 10 to 100 times, and 2) we can explicitly trace the reasoning-flow in terms of graph attentions. XNMs are so generic that they support a wide range of scene graph implementations with various qualities. For example, when the graphs are detected perfectly, XNMs achieve 100% accuracy on both CLEVR and CLEVR CoGenT, establishing an empirical performance upper-bound for visual reasoning; when the graphs are noisily detected from real-world images, XNMs are still robust to achieve a competitive 67.5% accuracy on VQAv2.0, surpassing the popular bag-of-objects attention models without graph structures.

Jiaxin Shi, Hanwang Zhang, Juanzi Li• 2018

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVQA 2.0 (val)
Accuracy (Overall)64.7
143
Visual Question AnsweringCLEVR (test)
Overall Accuracy97.7
61
Visual Question AnsweringVQA 2.0 (test)
Accuracy67.5
24
Visual Question AnsweringCLEVR (val)
Overall Accuracy100
15
Visual Question AnsweringFVQA (test)--
14
Visual Question AnsweringCLEVR-CoGenT (val)--
12
Visual Question AnsweringCLEVR-CoGenT (test)
Accuracy (Condition A)98.1
3
Showing 7 of 7 rows

Other info

Code

Follow for update