Situation Recognition with Graph Neural Networks
About
We address the problem of recognizing situations in images. Given an image, the task is to predict the most salient verb (action), and fill its semantic roles such as who is performing the action, what is the source and target of the action, etc. Different verbs have different roles (e.g. attacking has weapon), and each role can take on many possible values (nouns). We propose a model based on Graph Neural Networks that allows us to efficiently capture joint dependencies between roles using neural networks defined on a graph. Experiments with different graph connectivities show that our approach that propagates information between roles significantly outperforms existing work, as well as multiple baselines. We obtain roughly 3-5% improvement over previous work in predicting the full situation. We also provide a thorough qualitative analysis of our model and influence of different roles in the verbs.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Grounded Situation Recognition | SWiG (dev) | Value Accuracy68.89 | 51 | |
| Situation Recognition | imSitu (test) | Value Accuracy69.16 | 22 | |
| Grounded Situation Recognition | SWiG v1 (dev) | Top-1 Predicted Verb Accuracy36.93 | 21 | |
| Situation Recognition | imSitu (dev) | Value Accuracy68.89 | 18 | |
| Situation Prediction | imSitu (test) | Top-1 Verb Acc36.72 | 13 | |
| Grounded Situation Recognition | SWiG 1.0 (test) | Top-1 Verb Acc36.72 | 13 | |
| Situation Prediction | imSitu (dev) | Top-1 Verb Acc36.93 | 5 |