Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Modeling Context in Referring Expressions

About

Humans refer to objects in their environments all the time, especially in dialogue with other people. We explore generating and comprehending natural language referring expressions for objects in images. In particular, we focus on incorporating better measures of visual context into referring expression models and find that visual comparison to other objects within an image helps improve performance significantly. We also develop methods to tie the language generation process together, so that we generate expressions for all objects of a particular category jointly. Evaluation on three recent datasets - RefCOCO, RefCOCO+, and RefCOCOg, shows the advantages of our methods for both referring expression generation and comprehension.

Licheng Yu, Patrick Poirson, Shan Yang, Alexander C. Berg, Tamara L. Berg• 2016

Related benchmarks

TaskDatasetResultRank
Referring Expression ComprehensionRefCOCO+ (val)
Accuracy58.94
345
Referring Expression ComprehensionRefCOCO (val)
Accuracy76.18
335
Referring Expression ComprehensionRefCOCO (testA)
Accuracy0.7439
333
Referring Expression ComprehensionRefCOCOg (val)
Accuracy59.4
291
Referring Expression ComprehensionRefCOCO+ (test-A)
Accuracy61.29
172
Referring Expression ComprehensionRefCOCO+ (test-B)
Accuracy56.24
167
Referring Expression ComprehensionRefCOCO (test-B)
Accuracy77.3
160
Referring expression generationRefCOCOg (val)
METEOR14.5
31
Referring expression generationRefCOCO (testA)
Meteor18.5
13
Referring expression generationRefCOCO (testB)
Meteor24.7
13
Showing 10 of 13 rows

Other info

Follow for update