Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Image Retrieval on Real-life Images with Pre-trained Vision-and-Language Models

About

We extend the task of composed image retrieval, where an input query consists of an image and short textual description of how to modify the image. Existing methods have only been applied to non-complex images within narrow domains, such as fashion products, thereby limiting the scope of study on in-depth visual reasoning in rich image and language contexts. To address this issue, we collect the Compose Image Retrieval on Real-life images (CIRR) dataset, which consists of over 36,000 pairs of crowd-sourced, open-domain images with human-generated modifying text. To extend current methods to the open-domain, we propose CIRPLANT, a transformer based model that leverages rich pre-trained vision-and-language (V&L) knowledge for modifying visual features conditioned on natural language. Retrieval is then done by nearest neighbor lookup on the modified features. We demonstrate that with a relatively simple architecture, CIRPLANT outperforms existing methods on open-domain images, while matching state-of-the-art accuracy on the existing narrow datasets, such as fashion. Together with the release of CIRR, we believe this work will inspire further research on composed image retrieval.

Zheyuan Liu, Cristian Rodriguez-Opazo, Damien Teney, Stephen Gould• 2021

Related benchmarks

TaskDatasetResultRank
Composed Image RetrievalCIRR (test)
Recall@119.6
580
Composed Image RetrievalFashionIQ (val)
Average Recall@1018.9
489
Composed Image RetrievalFashion-IQ (test)
Average Recall@100.1887
169
Image-Text RetrievalRSICD--
119
Composed Image RetrievalFashion-IQ
Average Recall@5041.53
80
Scene ClassificationAID
Top-1 Acc64.25
69
Scene ClassificationUCM
Top-1 Accuracy69.67
28
Composed Image RetrievalCIRR Subset (test)
R@139.2
26
Image-to-Text RetrievalRSITMD
Rank4
24
Text-to-Image RetrievalRSITMD
mR4
24
Showing 10 of 27 rows

Other info

Code

Follow for update