Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Extending CLIP's Image-Text Alignment to Referring Image Segmentation

About

Referring Image Segmentation (RIS) is a cross-modal task that aims to segment an instance described by a natural language expression. Recent methods leverage large-scale pretrained unimodal models as backbones along with fusion techniques for joint reasoning across modalities. However, the inherent cross-modal nature of RIS raises questions about the effectiveness of unimodal backbones. We propose RISCLIP, a novel framework that effectively leverages the cross-modal nature of CLIP for RIS. Observing CLIP's inherent alignment between image and text features, we capitalize on this starting point and introduce simple but strong modules that enhance unimodal feature extraction and leverage rich alignment knowledge in CLIP's image-text shared-embedding space. RISCLIP exhibits outstanding results on all three major RIS benchmarks and also outperforms previous CLIP-based methods, demonstrating the efficacy of our strategy in extending CLIP's image-text alignment to RIS.

Seoyeon Kim, Minguk Kang, Dongwon Kim, Jaesik Park, Suha Kwak• 2023

Related benchmarks

TaskDatasetResultRank
Referring Image SegmentationRefCOCO+ (test-B)
mIoU60.7
200
Referring Image SegmentationRefCOCO (val)--
197
Referring Image SegmentationRefCOCO (test A)
mIoU78
178
Referring Image SegmentationRefCOCO (test-B)--
119
Referring Image SegmentationRefCOCO+ (val)--
117
Referring Image SegmentationRefCOCO+ (testA)
mIoU73.5
45
Referring Image SegmentationG-Ref u (val)
IoU67.6
19
Referring Image SegmentationG-Ref u (test)
mIoU68
16
Showing 8 of 8 rows

Other info

Follow for update