Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Grounded SAM: Assembling Open-World Models for Diverse Visual Tasks

About

We introduce Grounded SAM, which uses Grounding DINO as an open-set object detector to combine with the segment anything model (SAM). This integration enables the detection and segmentation of any regions based on arbitrary text inputs and opens a door to connecting various vision models. As shown in Fig.1, a wide range of vision tasks can be achieved by using the versatile Grounded SAM pipeline. For example, an automatic annotation pipeline based solely on input images can be realized by incorporating models such as BLIP and Recognize Anything. Additionally, incorporating Stable-Diffusion allows for controllable image editing, while the integration of OSX facilitates promptable 3D human motion analysis. Grounded SAM also shows superior performance on open-vocabulary benchmarks, achieving 48.7 mean AP on SegInW (Segmentation in the wild) zero-shot benchmark with the combination of Grounding DINO-Base and SAM-Huge models.

Tianhe Ren, Shilong Liu, Ailing Zeng, Jing Lin, Kunchang Li, He Cao, Jiayu Chen, Xinyu Huang, Yukang Chen, Feng Yan, Zhaoyang Zeng, Hao Zhang, Feng Li, Jie Yang, Hongyang Li, Qing Jiang, Lei Zhang• 2024

Related benchmarks

TaskDatasetResultRank
Semantic segmentationADE20K (val)
mIoU4.8
2731
Semantic segmentationPASCAL VOC (val)
mIoU46.3
338
Semantic segmentationPASCAL Context (val)
mIoU28.1
323
Referring Expression SegmentationRefCOCO+ (val)--
201
Referring Expression SegmentationRefCOCO+ (testA)--
190
Referring Expression SegmentationRefCOCO+ (testB)--
188
Reasoning SegmentationReasonSeg (val)
cIoU14.5
145
Semantic segmentationCOCO (val)
mIoU35.7
135
Semantic segmentationCOCO Stuff (val)
mIoU18.8
126
Reasoning SegmentationReasonSeg (test)
gIoU21.3
102
Showing 10 of 42 rows

Other info

Follow for update