Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Fine-Grained Visual Prompting

About

Vision-Language Models (VLMs), such as CLIP, have demonstrated impressive zero-shot transfer capabilities in image-level visual perception. However, these models have shown limited performance in instance-level tasks that demand precise localization and recognition. Previous works have suggested that incorporating visual prompts, such as colorful boxes or circles, can improve the ability of models to recognize objects of interest. Nonetheless, compared to language prompting, visual prompting designs are rarely explored. Existing approaches, which employ coarse visual cues such as colorful boxes or circles, often result in sub-optimal performance due to the inclusion of irrelevant and noisy pixels. In this paper, we carefully study the visual prompting designs by exploring more fine-grained markings, such as segmentation masks and their variations. In addition, we introduce a new zero-shot framework that leverages pixel-level annotations acquired from a generalist segmentation model for fine-grained visual prompting. Consequently, our investigation reveals that a straightforward application of blur outside the target mask, referred to as the Blur Reverse Mask, exhibits exceptional effectiveness. This proposed prompting strategy leverages the precise mask annotations to reduce focus on weakly related regions while retaining spatial coherence between the target and the surrounding background. Our Fine-Grained Visual Prompting (FGVP) demonstrates superior performance in zero-shot comprehension of referring expressions on the RefCOCO, RefCOCO+, and RefCOCOg benchmarks. It outperforms prior methods by an average margin of 3.0% to 4.6%, with a maximum improvement of 12.5% on the RefCOCO+ testA subset. Code is available at https://github.com/ylingfeng/FGVP.

Lingfeng Yang, Yueze Wang, Xiang Li, Xinlong Wang, Jian Yang• 2023

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVizWiz
Accuracy70.8
1525
Multimodal UnderstandingMMMU
Accuracy49.6
437
Referring Expression ComprehensionRefCOCO+ (val)
Accuracy50.4
354
Referring Expression ComprehensionRefCOCO (val)
Accuracy46.1
344
Referring Expression ComprehensionRefCOCO (testA)
Accuracy0.53
342
Multimodal UnderstandingMMStar--
324
Referring Expression ComprehensionRefCOCOg (test)
Accuracy54.1
300
Referring Expression ComprehensionRefCOCOg (val)
Accuracy54.5
300
Mathematical ReasoningMathVista
Accuracy68
257
Referring Expression ComprehensionRefCOCO+ (testB)
Accuracy42.6
244
Showing 10 of 21 rows

Other info

Follow for update