Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Instruction-Guided Visual Masking

About

Instruction following is crucial in contemporary LLM. However, when extended to multimodal setting, it often suffers from misalignment between specific textual instruction and targeted local region of an image. To achieve more accurate and nuanced multimodal instruction following, we introduce Instruction-guided Visual Masking (IVM), a new versatile visual grounding model that is compatible with diverse multimodal models, such as LMM and robot model. By constructing visual masks for instruction-irrelevant regions, IVM-enhanced multimodal models can effectively focus on task-relevant image regions to better align with complex instructions. Specifically, we design a visual masking data generation pipeline and create an IVM-Mix-1M dataset with 1 million image-instruction pairs. We further introduce a new learning technique, Discriminator Weighted Supervised Learning (DWSL) for preferential IVM training that prioritizes high-quality data samples. Experimental results on generic multimodal tasks such as VQA and embodied robotic control demonstrate the versatility of IVM, which as a plug-and-play tool, significantly boosts the performance of diverse multimodal models, yielding new state-of-the-art results across challenging multimodal benchmarks. Code, model and data are available at https://github.com/2toinf/IVM.

Jinliang Zheng, Jianxiong Li, Sijie Cheng, Yinan Zheng, Jiaming Li, Jihao Liu, Yu Liu, Jingjing Liu, Xianyuan Zhan• 2024

Related benchmarks

TaskDatasetResultRank
Referring Expression ComprehensionRefCOCO+ (val)--
345
Referring Expression ComprehensionRefCOCO (val)--
335
Referring Expression ComprehensionRefCOCOg (val)--
291
Science Question AnsweringScienceQA (test)
Average Accuracy70.2
208
Visual Question AnsweringVQA v2 (test)
Accuracy79
131
Visual Question AnsweringGQA (test)
Accuracy62.2
119
Object Hallucination EvaluationPOPE (test)
Accuracy87.2
44
Multi-modal EvaluationMME (test)--
32
Visual SearchV* bench (test)
Attribute Rate87
10
Ego-centric Visual ReasoningEgoThink (test)
Score54.5
4
Showing 10 of 10 rows

Other info

Code

Follow for update