Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

MMFuser: Multimodal Multi-Layer Feature Fuser for Fine-Grained Vision-Language Understanding

About

Despite significant advancements in Multimodal Large Language Models (MLLMs) for understanding complex human intentions through cross-modal interactions, capturing intricate image details remains challenging. Previous methods integrating multiple vision encoders to enhance visual detail introduce redundancy and computational overhead. We observe that most MLLMs utilize only the last-layer feature map of the vision encoder for visual representation, neglecting the rich fine-grained information in shallow feature maps. To address this issue, we propose \modelname, a simple yet effective multi-layer feature fuser that efficiently integrates deep and shallow features from Vision Transformers (ViTs). Specifically, it leverages semantically aligned deep features as queries to dynamically extract missing details from shallow features, thus preserving semantic alignment while enriching the representation with fine-grained information. Applied to the LLaVA-1.5 model, \modelname~achieves significant improvements in visual representation and benchmark performance, providing a more flexible and lightweight solution compared to multi-encoder ensemble methods. The code and model have been released at https://github.com/yuecao0119/MMFuser.

Yue Cao, Yangzhou Liu, Zhe Chen, Guangchen Shi, Wenhai Wang, Danhuai Zhao, Tong Lu• 2024

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVizWiz
Accuracy57.4
1525
Object Hallucination EvaluationPOPE
Accuracy87.5
1455
Visual Question AnsweringTextVQA
Accuracy59.9
1285
Visual Question AnsweringGQA
Accuracy63.4
1249
Multimodal EvaluationMME
Score1.59e+3
658
Visual Question AnsweringScienceQA
Accuracy68.7
370
Referring Expression ComprehensionRefCOCO+ (val)--
354
Multimodal Capability EvaluationMM-Vet
Score36.6
345
Referring Expression ComprehensionRefCOCO (val)--
344
Referring Expression ComprehensionRefCOCO (testA)--
342
Showing 10 of 38 rows

Other info

Code

Follow for update