Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

MM-Instruct: Generated Visual Instructions for Large Multimodal Model Alignment

About

This paper introduces MM-Instruct, a large-scale dataset of diverse and high-quality visual instruction data designed to enhance the instruction-following capabilities of large multimodal models (LMMs). While existing visual instruction datasets often focus on question-answering, they struggle to generalize to broader application scenarios such as creative writing, summarization, or image analysis. To address these limitations, we propose a novel approach to constructing MM-Instruct that leverages the strong instruction-following capabilities of existing LLMs to generate novel visual instruction data from large-scale but conventional image captioning datasets. MM-Instruct first leverages ChatGPT to automatically generate diverse instructions from a small set of seed instructions through augmenting and summarization. It then matches these instructions with images and uses an open-sourced large language model (LLM) to generate coherent answers to the instruction-image pairs. The LLM is grounded by the detailed text descriptions of images in the whole answer generation process to guarantee the alignment of the instruction data. Moreover, we introduce a benchmark based on the generated instruction data to evaluate the instruction-following capabilities of existing LMMs. We demonstrate the effectiveness of MM-Instruct by training a LLaVA-1.5 model on the generated data, denoted as LLaVA-Instruct, which exhibits significant improvements in instruction-following capabilities compared to LLaVA-1.5 models. The MM-Instruct dataset, benchmark, and pre-trained models are available at https://github.com/jihaonew/MM-Instruct.

Jihao Liu, Xin Huang, Jinliang Zheng, Boxiao Liu, Jia Wang, Osamu Yoshie, Yu Liu, Hongsheng Li• 2024

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVizWiz
Accuracy56.3
1525
Object Hallucination EvaluationPOPE--
1455
Visual Question AnsweringVQA v2
Accuracy80.3
1362
Visual Question AnsweringTextVQA
Accuracy61.8
1285
Visual Question AnsweringGQA
Accuracy63.8
1249
Multimodal UnderstandingMMBench
Accuracy67.9
637
Multimodal UnderstandingMM-Vet
MM-Vet Score37.1
531
Science Question AnsweringScienceQA IMG
Accuracy70.8
294
Multimodal UnderstandingMME
MME Score1.57e+3
207
Multimodal UnderstandingSEED-Bench Image
Accuracy69.4
121
Showing 10 of 12 rows

Other info

Code

Follow for update