Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ImageBind-LLM: Multi-modality Instruction Tuning

About

We present ImageBind-LLM, a multi-modality instruction tuning method of large language models (LLMs) via ImageBind. Existing works mainly focus on language and image instruction tuning, different from which, our ImageBind-LLM can respond to multi-modality conditions, including audio, 3D point clouds, video, and their embedding-space arithmetic by only image-text alignment training. During training, we adopt a learnable bind network to align the embedding space between LLaMA and ImageBind's image encoder. Then, the image features transformed by the bind network are added to word tokens of all layers in LLaMA, which progressively injects visual instructions via an attention-free and zero-initialized gating mechanism. Aided by the joint embedding of ImageBind, the simple image-text training enables our model to exhibit superior multi-modality instruction-following capabilities. During inference, the multi-modality inputs are fed into the corresponding ImageBind encoders, and processed by a proposed visual cache model for further cross-modal embedding enhancement. The training-free cache model retrieves from three million image features extracted by ImageBind, which effectively mitigates the training-inference modality discrepancy. Notably, with our approach, ImageBind-LLM can respond to instructions of diverse modalities and demonstrate significant language generation quality. Code is released at https://github.com/OpenGVLab/LLaMA-Adapter.

Jiaming Han, Renrui Zhang, Wenqi Shao, Peng Gao, Peng Xu, Han Xiao, Kaipeng Zhang, Chris Liu, Song Wen, Ziyu Guo, Xudong Lu, Shuai Ren, Yafei Wen, Xiaoxin Chen, Xiangyu Yue, Hongsheng Li, Yu Qiao• 2023

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringTextVQA
Accuracy24
1117
Visual Question AnsweringVizWiz
Accuracy51.4
1043
Visual Question AnsweringGQA
Accuracy41.1
963
Multimodal EvaluationMME
Score775.7
557
Multimodal UnderstandingMMBench
Accuracy23.5
367
Visual Question AnsweringOKVQA
Top-1 Accuracy51.66
283
Visual Question AnsweringScienceQA
Accuracy51.4
210
Multimodal UnderstandingMME
MME Score775.7
158
Visual Question AnsweringGQA (test)
Accuracy41.2
119
Image CaptioningFlickr30K
CIDEr Score23.5
111
Showing 10 of 19 rows

Other info

Follow for update