Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

LLaVA-Plus: Learning to Use Tools for Creating Multimodal Agents

About

LLaVA-Plus is a general-purpose multimodal assistant that expands the capabilities of large multimodal models. It maintains a skill repository of pre-trained vision and vision-language models and can activate relevant tools based on users' inputs to fulfill real-world tasks. LLaVA-Plus is trained on multimodal instruction-following data to acquire the ability to use tools, covering visual understanding, generation, external knowledge retrieval, and compositions. Empirical results show that LLaVA-Plus outperforms LLaVA in existing capabilities and exhibits new ones. It is distinct in that the image query is directly grounded and actively engaged throughout the entire human-AI interaction sessions, significantly improving tool use performance and enabling new scenarios.

Shilong Liu, Hao Cheng, Haotian Liu, Hao Zhang, Feng Li, Tianhe Ren, Xueyan Zou, Jianwei Yang, Hang Su, Jun Zhu, Lei Zhang, Jianfeng Gao, Chunyuan Li• 2023

Related benchmarks

TaskDatasetResultRank
Multimodal UnderstandingSEED-Bench--
343
Multimodal UnderstandingMM-VET (test)--
120
Multi-modal UnderstandingMMVet
Accuracy35
40
Multimodal Instruction FollowingViSiT-Bench Sept. 27th, 2023 (Leaderboard)
ELO1.20e+3
15
Visual UnderstandingBLINK sub-tasks
Jigsaw Accuracy58
14
Large Multi-modal Model EvaluationLLaVA-Bench Tool Use (test)
Grounding0.893
8
Multimodal Tool UseLLaVA-Bench Tool Use
Grounding89.3
8
Large Multi-modal Model EvaluationLLaVA-Bench In-the-Wild v1
Conversational Score65.5
6
Large Multi-modal Model EvaluationLLaVA-Bench COCO v1
Conv Score0.816
6
Image CaptioningCOCO Caption
BLEU-150.8
3
Showing 10 of 10 rows

Other info

Code

Follow for update