Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

LLaVA-Phi: Efficient Multi-Modal Assistant with Small Language Model

About

In this paper, we introduce LLaVA-$\phi$ (LLaVA-Phi), an efficient multi-modal assistant that harnesses the power of the recently advanced small language model, Phi-2, to facilitate multi-modal dialogues. LLaVA-Phi marks a notable advancement in the realm of compact multi-modal models. It demonstrates that even smaller language models, with as few as 2.7B parameters, can effectively engage in intricate dialogues that integrate both textual and visual elements, provided they are trained with high-quality corpora. Our model delivers commendable performance on publicly available benchmarks that encompass visual comprehension, reasoning, and knowledge-based perception. Beyond its remarkable performance in multi-modal dialogue tasks, our model opens new avenues for applications in time-sensitive environments and systems that require real-time interaction, such as embodied agents. It highlights the potential of smaller language models to achieve sophisticated levels of understanding and interaction, while maintaining greater resource efficiency.The project is available at {https://github.com/zhuyiche/llava-phi}.

Yichen Zhu, Minjie Zhu, Ning Liu, Zhicai Ou, Xiaofeng Mou, Jian Tang• 2024

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVQA v2
Accuracy71.4
1165
Visual Question AnsweringTextVQA
Accuracy48.6
1117
Visual Question AnsweringVizWiz
Accuracy35.9
1043
Visual Question AnsweringGQA
Accuracy56.5
963
Object Hallucination EvaluationPOPE
Accuracy85
935
Visual Question AnsweringVQA v2 (test-dev)
Overall Accuracy71.4
664
Multimodal EvaluationMME
Score1.34e+3
557
Multimodal UnderstandingMM-Vet
MM-Vet Score28.9
418
Multimodal UnderstandingMMBench--
367
Visual Question AnsweringTextVQA (val)
VQA Score48.6
309
Showing 10 of 47 rows

Other info

Code

Follow for update