Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Shikra: Unleashing Multimodal LLM's Referential Dialogue Magic

About

In human conversations, individuals can indicate relevant regions within a scene while addressing others. In turn, the other person can then respond by referring to specific regions if necessary. This natural referential ability in dialogue remains absent in current Multimodal Large Language Models (MLLMs). To fill this gap, this paper proposes an MLLM called Shikra, which can handle spatial coordinate inputs and outputs in natural language. Its architecture consists of a vision encoder, an alignment layer, and a LLM. It is designed to be straightforward and simple, without the need for extra vocabularies, position encoder, pre-/post-detection modules, or external plug-in models. All inputs and outputs are in natural language form. Referential dialogue is a superset of various vision-language (VL) tasks. Shikra can naturally handle location-related tasks like REC and PointQA, as well as conventional VL tasks such as Image Captioning and VQA. Experimental results showcase Shikra's promising performance. Furthermore, it enables numerous exciting applications, like providing mentioned objects' coordinates in chains of thoughts and comparing user-pointed regions similarities. Our code, model and dataset are accessed at https://github.com/shikras/shikra.

Keqin Chen, Zhao Zhang, Weili Zeng, Richong Zhang, Feng Zhu, Rui Zhao• 2023

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVQA v2
Accuracy77.4
1165
Visual Question AnsweringGQA
Accuracy58.8
963
Object Hallucination EvaluationPOPE
Accuracy84.7
935
Image CaptioningMS COCO Karpathy (test)
CIDEr1.175
682
Visual Question AnsweringVQA v2 (test-dev)
Overall Accuracy83.3
664
Object DetectionCOCO (val)--
613
Multimodal EvaluationMME--
557
Multimodal UnderstandingMMBench
Accuracy58.8
367
Referring Expression ComprehensionRefCOCO+ (val)
Accuracy82.89
345
Referring Expression ComprehensionRefCOCO (val)
Accuracy87.83
335
Showing 10 of 174 rows
...

Other info

Code

Follow for update