Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

InternLM-XComposer-2.5: A Versatile Large Vision Language Model Supporting Long-Contextual Input and Output

About

We present InternLM-XComposer-2.5 (IXC-2.5), a versatile large-vision language model that supports long-contextual input and output. IXC-2.5 excels in various text-image comprehension and composition applications, achieving GPT-4V level capabilities with merely 7B LLM backend. Trained with 24K interleaved image-text contexts, it can seamlessly extend to 96K long contexts via RoPE extrapolation. This long-context capability allows IXC-2.5 to excel in tasks requiring extensive input and output contexts. Compared to its previous 2.0 version, InternLM-XComposer-2.5 features three major upgrades in vision-language comprehension: (1) Ultra-High Resolution Understanding, (2) Fine-Grained Video Understanding, and (3) Multi-Turn Multi-Image Dialogue. In addition to comprehension, IXC-2.5 extends to two compelling applications using extra LoRA parameters for text-image composition: (1) Crafting Webpages and (2) Composing High-Quality Text-Image Articles. IXC-2.5 has been evaluated on 28 benchmarks, outperforming existing open-source state-of-the-art models on 16 benchmarks. It also surpasses or competes closely with GPT-4V and Gemini Pro on 16 key tasks. The InternLM-XComposer-2.5 is publicly available at https://github.com/InternLM/InternLM-XComposer.

Pan Zhang, Xiaoyi Dong, Yuhang Zang, Yuhang Cao, Rui Qian, Lin Chen, Qipeng Guo, Haodong Duan, Bin Wang, Linke Ouyang, Songyang Zhang, Wenwei Zhang, Yining Li, Yang Gao, Peng Sun, Xinyue Zhang, Wei Li, Jingwen Li, Wenhai Wang, Hang Yan, Conghui He, Xingcheng Zhang, Kai Chen, Jifeng Dai, Yu Qiao, Dahua Lin, Jiaqi Wang• 2024

Related benchmarks

TaskDatasetResultRank
Multimodal EvaluationMME--
658
Science Question AnsweringScienceQA--
502
Video UnderstandingMVBench
Accuracy69.1
425
Video Question AnsweringActivityNet-QA
Accuracy52.8
376
Visual Question AnsweringTextVQA (val)
VQA Score78.2
343
OCR EvaluationOCRBench
Score690
329
Multi-discipline Multimodal UnderstandingMMMU
Accuracy42.9
317
Video Question AnsweringActivityNet-QA (test)
Accuracy52.8
288
Visual Mathematical ReasoningMathVista
Accuracy63.7
278
Text-based Visual Question AnsweringTextVQA (val)
Accuracy78.2
262
Showing 10 of 96 rows
...

Other info

Code

Follow for update