Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Vary: Scaling up the Vision Vocabulary for Large Vision-Language Models

About

Modern Large Vision-Language Models (LVLMs) enjoy the same vision vocabulary -- CLIP, which can cover most common vision tasks. However, for some special vision task that needs dense and fine-grained vision perception, e.g., document-level OCR or chart understanding, especially in non-English scenarios, the CLIP-style vocabulary may encounter low efficiency in tokenizing the vision knowledge and even suffer out-of-vocabulary problem. Accordingly, we propose Vary, an efficient and effective method to scale up the vision vocabulary of LVLMs. The procedures of Vary are naturally divided into two folds: the generation and integration of a new vision vocabulary. In the first phase, we devise a vocabulary network along with a tiny decoder-only transformer to produce the desired vocabulary via autoregression. In the next, we scale up the vanilla vision vocabulary by merging the new one with the original one (CLIP), enabling the LVLMs can quickly garner new features. Compared to the popular BLIP-2, MiniGPT4, and LLaVA, Vary can maintain its vanilla capabilities while enjoying more excellent fine-grained perception and understanding ability. Specifically, Vary is competent in new document parsing features (OCR or markdown conversion) while achieving 78.2% ANLS in DocVQA and 36.2% in MMVet. Our code will be publicly available on the homepage.

Haoran Wei, Lingyu Kong, Jinyue Chen, Liang Zhao, Zheng Ge, Jinrong Yang, Jianjian Sun, Chunrui Han, Xiangyu Zhang• 2023

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringChartQA
Accuracy65.3
239
Chart Question AnsweringChartQA
Accuracy66.1
229
Document Visual Question AnsweringDocVQA (test)
ANLS76.3
192
Document Visual Question AnsweringDocVQA
ANLS76.3
164
Multimodal UnderstandingMM-VET (test)
Total Score36.2
114
Visual Question AnsweringDocVQA
Accuracy76.3
103
Visual Question AnsweringDocVQA (val)
ANLS78.2
31
Text RecognitionSROIE Task 2 (test)
F1 Score9.84
19
Document Image RetrievalNL-DIR (test)
Recall@10.01
15
Document-level OCRCORD 100 images (test)
F1 Score12.89
5
Showing 10 of 18 rows

Other info

Code

Follow for update