Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

TextHawk2: A Large Vision-Language Model Excels in Bilingual OCR and Grounding with 16x Fewer Tokens

About

Reading dense text and locating objects within images are fundamental abilities for Large Vision-Language Models (LVLMs) tasked with advanced jobs. Previous LVLMs, including superior proprietary models like GPT-4o, have struggled to excel in both tasks simultaneously. Moreover, previous LVLMs with fine-grained perception cost thousands of tokens per image, making them resource-intensive. We present TextHawk2, a bilingual LVLM featuring efficient fine-grained perception and demonstrating cutting-edge performance across general-purpose, OCR, and grounding tasks with 16 times fewer image tokens. Critical improvements include: (1) Token Compression: Building on the efficient architecture of its predecessor, TextHawk2 significantly reduces the number of tokens per image by 16 times, facilitating training and deployment of the TextHawk series with minimal resources. (2) Visual Encoder Reinforcement: We enhance the visual encoder through LVLM co-training, unlocking its potential for previously unseen tasks like Chinese OCR and grounding. (3) Data Diversity: We maintain a comparable scale of 100 million samples while diversifying the sources of pre-training data. We assess TextHawk2 across multiple benchmarks, where it consistently delivers superior performance and outperforms closed-source models of similar scale, such as achieving 78.4% accuracy on OCRBench, 81.4% accuracy on ChartQA, 89.6% ANLS on DocVQA, and 88.1% accuracy@0.5 on RefCOCOg-test.

Ya-Qi Yu, Minghui Liao, Jiwen Zhang, Jihao Wu• 2024

Related benchmarks

TaskDatasetResultRank
Text-based Visual Question AnsweringTextVQA
Accuracy75.1
807
Chart Question AnsweringChartQA
Accuracy81.4
356
Referring Expression ComprehensionRefCOCO+ (val)
Accuracy86.2
354
Referring Expression ComprehensionRefCOCO (val)
Accuracy91.9
344
Referring Expression ComprehensionRefCOCO (testA)
Accuracy0.93
342
Referring Expression ComprehensionRefCOCOg (test)
Accuracy88.1
300
Referring Expression ComprehensionRefCOCOg (val)
Accuracy88.2
300
Document Visual Question AnsweringDocVQA
ANLS89.6
263
Text-based Visual Question AnsweringTextVQA (val)
Accuracy75.1
262
Referring Expression ComprehensionRefCOCO+ (testB)
Accuracy80.4
244
Showing 10 of 19 rows

Other info

Follow for update