Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

DocLayLLM: An Efficient Multi-modal Extension of Large Language Models for Text-rich Document Understanding

About

Text-rich document understanding (TDU) requires comprehensive analysis of documents containing substantial textual content and complex layouts. While Multimodal Large Language Models (MLLMs) have achieved fast progress in this domain, existing approaches either demand significant computational resources or struggle with effective multi-modal integration. In this paper, we introduce DocLayLLM, an efficient multi-modal extension of LLMs specifically designed for TDU. By lightly integrating visual patch tokens and 2D positional tokens into LLMs' input and encoding the document content using the LLMs themselves, we fully take advantage of the document comprehension capability of LLMs and enhance their perception of OCR information. We have also deeply considered the role of chain-of-thought (CoT) and innovatively proposed the techniques of CoT Pre-training and CoT Annealing. Our DocLayLLM can achieve remarkable performances with lightweight training settings, showcasing its efficiency and effectiveness. Experimental results demonstrate that our DocLayLLM outperforms existing OCR-dependent methods and OCR-free competitors. Code and model are available at https://github.com/whlscut/DocLayLLM.

Wenhui Liao, Jiapeng Wang, Hongliang Li, Chengyu Wang, Jun Huang, Lianwen Jin• 2024

Related benchmarks

TaskDatasetResultRank
Document Visual Question AnsweringDocVQA
ANLS78.4
164
Document Visual Question AnsweringVisualMRC
ANLS55
12
Document Visual Question AnsweringFUNSD
ANLS84.1
12
Document Visual Question AnsweringCORD
ANLS71.3
12
Document Visual Question AnsweringSROIE
ANLS84.3
12
Showing 5 of 5 rows

Other info

Follow for update