Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

A Bounding Box is Worth One Token: Interleaving Layout and Text in a Large Language Model for Document Understanding

About

Recently, many studies have demonstrated that exclusively incorporating OCR-derived text and spatial layouts with large language models (LLMs) can be highly effective for document understanding tasks. However, existing methods that integrate spatial layouts with text have limitations, such as producing overly long text sequences or failing to fully leverage the autoregressive traits of LLMs. In this work, we introduce Interleaving Layout and Text in a Large Language Model (LayTextLLM)} for document understanding. LayTextLLM projects each bounding box to a single embedding and interleaves it with text, efficiently avoiding long sequence issues while leveraging autoregressive traits of LLMs. LayTextLLM not only streamlines the interaction of layout and textual data but also shows enhanced performance in KIE and VQA. Comprehensive benchmark evaluations reveal significant improvements of LayTextLLM, with a 15.2% increase on KIE tasks and 10.7% on VQA tasks compared to previous SOTA OCR-based LLMs. All resources are available at https://github.com/LayTextLLM/LayTextLLM.

Jinghui Lu, Haiyang Yu, Yanjie Wang, Yongjie Ye, Jingqun Tang, Ziwei Yang, Binghong Wu, Qi Liu, Hao Feng, Han Wang, Hao Liu, Can Huang• 2024

Related benchmarks

TaskDatasetResultRank
Document Visual Question AnsweringDocVQA
ANLS77.2
164
Deepfake DetectionDFDC
AUC69.95
135
Deepfake DetectionDFD
AUC0.812
77
Deepfake DetectionFF++ Intra-dataset c23
AUC98.91
24
Image Deepfake DetectionDFo
AUC0.8638
20
Deepfake DetectionDFDCP--
20
Document Visual Question AnsweringSROIE
ANLS96.1
12
Deepfake DetectionCDF
AUC75.52
12
Document Visual Question AnsweringCORD
ANLS82.5
12
Document Visual Question AnsweringFUNSD
ANLS81
12
Showing 10 of 15 rows

Other info

Follow for update