Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

LLaMA-VID: An Image is Worth 2 Tokens in Large Language Models

About

In this work, we present a novel method to tackle the token generation challenge in Vision Language Models (VLMs) for video and image understanding, called LLaMA-VID. Current VLMs, while proficient in tasks like image captioning and visual question answering, face computational burdens when processing long videos due to the excessive visual tokens. LLaMA-VID addresses this issue by representing each frame with two distinct tokens, namely context token and content token. The context token encodes the overall image context based on user input, whereas the content token encapsulates visual cues in each frame. This dual-token strategy significantly reduces the overload of long videos while preserving critical information. Generally, LLaMA-VID empowers existing frameworks to support hour-long videos and pushes their upper limit with an extra context token. It is proved to surpass previous methods on most of video- or image-based benchmarks. Code is available https://github.com/dvlab-research/LLaMA-VID}{https://github.com/dvlab-research/LLaMA-VID

Yanwei Li, Chengyao Wang, Jiaya Jia• 2023

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVQA v2
Accuracy80
1165
Visual Question AnsweringVizWiz
Accuracy54.3
1043
Visual Question AnsweringGQA
Accuracy65
963
Object Hallucination EvaluationPOPE
Accuracy79.1
935
Video Question AnsweringMSRVTT-QA
Accuracy58.9
481
Visual Question AnsweringGQA
Accuracy52.9
374
Video Question AnsweringMSRVTT-QA (test)
Accuracy58.9
371
Multimodal UnderstandingMMBench
Accuracy55.5
367
Video Question AnsweringMSVD-QA
Accuracy70
340
Video Question AnsweringActivityNet-QA
Accuracy49.1
319
Showing 10 of 130 rows
...

Other info

Code

Follow for update