Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

World Model on Million-Length Video And Language With Blockwise RingAttention

About

Enabling long-context understanding remains a key challenge in scaling existing sequence models -- a crucial component in developing generally intelligent models that can process and operate over long temporal horizons that potentially consist of millions of tokens. In this paper, we aim to address these challenges by providing a comprehensive exploration of the full development process for producing 1M context language models and video-language models, setting new benchmarks in language retrieval and new capabilities in long video understanding. We detail our long context data curation process, progressive context extension from 4K to 1M tokens, and present an efficient open-source implementation for scalable training on long sequences. Additionally, we open-source a family of 7B parameter models capable of processing long text documents and videos exceeding 1M tokens.

Hao Liu, Wilson Yan, Matei Zaharia, Pieter Abbeel• 2024

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVizWiz
Accuracy11.6
1525
Object Hallucination EvaluationPOPE
Accuracy75.2
1455
Visual Question AnsweringGQA
Accuracy55.8
1249
Text-based Visual Question AnsweringTextVQA
Accuracy18.8
807
Multimodal UnderstandingMM-Vet
MM-Vet Score9.6
531
Text-to-Image GenerationGenEval
Overall Score47
506
Visual Question AnsweringGQA
Accuracy44.8
505
Science Question AnsweringScienceQA--
502
Video Question AnsweringMSRVTT-QA
Accuracy44.1
491
Text-to-Image GenerationGenEval
Overall Score47
391
Showing 10 of 41 rows

Other info

Follow for update