Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

World Model on Million-Length Video And Language With Blockwise RingAttention

About

Enabling long-context understanding remains a key challenge in scaling existing sequence models -- a crucial component in developing generally intelligent models that can process and operate over long temporal horizons that potentially consist of millions of tokens. In this paper, we aim to address these challenges by providing a comprehensive exploration of the full development process for producing 1M context language models and video-language models, setting new benchmarks in language retrieval and new capabilities in long video understanding. We detail our long context data curation process, progressive context extension from 4K to 1M tokens, and present an efficient open-source implementation for scalable training on long sequences. Additionally, we open-source a family of 7B parameter models capable of processing long text documents and videos exceeding 1M tokens.

Hao Liu, Wilson Yan, Matei Zaharia, Pieter Abbeel• 2024

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVizWiz
Accuracy11.6
1043
Visual Question AnsweringGQA
Accuracy55.8
963
Object Hallucination EvaluationPOPE
Accuracy75.2
935
Text-based Visual Question AnsweringTextVQA
Accuracy18.8
496
Video Question AnsweringMSRVTT-QA
Accuracy44.1
481
Text-to-Image GenerationGenEval
Overall Score47
467
Multimodal UnderstandingMM-Vet
MM-Vet Score9.6
418
Visual Question AnsweringGQA
Accuracy44.8
374
Video Question AnsweringMSVD-QA
Accuracy55.9
340
Multimodal Capability EvaluationMM-Vet
Score9.6
282
Showing 10 of 31 rows

Other info

Follow for update