World Model on Million-Length Video And Language With Blockwise RingAttention
About
Enabling long-context understanding remains a key challenge in scaling existing sequence models -- a crucial component in developing generally intelligent models that can process and operate over long temporal horizons that potentially consist of millions of tokens. In this paper, we aim to address these challenges by providing a comprehensive exploration of the full development process for producing 1M context language models and video-language models, setting new benchmarks in language retrieval and new capabilities in long video understanding. We detail our long context data curation process, progressive context extension from 4K to 1M tokens, and present an efficient open-source implementation for scalable training on long sequences. Additionally, we open-source a family of 7B parameter models capable of processing long text documents and videos exceeding 1M tokens.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Visual Question Answering | VizWiz | Accuracy11.6 | 1525 | |
| Object Hallucination Evaluation | POPE | Accuracy75.2 | 1455 | |
| Visual Question Answering | GQA | Accuracy55.8 | 1249 | |
| Text-based Visual Question Answering | TextVQA | Accuracy18.8 | 807 | |
| Multimodal Understanding | MM-Vet | MM-Vet Score9.6 | 531 | |
| Text-to-Image Generation | GenEval | Overall Score47 | 506 | |
| Visual Question Answering | GQA | Accuracy44.8 | 505 | |
| Science Question Answering | ScienceQA | -- | 502 | |
| Video Question Answering | MSRVTT-QA | Accuracy44.1 | 491 | |
| Text-to-Image Generation | GenEval | Overall Score47 | 391 |