xGen-MM-Vid (BLIP-3-Video): You Only Need 32 Tokens to Represent a Video Even in VLMs
About
We present xGen-MM-Vid (BLIP-3-Video): a multimodal language model for videos, particularly designed to efficiently capture temporal information over multiple frames. BLIP-3-Video takes advantage of the 'temporal encoder' in addition to the conventional visual tokenizer, which maps a sequence of tokens over multiple frames into a compact set of visual tokens. This enables BLIP3-Video to use much fewer visual tokens than its competing models (e.g., 32 vs. 4608 tokens). We explore different types of temporal encoders, including learnable spatio-temporal pooling as well as sequential models like Token Turing Machines. We experimentally confirm that BLIP-3-Video obtains video question-answering accuracies comparable to much larger state-of-the-art models (e.g., 34B), while being much smaller (i.e., 4B) and more efficient by using fewer visual tokens. The project website is at https://www.salesforceairesearch.com/opensource/xGen-MM-Vid/index.html
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Video Question Answering | ActivityNet-QA | Accuracy56.9 | 319 | |
| Video Question Answering | NEXT-QA | Overall Accuracy77.1 | 105 | |
| Video Question Answering | MSVD-QA zero-shot (test) | Accuracy77.7 | 56 | |
| Video Question Answering | ActivityNet-QA zero-shot (test) | Accuracy55.7 | 55 | |
| Video Question Answering | MSRVTT-QA zero-shot (test) | Accuracy60 | 55 | |
| Video Question Answering | TGIF-QA zero-shot (test) | Accuracy76.5 | 15 |