SlowFast-LLaVA: A Strong Training-Free Baseline for Video Large Language Models
About
We propose SlowFast-LLaVA (or SF-LLaVA for short), a training-free video large language model (LLM) that can jointly capture detailed spatial semantics and long-range temporal context without exceeding the token budget of commonly used LLMs. This is realized by using a two-stream SlowFast design of inputs for Video LLMs to aggregate features from sampled frames in an effective way. Specifically, the Slow pathway extracts features at a low frame rate while keeping as much spatial detail as possible (e.g., with 12x24 tokens), and the Fast pathway operates on a high frame rate but uses a larger spatial pooling stride (e.g., downsampling 6x) to focus on the motion cues. As a result, this design allows us to adequately capture both spatial and temporal features that are beneficial for detailed video understanding. Experimental results show that SF-LLaVA outperforms existing training-free methods on a wide range of video tasks. On some benchmarks, it achieves comparable or even better performance compared to state-of-the-art Video LLMs that are fine-tuned on video datasets. Code has been made available at: https://github.com/apple/ml-slowfast-llava.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Video Question Answering | MSRVTT-QA | Accuracy67.4 | 481 | |
| Video Question Answering | MSRVTT-QA (test) | Accuracy67.4 | 371 | |
| Video Question Answering | ActivityNet-QA (test) | Accuracy59.2 | 275 | |
| Video Question Answering | MSVD-QA (test) | Accuracy79.9 | 274 | |
| Video Question Answering | NExT-QA (test) | Accuracy72 | 204 | |
| Video Question Answering | NExT-QA (val) | Overall Acc64.2 | 176 | |
| Video Question Answering | TGIF-QA | Accuracy80.6 | 147 | |
| Video Question Answering | NExT-QA Multi-choice | Accuracy72 | 102 | |
| Video Question Answering | TGIF-QA (test) | Accuracy80.6 | 89 | |
| Video Question Answering | EgoSchema | Accuracy55.8 | 88 |