Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Make Your Training Flexible: Towards Deployment-Efficient Video Models

About

Popular video training methods mainly operate on a fixed number of tokens sampled from a predetermined spatiotemporal grid, resulting in sub-optimal accuracy-computation trade-offs due to inherent video redundancy. They also lack adaptability to varying computational budgets for downstream tasks, hindering applications of the most competitive model in real-world scenes. We thus propose a new test setting, Token Optimization, for maximized input information across budgets, which optimizes the size-limited set of input tokens through token selection from more suitably sampled videos. To this end, we propose a novel augmentation tool termed Flux. By making the sampling grid flexible and leveraging token selection, it is easily adopted in most popular video training frameworks, boosting model robustness with nearly no additional cost. We integrate Flux in large-scale video pre-training, and the resulting FluxViT establishes new state-of-the-art results across extensive tasks at standard costs. Notably, with 1/4 tokens only, it can still match the performance of previous state-of-the-art models with Token Optimization, yielding nearly 90\% savings. All models and data are available at https://github.com/OpenGVLab/FluxViT.

Chenting Wang, Kunchang Li, Tianxiang Jiang, Xiangyu Zeng, Yi Wang, Limin Wang• 2025

Related benchmarks

TaskDatasetResultRank
Action RecognitionSomething-Something v2 (val)
Top-1 Accuracy75.6
535
Action RecognitionKinetics-400
Top-1 Acc90
413
Text-to-Video RetrievalDiDeMo
R@10.535
360
Text-to-Video RetrievalMSVD
R@154.2
218
Text-to-Video RetrievalActivityNet
R@10.567
197
Text-to-Video RetrievalLSMDC
R@125.4
154
Text-to-Video RetrievalMSRVTT
Recall@149.9
48
Video ClassificationCOIN (test)
Top-1 Accuracy94.1
20
Fine-grained captioningDream1k
F1 Score29.5
11
General spatiotemporal perceptionMVBench
Score49
11
Showing 10 of 10 rows

Other info

Code

Follow for update