Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Video-GPT via Next Clip Diffusion

About

GPT has shown its remarkable success in natural language processing. However, the language sequence is not sufficient to describe spatial-temporal details in the visual world. Alternatively, the video sequence is good at capturing such details. Motivated by this fact, we propose a concise Video-GPT in this paper by treating video as new language for visual world modeling. By analogy to next token prediction in GPT, we introduce a novel next clip diffusion paradigm for pretraining Video-GPT. Different from the previous works, this distinct paradigm allows Video-GPT to tackle both short-term generation and long-term prediction, by autoregressively denoising the noisy clip according to the clean clips in the history. Extensive experiments show our Video-GPT achieves the state-of-the-art performance on video prediction, which is the key factor towards world modeling (Physics-IQ Benchmark: Video-GPT 34.97 vs. Kling 23.64 vs. Wan 20.89). Moreover, it can be well adapted on 6 mainstream video tasks in both video generation and understanding, showing its great generalization capacity in downstream. The project page is at https://zhuangshaobin.github.io/Video-GPT.github.io/.

Shaobin Zhuang, Zhipeng Huang, Ying Zhang, Fangyikang Wang, Canmiao Fu, Binxin Yang, Chong Sun, Chen Li, Yali Wang• 2025

Related benchmarks

TaskDatasetResultRank
Text-to-Video RetrievalMSR-VTT (test)
R@122.8
234
Video GenerationPhysics-IQ
Phys. IQ Score34.97
45
Video ClassificationUCF-101--
19
Class-to-video generationUCF-101
FVD53
13
Video GenerationKinetics-600
FVD315.4
4
Showing 5 of 5 rows

Other info

Code

Follow for update