Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Video Diffusion Models

About

Generating temporally coherent high fidelity video is an important milestone in generative modeling research. We make progress towards this milestone by proposing a diffusion model for video generation that shows very promising initial results. Our model is a natural extension of the standard image diffusion architecture, and it enables jointly training from image and video data, which we find to reduce the variance of minibatch gradients and speed up optimization. To generate long and higher resolution videos we introduce a new conditional sampling technique for spatial and temporal video extension that performs better than previously proposed methods. We present the first results on a large text-conditioned video generation task, as well as state-of-the-art results on established benchmarks for video prediction and unconditional video generation. Supplementary material is available at https://video-diffusion.github.io/

Jonathan Ho, Tim Salimans, Alexey Gritsenko, William Chan, Mohammad Norouzi, David J. Fleet• 2022

Related benchmarks

TaskDatasetResultRank
Video GenerationUCF-101 (test)
Inception Score57.8
105
Video GenerationUCF101--
54
Video PredictionKinetics-600 (test)
FVD16.2
46
Video PredictionBAIR Robot Pushing
FVD66.92
38
Video Frame PredictionKinetics-600
gFVD16.2
28
Video GenerationVideo Generation
Sampling Time (s)125
21
Video PredictionKinetics-600
FVD16.2
18
Video GenerationUCF101 128x128 16 frames
Inception Score57
17
Text-to-Video GenerationUCF-101 (fine-tuning)
IS57.8
13
Conditional Video GenerationKinetics600 (test)
FVD (50k)16.6
10
Showing 10 of 17 rows

Other info

Code

Follow for update