Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Temporal Generative Adversarial Nets with Singular Value Clipping

About

In this paper, we propose a generative model, Temporal Generative Adversarial Nets (TGAN), which can learn a semantic representation of unlabeled videos, and is capable of generating videos. Unlike existing Generative Adversarial Nets (GAN)-based methods that generate videos with a single generator consisting of 3D deconvolutional layers, our model exploits two different types of generators: a temporal generator and an image generator. The temporal generator takes a single latent variable as input and outputs a set of latent variables, each of which corresponds to an image frame in a video. The image generator transforms a set of such latent variables into a video. To deal with instability in training of GAN with such advanced networks, we adopt a recently proposed model, Wasserstein GAN, and propose a novel method to train it stably in an end-to-end manner. The experimental results demonstrate the effectiveness of our methods.

Masaki Saito, Eiichi Matsumoto, Shunta Saito• 2016

Related benchmarks

TaskDatasetResultRank
Video GenerationUCF-101 (test)
Inception Score15.83
105
Video GenerationUCF101
FVD1.32e+3
54
Class-conditioned Video GenerationUCF101 (test)--
19
Video GenerationUCF101 128x128 16 frames
Inception Score11.85
17
Unconditional video synthesisUCF-101 128x128
Inception Score11.85
12
Video GenerationUCF-101 16-frame
IS11.85
12
Video GenerationHuman Actions
Inception Score3.65
9
Text-to-Video GenerationMUG
Image Similarity (IS)4.63
7
Video GenerationMUG (test)
FID97.07
6
Video GenerationWeizmann
FID99.85
6
Showing 10 of 15 rows

Other info

Follow for update