Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Generative Modeling by Estimating Gradients of the Data Distribution

About

We introduce a new generative model where samples are produced via Langevin dynamics using gradients of the data distribution estimated with score matching. Because gradients can be ill-defined and hard to estimate when the data resides on low-dimensional manifolds, we perturb the data with different levels of Gaussian noise, and jointly estimate the corresponding scores, i.e., the vector fields of gradients of the perturbed data distribution for all noise levels. For sampling, we propose an annealed Langevin dynamics where we use gradients corresponding to gradually decreasing noise levels as the sampling process gets closer to the data manifold. Our framework allows flexible model architectures, requires no sampling during training or the use of adversarial methods, and provides a learning objective that can be used for principled model comparisons. Our models produce samples comparable to GANs on MNIST, CelebA and CIFAR-10 datasets, achieving a new state-of-the-art inception score of 8.87 on CIFAR-10. Additionally, we demonstrate that our models learn effective representations via image inpainting experiments.

Yang Song, Stefano Ermon• 2019

Related benchmarks

TaskDatasetResultRank
Image GenerationCIFAR-10 (test)
FID25.32
471
Unconditional Image GenerationCIFAR-10 (test)
FID25.32
216
Image GenerationCelebA 64 x 64 (test)
FID25.3
203
Unconditional Image GenerationCIFAR-10 unconditional
FID25.32
159
Image GenerationCIFAR10 32x32 (test)
FID25.3
154
Unconditional GenerationCIFAR-10 (test)
FID25.3
102
Image SynthesisCIFAR-10
FID25.32
79
Image GenerationCIFAR-10 (train/test)
FID25.32
78
SuperresolutionCelebA-HQ (test)
PSNR26.83
25
Image GenerationCIFAR-10
FID25.32
25
Showing 10 of 14 rows

Other info

Follow for update