Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Elucidating the Exposure Bias in Diffusion Models

About

Diffusion models have demonstrated impressive generative capabilities, but their \textit{exposure bias} problem, described as the input mismatch between training and sampling, lacks in-depth exploration. In this paper, we systematically investigate the exposure bias problem in diffusion models by first analytically modelling the sampling distribution, based on which we then attribute the prediction error at each sampling step as the root cause of the exposure bias issue. Furthermore, we discuss potential solutions to this issue and propose an intuitive metric for it. Along with the elucidation of exposure bias, we propose a simple, yet effective, training-free method called Epsilon Scaling to alleviate the exposure bias. We show that Epsilon Scaling explicitly moves the sampling trajectory closer to the vector field learned in the training phase by scaling down the network output, mitigating the input mismatch between training and sampling. Experiments on various diffusion frameworks (ADM, DDIM, EDM, LDM, DiT, PFGM++) verify the effectiveness of our method. Remarkably, our ADM-ES, as a state-of-the-art stochastic sampler, obtains 2.17 FID on CIFAR-10 under 100-step unconditional generation. The code is available at \url{https://github.com/forever208/ADM-ES} and \url{https://github.com/forever208/EDM-ES}.

Mang Ning, Mingxiao Li, Jianlin Su, Albert Ali Salah, Itir Onal Ertugrul• 2023

Related benchmarks

TaskDatasetResultRank
Class-conditional Image GenerationImageNet 64x64
FID2.39
126
Unconditional Image GenerationCIFAR-10 32x32 (test)
FID2.17
94
Unconditional Image GenerationCIFAR-10 32 x 32
FID2.17
47
Class-conditional Image GenerationImageNet 128x128
FID3.37
27
Unconditional Image GenerationLSUN tower unconditional 64x64
FID2.91
7
Unconditional Image GenerationFFHQ unconditional 128x128
FID6.77
7
Showing 6 of 6 rows

Other info

Code

Follow for update