Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Perception Prioritized Training of Diffusion Models

About

Diffusion models learn to restore noisy data, which is corrupted with different levels of noise, by optimizing the weighted sum of the corresponding loss terms, i.e., denoising score matching loss. In this paper, we show that restoring data corrupted with certain noise levels offers a proper pretext task for the model to learn rich visual concepts. We propose to prioritize such noise levels over other levels during training, by redesigning the weighting scheme of the objective function. We show that our simple redesign of the weighting scheme significantly improves the performance of diffusion models regardless of the datasets, architectures, and sampling strategies.

Jooyoung Choi, Jungbeom Lee, Chaehun Shin, Sungwon Kim, Hyunwoo Kim, Sungroh Yoon• 2022

Related benchmarks

TaskDatasetResultRank
Unconditional Image GenerationCIFAR-10 (test)
FID5.63
223
Text-to-Image GenerationMS-COCO
FID13.23
131
Unconditional Image GenerationCelebA unconditional 64 x 64
FID7.22
95
Image GenerationFFHQ
FID6.92
91
Image GenerationCelebA-HQ
FID6.91
79
Unconditional Image GenerationFFHQ 256x256
FID6.97
64
Text-to-Image GenerationPartiPrompts
CLIP Score29.5
26
Unconditional Image GenerationFFHQ 256x256 (test)
FID7
25
Unconditional Image GenerationLSUN Church (test)
FID10.77
17
Unconditional Image GenerationLSUN Bedroom (test)
FID6.53
14
Showing 10 of 12 rows

Other info

Code

Follow for update