Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Perception Prioritized Training of Diffusion Models

About

Diffusion models learn to restore noisy data, which is corrupted with different levels of noise, by optimizing the weighted sum of the corresponding loss terms, i.e., denoising score matching loss. In this paper, we show that restoring data corrupted with certain noise levels offers a proper pretext task for the model to learn rich visual concepts. We propose to prioritize such noise levels over other levels during training, by redesigning the weighting scheme of the objective function. We show that our simple redesign of the weighting scheme significantly improves the performance of diffusion models regardless of the datasets, architectures, and sampling strategies.

Jooyoung Choi, Jungbeom Lee, Chaehun Shin, Sungwon Kim, Hyunwoo Kim, Sungroh Yoon• 2022

Related benchmarks

TaskDatasetResultRank
Unconditional Image GenerationCIFAR-10 (test)
FID5.63
216
Unconditional Image GenerationCelebA unconditional 64 x 64
FID7.22
95
Text-to-Image GenerationMS-COCO
FID13.23
75
Unconditional Image GenerationFFHQ 256x256
FID6.97
64
Image GenerationFFHQ
FID6.92
52
Text-to-Image GenerationPartiPrompts
CLIP Score29.5
26
Unconditional Image GenerationFFHQ 256x256 (test)
FID7
25
Image GenerationCelebA-HQ
FID6.91
23
Unconditional Image GenerationLSUN Church (test)
FID10.77
17
Unconditional Image GenerationLSUN Bedroom (test)
FID6.53
14
Showing 10 of 12 rows

Other info

Code

Follow for update