Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Feature Quantization Improves GAN Training

About

The instability in GAN training has been a long-standing problem despite remarkable research efforts. We identify that instability issues stem from difficulties of performing feature matching with mini-batch statistics, due to a fragile balance between the fixed target distribution and the progressively generated distribution. In this work, we propose Feature Quantization (FQ) for the discriminator, to embed both true and fake data samples into a shared discrete space. The quantized values of FQ are constructed as an evolving dictionary, which is consistent with feature statistics of the recent distribution history. Hence, FQ implicitly enables robust feature matching in a compact space. Our method can be easily plugged into existing GAN models, with little computational overhead in training. We apply FQ to 3 representative GAN models on 9 benchmarks: BigGAN for image generation, StyleGAN for face synthesis, and U-GAT-IT for unsupervised image-to-image translation. Extensive experimental results show that the proposed FQ-GAN can improve the FID scores of baseline methods by a large margin on a variety of tasks, achieving new state-of-the-art performance.

Yang Zhao, Chunyuan Li, Ping Yu, Jianfeng Gao, Changyou Chen• 2020

Related benchmarks

TaskDatasetResultRank
Image GenerationCIFAR-10 (test)
FID5.34
471
Image GenerationCIFAR-10
Inception Score9.16
178
Image GenerationImageNet 64x64 (train val)
FID8.15
83
Image GenerationCIFAR100
FID8.23
51
Image GenerationImageNet 128x128
FID13.77
51
Image GenerationCIFAR-100 (test)
IS11.05
35
Image GenerationCIFAR10 (train)
FID7.65
32
Image GenerationFFHQ 1024x1024 (train)
FID3.01
23
Image-to-Image Translationselfie2anime
KID0.114
11
Image-to-Image Translationanime2selfie
KID0.1023
10
Showing 10 of 24 rows

Other info

Code

Follow for update