Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Dual Contradistinctive Generative Autoencoder

About

We present a new generative autoencoder model with dual contradistinctive losses to improve generative autoencoder that performs simultaneous inference (reconstruction) and synthesis (sampling). Our model, named dual contradistinctive generative autoencoder (DC-VAE), integrates an instance-level discriminative loss (maintaining the instance-level fidelity for the reconstruction/synthesis) with a set-level adversarial loss (encouraging the set-level fidelity for there construction/synthesis), both being contradistinctive. Extensive experimental results by DC-VAE across different resolutions including 32x32, 64x64, 128x128, and 512x512 are reported. The two contradistinctive losses in VAE work harmoniously in DC-VAE leading to a significant qualitative and quantitative performance enhancement over the baseline VAEs without architectural changes. State-of-the-art or competitive results among generative autoencoders for image reconstruction, image synthesis, image interpolation, and representation learning are observed. DC-VAE is a general-purpose VAE model, applicable to a wide variety of downstream tasks in computer vision and machine learning.

Gaurav Parmar, Dacheng Li, Kwonjoon Lee, Zhuowen Tu• 2020

Related benchmarks

TaskDatasetResultRank
Image GenerationCIFAR-10 (test)
FID17.9
471
Unconditional Image GenerationCIFAR-10 unconditional--
159
Image GenerationCelebA
FID14.3
110
Image GenerationSTL-10 (test)
Inception Score8.4
59
Image GenerationLSUN bedroom
FID14.3
56
Image GenerationCelebA-HQ 256x256
FID15.8
51
Image GenerationCelebA (test)
FID19.9
49
Image GenerationCelebA-HQ (test)
FID15.81
42
Image GenerationCelebA-HQ 256x256 (test)
FID15.8
34
Unconditional image synthesisCelebA-HQ 256 x 256 (test)
FID15.8
22
Showing 10 of 16 rows

Other info

Follow for update