Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

DEGAS: Differentiable Efficient Generator Search

About

Network architecture search (NAS) achieves state-of-the-art results in various tasks such as classification and semantic segmentation. Recently, a reinforcement learning-based approach has been proposed for Generative Adversarial Networks (GANs) search. In this work, we propose an alternative strategy for GAN search by using a method called DEGAS (Differentiable Efficient GenerAtor Search), which focuses on efficiently finding the generator in the GAN. Our search algorithm is inspired by the differential architecture search strategy and the Global Latent Optimization (GLO) procedure. This leads to both an efficient and stable GAN search. After the generator architecture is found, it can be plugged into any existing framework for GAN training. For CTGAN, which we use in this work, the new model outperforms the original inception score results by 0.25 for CIFAR-10 and 0.77 for STL. It also gets better results than the RL based GAN search methods in shorter search time.

Sivan Doveh, Raja Giryes• 2019

Related benchmarks

TaskDatasetResultRank
Image GenerationCIFAR-10 (test)
FID12.01
471
Image GenerationCIFAR-10 (train/test)
FID9.83
78
Image GenerationSTL-10 (train test)
FID28.76
15
Unsupervised Image GenerationSTL (unsupervised)
FID28.76
6
Showing 4 of 4 rows

Other info

Follow for update