Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Precision-Recall Divergence Optimization for Generative Modeling with GANs and Normalizing Flows

About

Achieving a balance between image quality (precision) and diversity (recall) is a significant challenge in the domain of generative models. Current state-of-the-art models primarily rely on optimizing heuristics, such as the Fr\'echet Inception Distance. While recent developments have introduced principled methods for evaluating precision and recall, they have yet to be successfully integrated into the training of generative models. Our main contribution is a novel training method for generative models, such as Generative Adversarial Networks and Normalizing Flows, which explicitly optimizes a user-defined trade-off between precision and recall. More precisely, we show that achieving a specified precision-recall trade-off corresponds to minimizing a unique $f$-divergence from a family we call the \textit{PR-divergences}. Conversely, any $f$-divergence can be written as a linear combination of PR-divergences and corresponds to a weighted precision-recall trade-off. Through comprehensive evaluations, we show that our approach improves the performance of existing state-of-the-art models like BigGAN in terms of either precision or recall when tested on datasets such as ImageNet.

Alexandre Verine, Benjamin Negrevergne, Muni Sreenivas Pydi, Yann Chevaleyre• 2023

Related benchmarks

TaskDatasetResultRank
Image GenerationImageNet 128x128
FID9.92
51
Image GenerationCIFAR-10 32x32 (train)
FID11.62
20
Image GenerationFFHQ 256x256
FID35.24
12
Image GenerationCelebA 64x64 (train)
Precision83.96
11
Showing 4 of 4 rows

Other info

Code

Follow for update