Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Conditional Image Generation by Conditioning Variational Auto-Encoders

About

We present a conditional variational auto-encoder (VAE) which, to avoid the substantial cost of training from scratch, uses an architecture and training objective capable of leveraging a foundation model in the form of a pretrained unconditional VAE. To train the conditional VAE, we only need to train an artifact to perform amortized inference over the unconditional VAE's latent variables given a conditioning input. We demonstrate our approach on tasks including image inpainting, for which it outperforms state-of-the-art GAN-based approaches at faithfully representing the inherent uncertainty. We conclude by describing a possible application of our inpainting model, in which it is used to perform Bayesian experimental design for the purpose of guiding a sensor.

William Harvey, Saeid Naderiparizi, Frank Wood• 2021

Related benchmarks

TaskDatasetResultRank
Sequential RecommendationML 1M
NDCG@100.2898
130
Sequential RecommendationML-10M
HR@559.41
15
Sequential RecommendationMusic4All
HR@50.4957
15
Sequential RecommendationGoodreads
HR@544.68
15
Showing 4 of 4 rows

Other info

Follow for update