Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Alleviating Adversarial Attacks on Variational Autoencoders with MCMC

About

Variational autoencoders (VAEs) are latent variable models that can generate complex objects and provide meaningful latent representations. Moreover, they could be further used in downstream tasks such as classification. As previous work has shown, one can easily fool VAEs to produce unexpected latent representations and reconstructions for a visually slightly modified input. Here, we examine several objective functions for adversarial attack construction proposed previously and present a solution to alleviate the effect of these attacks. Our method utilizes the Markov Chain Monte Carlo (MCMC) technique in the inference step that we motivate with a theoretical analysis. Thus, we do not incorporate any extra costs during training, and the performance on non-attacked inputs is not decreased. We validate our approach on a variety of datasets (MNIST, Fashion MNIST, Color MNIST, CelebA) and VAE configurations ($\beta$-VAE, NVAE, $\beta$-TCVAE), and show that our approach consistently improves the model robustness to adversarial attacks.

Anna Kuzina, Max Welling, Jakub M. Tomczak• 2022

Related benchmarks

TaskDatasetResultRank
Color ClassificationColorMNIST
Adversarial Accuracy100
30
Digit ClassificationColorMNIST
Adversarial Accuracy42
30
Generative ModelingColorMNIST
MSE206
10
ReconstructionColorMNIST
MSSSIM0.96
8
Showing 4 of 4 rows

Other info

Code

Follow for update