Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Selective Amnesia: A Continual Learning Approach to Forgetting in Deep Generative Models

About

The recent proliferation of large-scale text-to-image models has led to growing concerns that such models may be misused to generate harmful, misleading, and inappropriate content. Motivated by this issue, we derive a technique inspired by continual learning to selectively forget concepts in pretrained deep generative models. Our method, dubbed Selective Amnesia, enables controllable forgetting where a user can specify how a concept should be forgotten. Selective Amnesia can be applied to conditional variational likelihood models, which encompass a variety of popular deep generative frameworks, including variational autoencoders and large-scale text-to-image diffusion models. Experiments across different models demonstrate that our approach induces forgetting on a variety of concepts, from entire classes in standard datasets to celebrity and nudity prompts in text-to-image models. Our code is publicly available at https://github.com/clear-nus/selective-amnesia.

Alvin Heng, Harold Soh• 2023

Related benchmarks

TaskDatasetResultRank
Concept UnlearningUnlearnDiffAtk
UnlearnDiffAtk0.268
36
Explicit Content RemovalI2P
Armpits Count72
28
Concept UnlearningRing-a-Bell
Ring-A-Bell Score32.9
20
Safe Text-to-Image GenerationMMA-Diffusion
Automatic Safety Rate20.5
20
Text-to-Image GenerationNon-targeted concepts
CLIP Score30.6
18
Concept UnlearningI2P
I2P0.062
17
Concept UnlearningMMA-Diffusion
MMA-Diffusion20.5
16
Concept UnlearningP4D
P4D0.623
14
Safe Text-to-Image GenerationI2P
ASR0.062
13
Nudity Concept ErasureMMA Adversarial Prompts
Erase Rate (%)89.6
13
Showing 10 of 20 rows

Other info

Follow for update