Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Ablating Concepts in Text-to-Image Diffusion Models

About

Large-scale text-to-image diffusion models can generate high-fidelity images with powerful compositional ability. However, these models are typically trained on an enormous amount of Internet data, often containing copyrighted material, licensed images, and personal photos. Furthermore, they have been found to replicate the style of various living artists or memorize exact training samples. How can we remove such copyrighted concepts or images without retraining the model from scratch? To achieve this goal, we propose an efficient method of ablating concepts in the pretrained model, i.e., preventing the generation of a target concept. Our algorithm learns to match the image distribution for a target style, instance, or text prompt we wish to ablate to the distribution corresponding to an anchor concept. This prevents the model from generating target concepts given its text condition. Extensive experiments show that our method can successfully prevent the generation of the ablated concept while preserving closely related concepts in the model.

Nupur Kumari, Bingliang Zhang, Sheng-Yu Wang, Eli Shechtman, Richard Zhang, Jun-Yan Zhu• 2023

Related benchmarks

TaskDatasetResultRank
Coarse-grained UnlearningImagenette
Atar100
70
Text-to-Image GenerationMSCOCO 30K
FID14.08
42
Concept UnlearningUnlearnDiffAtk
UnlearnDiffAtk0.4296
36
Explicit Content RemovalI2P
Armpits Count153
28
Image GenerationMS-COCO 30k (val)
FID14.13
22
Concept ErasureOxford Flowers (test)
Aer70.2
21
Concept ErasureStanford Dogs (test)
Aer79.2
21
Concept ErasureCUB (test)
Aer77.62
21
Safe Text-to-Image GenerationMMA-Diffusion
Automatic Safety Rate85.5
20
Concept UnlearningRing-a-Bell
Ring-A-Bell Score77.3
20
Showing 10 of 78 rows
...

Other info

Follow for update