Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Ablating Concepts in Text-to-Image Diffusion Models

About

Large-scale text-to-image diffusion models can generate high-fidelity images with powerful compositional ability. However, these models are typically trained on an enormous amount of Internet data, often containing copyrighted material, licensed images, and personal photos. Furthermore, they have been found to replicate the style of various living artists or memorize exact training samples. How can we remove such copyrighted concepts or images without retraining the model from scratch? To achieve this goal, we propose an efficient method of ablating concepts in the pretrained model, i.e., preventing the generation of a target concept. Our algorithm learns to match the image distribution for a target style, instance, or text prompt we wish to ablate to the distribution corresponding to an anchor concept. This prevents the model from generating target concepts given its text condition. Extensive experiments show that our method can successfully prevent the generation of the ablated concept while preserving closely related concepts in the model.

Nupur Kumari, Bingliang Zhang, Sheng-Yu Wang, Eli Shechtman, Richard Zhang, Jun-Yan Zhu• 2023

Related benchmarks

TaskDatasetResultRank
Text-to-Image GenerationMS-COCO
FID22.63
131
Coarse-grained UnlearningImagenette
Atar100
70
Text-to-Image GenerationMS-COCO (30K)
FID (30K)21.55
62
Text-to-Image GenerationMSCOCO 30K
FID14.08
54
Concept ErasureVan Gogh style
FID17.5
39
Nudity ErasureI2P
Total Count390
38
Concept UnlearningUnlearnDiffAtk
UnlearnDiffAtk0.4296
36
Style ErasurePaul Gauguin
CS29.68
28
Style ErasurePicasso
Contrastive Similarity (CS)27.43
28
Style ErasureMonet
Contrastive Similarity (CS)28.72
28
Showing 10 of 139 rows
...

Other info

Follow for update