Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

ManiGAN: Text-Guided Image Manipulation

About

The goal of our paper is to semantically edit parts of an image matching a given text that describes desired attributes (e.g., texture, colour, and background), while preserving other contents that are irrelevant to the text. To achieve this, we propose a novel generative adversarial network (ManiGAN), which contains two key components: text-image affine combination module (ACM) and detail correction module (DCM). The ACM selects image regions relevant to the given text and then correlates the regions with corresponding semantic words for effective manipulation. Meanwhile, it encodes original image features to help reconstruct text-irrelevant contents. The DCM rectifies mismatched attributes and completes missing contents of the synthetic image. Finally, we suggest a new metric for evaluating image manipulation results, in terms of both the generation of new attributes and the reconstruction of text-irrelevant contents. Extensive experiments on the CUB and COCO datasets demonstrate the superior performance of the proposed method. Code is available at https://github.com/mrlibw/ManiGAN.

Bowen Li, Xiaojuan Qi, Thomas Lukasiewicz, Philip H. S. Torr• 2019

Related benchmarks

TaskDatasetResultRank
Affective Image FilterAIF
SSIM50.72
11
Semantic Image TranslationImageNet (test)
LPIPS21.7
6
Affective Image FilteringUser Study (test)
EPS (%)7.63
6
Text-Guided Image ManipulationMulti-modal CelebA-HQ
FID117.9
5
Text-Guided Image ManipulationMulti-Modal CelebA-HQ Non-CelebA
FID143.4
5
Text-Guided Image ManipulationCUB (test)
CLIP Score21.3
3
Text-Guided Image ManipulationOxford (test)
CLIP Score21.59
3
Text-Guided Image ManipulationMulti-Modal CelebA-HQ Open-Text
FID141.5
3
Text-Guided Image ManipulationCOCO (test)
CLIP Score11.91
3
Text-driven Image EditingCOCO (random edits)
IS14.96
2
Showing 10 of 10 rows

Other info

Follow for update