Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Neural Texture Extraction and Distribution for Controllable Person Image Synthesis

About

We deal with the controllable person image synthesis task which aims to re-render a human from a reference image with explicit control over body pose and appearance. Observing that person images are highly structured, we propose to generate desired images by extracting and distributing semantic entities of reference images. To achieve this goal, a neural texture extraction and distribution operation based on double attention is described. This operation first extracts semantic neural textures from reference feature maps. Then, it distributes the extracted neural textures according to the spatial distributions learned from target poses. Our model is trained to predict human images in arbitrary poses, which encourages it to extract disentangled and expressive neural textures representing the appearance of different semantic entities. The disentangled representation further enables explicit appearance control. Neural textures of different reference images can be fused to control the appearance of the interested areas. Experimental comparisons show the superiority of the proposed model. Code is available at https://github.com/RenYurui/Neural-Texture-Extraction-Distribution.

Yurui Ren, Xiaoqing Fan, Ge Li, Shan Liu, Thomas H. Li• 2022

Related benchmarks

TaskDatasetResultRank
ReposingDeepFashion In-Domain
FID7.685
10
ReposingWPose (Out-of-Domain)
FID90.542
10
Person Image SynthesisDeepFashion 256 x 176 (test)
FID8.6838
9
Pose TransferDeepFashion reduced (test)
FID8.813
7
Person Image SynthesisDeepFashion 512 x 352 (test)
FID7.7821
3
Showing 5 of 5 rows

Other info

Follow for update