Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Portrait4D-v2: Pseudo Multi-View Data Creates Better 4D Head Synthesizer

About

In this paper, we propose a novel learning approach for feed-forward one-shot 4D head avatar synthesis. Different from existing methods that often learn from reconstructing monocular videos guided by 3DMM, we employ pseudo multi-view videos to learn a 4D head synthesizer in a data-driven manner, avoiding reliance on inaccurate 3DMM reconstruction that could be detrimental to the synthesis performance. The key idea is to first learn a 3D head synthesizer using synthetic multi-view images to convert monocular real videos into multi-view ones, and then utilize the pseudo multi-view videos to learn a 4D head synthesizer via cross-view self-reenactment. By leveraging a simple vision transformer backbone with motion-aware cross-attentions, our method exhibits superior performance compared to previous methods in terms of reconstruction fidelity, geometry consistency, and motion control accuracy. We hope our method offers novel insights into integrating 3D priors with 2D supervisions for improved 4D head avatar creation.

Yu Deng, Duomin Wang, Baoyuan Wang• 2024

Related benchmarks

TaskDatasetResultRank
3D Head Avatar ReconstructionAva 256
PSNR11.9
13
Video-driven Talking Head Generation (Self-Reenactment)HDTF
FID27.83
12
Cross-ReenactmentVOODOO-XP (test)
MEt3R0.035
10
Self-ReenactmentVOODOO-XP (test)
MEt3R0.03
10
Cross-identity reenactmentVFHQ (test)
CSIM0.6731
8
Cross-identity reenactmentHDTF 55 (test)
CSIM0.8669
8
3D Face ReenactmentCelebVHQ
Sc Score5.823
8
Self-ReenactmentHDTF 55 (test)
PSNR20.86
8
Self-ReenactmentVFHQ (test)
PSNR17.65
8
Talking head video generationTalkinghead1kh
FID37.38
8
Showing 10 of 24 rows

Other info

Follow for update