HairWeaver: Few-Shot Photorealistic Hair Motion Synthesis with Sim-to-Real Guided Video Diffusion
About
We present HairWeaver, a diffusion-based pipeline that animates a single human image with realistic and expressive hair dynamics. While existing methods successfully control body pose, they lack specific control over hair, and as a result, fail to capture the intricate hair motions, resulting in stiff and unrealistic animations. HairWeaver overcomes this limitation using two specialized modules: a Motion-Context-LoRA to integrate motion conditions and a Sim2Real-Domain-LoRA to preserve the subject's photoreal appearance across different data domains. These lightweight components are designed to guide a video diffusion backbone while maintaining its core generative capabilities. By training on a specialized dataset of dynamic human motion generated from a CG simulator, HairWeaver affords fine control over hair motion and ultimately learns to produce highly realistic hair that responds naturally to movement. Comprehensive evaluations demonstrate that our approach sets a new state of the art, producing lifelike human hair animations with dynamic details.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Human Video Animation | Self-collected hair motion CG (test) | Average Vote Percentage49.9 | 6 | |
| Human Video Generation | User Study 20 video subjects (test) | Average Vote Percentage68.17 | 6 | |
| motion-conditioned image-to-video animation | self-collected hair motion (test) | SSIM (Hair)0.9794 | 6 | |
| motion-conditioned image-to-video animation | NeRSemble (test) | Hair SSIM0.967 | 6 |