Wan-S2V: Audio-Driven Cinematic Video Generation
About
Current state-of-the-art (SOTA) methods for audio-driven character animation demonstrate promising performance for scenarios primarily involving speech and singing. However, they often fall short in more complex film and television productions, which demand sophisticated elements such as nuanced character interactions, realistic body movements, and dynamic camera work. To address this long-standing challenge of achieving film-level character animation, we propose an audio-driven model, which we refere to as Wan-S2V, built upon Wan. Our model achieves significantly enhanced expressiveness and fidelity in cinematic contexts compared to existing approaches. We conducted extensive experiments, benchmarking our method against cutting-edge models such as Hunyuan-Avatar and Omnihuman. The experimental results consistently demonstrate that our approach significantly outperforms these existing solutions. Additionally, we explore the versatility of our method through its applications in long-form video generation and precise video lip-sync editing.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Talking Head Generation | HDTF (test) | FID23.85 | 49 | |
| Talking Avatar Generation | CelebV-HQ (clips) | FID38.21 | 10 | |
| Talking Avatar Generation | long-form videos (test) | FID88.73 | 10 | |
| Talking head video generation | Action Bench (test) | Sync-C6.473 | 9 | |
| Audio-driven video generation | Custom evaluation dataset | Sync-C4.05 | 9 | |
| Audio-driven Avatar Generation | GenBench ShortVideo (user study) | Naturalness84.3 | 7 | |
| Audio-driven Avatar Generation | GenBench-ShortVideo (test) | ASE3.36 | 7 | |
| Talking Head Generation | Foundation capability evaluation set | IQA4.49 | 7 | |
| Audio-driven video generation | EMTD (test) | FID15.66 | 6 | |
| Joint audio-video generation | Custom human-centric audio-video real and anime images (test) | PQ8.15 | 6 |