MoCha:End-to-End Video Character Replacement without Structural Guidance
About
Controllable video character replacement with a user-provided identity remains a challenging problem due to the lack of paired video data. Prior works have predominantly relied on a reconstruction-based paradigm that requires per-frame segmentation masks and explicit structural guidance (e.g., skeleton, depth). This reliance, however, severely limits their generalizability in complex scenarios involving occlusions, character-object interactions, unusual poses, or challenging illumination, often leading to visual artifacts and temporal inconsistencies. In this paper, we propose MoCha, a pioneering framework that bypasses these limitations by requiring only a single arbitrary frame mask. To effectively adapt the multi-modal input condition and enhance facial identity, we introduce a condition-aware RoPE and employ an RL-based post-training stage. Furthermore, to overcome the scarcity of qualified paired-training data, we propose a comprehensive data construction pipeline. Specifically, we design three specialized datasets: a high-fidelity rendered dataset built with Unreal Engine 5 (UE5), an expression-driven dataset synthesized by current portrait animation techniques, and an augmented dataset derived from existing video-mask pairs. Extensive experiments demonstrate that our method substantially outperforms existing state-of-the-art approaches. We will release the code to facilitate further research. Please refer to our project page for more details: orange-3dv-team.github.io/MoCha
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Human-to-Robot Video Generation | Human2Robot | Task Success Rate (SR)69 | 5 | |
| Character Replacement | Synthesized benchmark | SSIM0.746 | 4 | |
| Video Character Replacement | VBench real-world | Subject Consistency92.25 | 4 | |
| Image-Conditioned Video Editing | Ego-Exo4D (test) | Motion Consistency0.207 | 4 | |
| Image-Conditioned Video Editing | Synthetic Videos (val) | PSNR17.163 | 4 |