Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

AnimateAnything: Consistent and Controllable Animation for Video Generation

About

We present a unified controllable video generation approach AnimateAnything that facilitates precise and consistent video manipulation across various conditions, including camera trajectories, text prompts, and user motion annotations. Specifically, we carefully design a multi-scale control feature fusion network to construct a common motion representation for different conditions. It explicitly converts all control information into frame-by-frame optical flows. Then we incorporate the optical flows as motion priors to guide final video generation. In addition, to reduce the flickering issues caused by large-scale motion, we propose a frequency-based stabilization module. It can enhance temporal coherence by ensuring the video's frequency domain consistency. Experiments demonstrate that our method outperforms the state-of-the-art approaches. For more details and videos, please refer to the webpage: https://yu-shaonian.github.io/Animate_Anything/.

Guojun Lei, Chi Wang, Hong Li, Rong Zhang, Yikai Wang, Weiwei Xu• 2024

Related benchmarks

TaskDatasetResultRank
Camera Trajectory ControlReal10K Basic Trajectory
Translational Error (T-Err)0.041
9
Camera Trajectory ControlReal10K Difficult Trajectory
Translation Error (T-Err)0.053
9
Language-driven motion control in Text-to-Video generationSSv2 (val)
FVD90.3
8
Video GenerationWebVid (test)
LPIPS0.135
7
Video GenerationOpenVid (test)
LPIPS0.113
7
Video GenerationWebVid (test)
SubC92.8
5
Showing 6 of 6 rows

Other info

Code

Follow for update