Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

MT-VAE: Learning Motion Transformations to Generate Multimodal Human Dynamics

About

Long-term human motion can be represented as a series of motion modes---motion sequences that capture short-term temporal dynamics---with transitions between them. We leverage this structure and present a novel Motion Transformation Variational Auto-Encoders (MT-VAE) for learning motion sequence generation. Our model jointly learns a feature embedding for motion modes (that the motion sequence can be reconstructed from) and a feature transformation that represents the transition of one motion mode to the next motion mode. Our model is able to generate multiple diverse and plausible motion sequences in the future from the same input. We apply our approach to both facial and full body motion, and demonstrate applications like analogy-based motion transfer and video synthesis.

Xinchen Yan, Akash Rastogi, Ruben Villegas, Kalyan Sunkavalli, Eli Shechtman, Sunil Hadap, Ersin Yumer, Honglak Lee• 2018

Related benchmarks

TaskDatasetResultRank
Human Motion PredictionHuman3.6M (test)--
85
Human Motion PredictionHumanEva-I (test)
ADE0.345
48
3D Human Pose PredictionHuman3.6M Setting-A
ADE457
13
3D Human Pose PredictionHumanEva I
ADE345
12
Diverse Human Motion PredictionHuman3.6M 30
APD0.403
11
Multimodal motion generationAff-Wild facial expression coefficients (test)
S-MSE15.4
6
Multimodal motion generationHuman3.6M 2D joints (test)
S-MSE2.84
6
Multimodal motion generationAff-Wild facial expression coefficients (train)
S-MSE10.1
6
Multimodal motion generationHuman3.6M 2D joints (train)
S-MSE2.26
6
Trajectory ForecastingSynthetic trajectory data Imbalanced
ADE0.332
5
Showing 10 of 13 rows

Other info

Code

Follow for update