Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Language2Pose: Natural Language Grounded Pose Forecasting

About

Generating animations from natural language sentences finds its applications in a a number of domains such as movie script visualization, virtual human animation and, robot motion planning. These sentences can describe different kinds of actions, speeds and direction of these actions, and possibly a target destination. The core modeling challenge in this language-to-pose application is how to map linguistic concepts to motion animations. In this paper, we address this multimodal problem by introducing a neural architecture called Joint Language to Pose (or JL2P), which learns a joint embedding of language and pose. This joint embedding space is learned end-to-end using a curriculum learning approach which emphasizes shorter and easier sequences first before moving to longer and harder ones. We evaluate our proposed model on a publicly available corpus of 3D pose data and human-annotated sentences. Both objective metrics and human judgment evaluation confirm that our proposed approach is able to generate more accurate animations and are deemed visually more representative by humans than other data driven approaches.

Chaitanya Ahuja, Louis-Philippe Morency• 2019

Related benchmarks

TaskDatasetResultRank
Text-to-motion generationHumanML3D (test)
FID11.02
331
text-to-motion mappingKIT-ML (test)
R Precision (Top 3)0.483
275
text-to-motion mappingHumanML3D (test)
FID11.02
243
Text-to-motion generationKIT-ML (test)
FID6.545
115
Text-to-Motion SynthesisHumanML3D--
43
Text-to-motion generationKIT (test)
R-Precision Top-122.1
14
Text-to-Motion SynthesisKIT-ML
R Precision Top 122.1
10
Co-speech gesture synthesisTED (test)
FGD22.083
9
Co-speech gesture generationTED Gesture & TED Expressive User Study (test)
Naturalness1.22
7
Co-speech gesture generationTED Gesture
FGD22.083
7
Showing 10 of 19 rows

Other info

Follow for update