LISA: Learning Implicit Shape and Appearance of Hands
About
This paper proposes a do-it-all neural model of human hands, named LISA. The model can capture accurate hand shape and appearance, generalize to arbitrary hand subjects, provide dense surface correspondences, be reconstructed from images in the wild and easily animated. We train LISA by minimizing the shape and appearance losses on a large set of multi-view RGB image sequences annotated with coarse 3D poses of the hand skeleton. For a 3D point in the hand local coordinate, our model predicts the color and the signed distance with respect to each hand bone independently, and then combines the per-bone predictions using predicted skinning weights. The shape, color and pose representations are disentangled by design, allowing to estimate or animate only selected parameters. We experimentally demonstrate that LISA can accurately reconstruct a dynamic hand from monocular or multi-view sequences, achieving a noticeably higher quality of reconstructed hand shapes compared to baseline approaches. Project page: https://www.iri.upc.edu/people/ecorona/lisa/.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| 3D Hand Shape and Color Reconstruction | DeepHandMesh (test) | V2V Distance3.53 | 17 | |
| Color reconstruction | InterHand2.6M | PSNR28.4 | 15 | |
| Shape reconstruction from point clouds | 3DH 62 (test) | V2V Distance (mm)0.63 | 14 | |
| Shape reconstruction from point clouds | MANO 53 (test) | V2V Error (mm)0.36 | 14 | |
| 3D Hand Reconstruction | DHM (test) | P2S Error (mm)3.38 | 11 |