Mesh Graphormer
About
We present a graph-convolution-reinforced transformer, named Mesh Graphormer, for 3D human pose and mesh reconstruction from a single image. Recently both transformers and graph convolutional neural networks (GCNNs) have shown promising progress in human mesh reconstruction. Transformer-based approaches are effective in modeling non-local interactions among 3D mesh vertices and body joints, whereas GCNNs are good at exploiting neighborhood vertex interactions based on a pre-specified mesh topology. In this paper, we study how to combine graph convolutions and self-attentions in a transformer to model both local and global interactions. Experimental results show that our proposed method, Mesh Graphormer, significantly outperforms the previous state-of-the-art methods on multiple benchmarks, including Human3.6M, 3DPW, and FreiHAND datasets. Code and pre-trained models are available at https://github.com/microsoft/MeshGraphormer
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| 3D Human Pose Estimation | Human3.6M (test) | -- | 547 | |
| 3D Human Pose Estimation | 3DPW (test) | PA-MPJPE45.6 | 505 | |
| 3D Human Mesh Recovery | 3DPW (test) | PA-MPJPE45.6 | 264 | |
| 3D Human Pose Estimation | Human3.6M | MPJPE51.2 | 160 | |
| 3D Human Pose and Shape Estimation | 3DPW (test) | MPJPE-PA45.6 | 158 | |
| 3D Hand Reconstruction | FreiHAND (test) | F@15mm98.7 | 148 | |
| Human Mesh Recovery | 3DPW | PA-MPJPE45.6 | 123 | |
| 3D Human Mesh Recovery | Human3.6M (test) | PA-MPJPE34.5 | 120 | |
| 3D Human Pose and Shape Estimation | Human3.6M (test) | PA-MPJPE34.5 | 119 | |
| 3D Human Pose Estimation | 3DPW | PA-MPJPE45.6 | 119 |