AWR: Adaptive Weighting Regression for 3D Hand Pose Estimation
About
In this paper, we propose an adaptive weighting regression (AWR) method to leverage the advantages of both detection-based and regression-based methods. Hand joint coordinates are estimated as discrete integration of all pixels in dense representation, guided by adaptive weight maps. This learnable aggregation process introduces both dense and joint supervision that allows end-to-end training and brings adaptability to weight maps, making the network more accurate and robust. Comprehensive exploration experiments are conducted to validate the effectiveness and generality of AWR under various experimental settings, especially its usefulness for different types of dense representation and input modality. Our method outperforms other state-of-the-art methods on four publicly available datasets, including NYU, ICVL, MSRA and HANDS 2017 dataset.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| 3D Hand Pose Estimation | ICVL (test) | Mean Error (mm)5.98 | 91 | |
| 3D Hand Pose Estimation | HANDS 2017 (test) | SEEN Error (mm)5.21 | 17 | |
| 3D Hand Pose Estimation | NYU Hand Pose Dataset (test) | Mean Joint 3D Error (mm)7.37 | 11 | |
| 3D Hand Pose Estimation | Hands19 Task1 (Extrapolation) | Mean Joint Error (mm)13.76 | 9 |