Domain and View-point Agnostic Hand Action Recognition
About
Hand action recognition is a special case of action recognition with applications in human-robot interaction, virtual reality or life-logging systems. Building action classifiers able to work for such heterogeneous action domains is very challenging. There are very subtle changes across different actions from a given application but also large variations across domains (e.g. virtual reality vs life-logging). This work introduces a novel skeleton-based hand motion representation model that tackles this problem. The framework we propose is agnostic to the application domain or camera recording view-point. When working on a single domain (intra-domain action classification) our approach performs better or similar to current state-of-the-art methods on well-known hand action recognition benchmarks. And, more importantly, when performing hand action recognition for action domains and camera perspectives which our approach has not been trained for (cross-domain action classification), our proposed framework achieves comparable performance to intra-domain state-of-the-art methods. These experiments show the robustness and generalization capabilities of our framework.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Hand Gesture Recognition | SHREC 14 Gestures 17 | Accuracy93.57 | 42 | |
| Hand Gesture Recognition | SHREC 28 Gestures '17 | Accuracy91.43 | 26 | |
| Hand Gesture Recognition | SHREC 2017 (val) | Accuracy (14G)93.57 | 15 | |
| Action Recognition | F-PHAB 1:1 split | Accuracy95.93 | 12 | |
| Hand Gesture Recognition | FPHA (val) | Accuracy95.93 | 10 | |
| Hand Gesture Recognition | FPHA 1:1 evaluation protocol (val) | Accuracy95.93 | 10 | |
| Action Recognition | F-PHAB 1:3 split | Accuracy92.9 | 7 | |
| Action Recognition | F-PHAB 3:1 split | Accuracy96.76 | 7 | |
| Action Recognition | F-PHAB cross-person | Accuracy88.7 | 7 | |
| Action Recognition | FPHA (test) | Accuracy0.9593 | 6 |