TinierHAR: Towards Ultra-Lightweight Deep Learning Models for Efficient Human Activity Recognition on Edge Devices
About
Human Activity Recognition (HAR) on resource-constrained wearable devices demands inference models that harmonize accuracy with computational efficiency. This paper introduces TinierHAR, an ultra-lightweight deep learning architecture that synergizes residual depthwise separable convolutions, gated recurrent units (GRUs), and temporal aggregation to achieve SOTA efficiency without compromising performance. Evaluated across 14 public HAR datasets, TinierHAR reduces Parameters by 2.7x (vs. TinyHAR) and 43.3x (vs. DeepConvLSTM), and MACs by 6.4x and 58.6x, respectively, while maintaining the averaged F1-scores. Beyond quantitative gains, this work provides the first systematic ablation study dissecting the contributions of spatial-temporal components across proposed TinierHAR, prior SOTA TinyHAR, and the classical DeepConvLSTM, offering actionable insights for designing efficient HAR systems. We finally discussed the findings and suggested principled design guidelines for future efficient HAR. To catalyze edge-HAR research, we open-source all materials in this work for future benchmarking\footnote{https://github.com/zhaxidele/TinierHAR}
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Human Activity Recognition | PAMAP2 | -- | 26 | |
| Human Activity Recognition | Opportunity | Macro F187.09 | 23 | |
| Human Activity Recognition | WISDM | Macro F183.06 | 23 | |
| Human Activity Recognition | UCI-HAR | -- | 15 | |
| Human Activity Recognition | Daphnet | Macro F1 Score89.84 | 9 | |
| Human Activity Recognition | SKODA | Macro F196.99 | 9 | |
| Human Activity Recognition | MotionSense | Macro-F191.99 | 9 | |
| Human Activity Recognition | UniMiB | Macro F179.67 | 9 | |
| Human Activity Recognition | UCI-HAR | Memory Footprint (KB)34.1 | 4 | |
| Human Activity Recognition | MotionSense | Memory Footprint (KB)33.6 | 4 |