Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Generating Virtual On-body Accelerometer Data from Virtual Textual Descriptions for Human Activity Recognition

About

The development of robust, generalized models in human activity recognition (HAR) has been hindered by the scarcity of large-scale, labeled data sets. Recent work has shown that virtual IMU data extracted from videos using computer vision techniques can lead to substantial performance improvements when training HAR models combined with small portions of real IMU data. Inspired by recent advances in motion synthesis from textual descriptions and connecting Large Language Models (LLMs) to various AI models, we introduce an automated pipeline that first uses ChatGPT to generate diverse textual descriptions of activities. These textual descriptions are then used to generate 3D human motion sequences via a motion synthesis model, T2M-GPT, and later converted to streams of virtual IMU data. We benchmarked our approach on three HAR datasets (RealWorld, PAMAP2, and USC-HAD) and demonstrate that the use of virtual IMU training data generated using our new approach leads to significantly improved HAR model performance compared to only using real IMU data. Our approach contributes to the growing field of cross-modality transfer methods and illustrate how HAR models can be improved through the generation of virtual training data that do not require any manual effort.

Zikang Leng, Hyeokhyen Kwon, Thomas Pl\"otz• 2023

Related benchmarks

TaskDatasetResultRank
Activity RecognitionPAMAP2
Accuracy8.9
22
Activity RecognitionmHealth
F1 Score2.5
17
Activity Recognitionw-HAR
Accuracy67.2
9
Activity RecognitionUT-Complex
Acc11.6
9
Activity RecognitionWISDM
Acc8.3
9
Activity RecognitionDSADS
Acc7.5
9
Activity RecognitionUTD-MHAD
Accuracy3.7
9
Activity RecognitionRealWorld
Accuracy16.9
9
Activity RecognitionTNDA-HAR
Accuracy14.3
9
Activity RecognitionWharf
Accuracy2.7
9
Showing 10 of 18 rows

Other info

Follow for update