Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Walking Further: Semantic-aware Multimodal Gait Recognition Under Long-Range Conditions

About

Gait recognition is an emerging biometric technology that enables non-intrusive and hard-to-spoof human identification. However, most existing methods are confined to short-range, unimodal settings and fail to generalize to long-range and cross-distance scenarios under real-world conditions. To address this gap, we present \textbf{LRGait}, the first LiDAR-Camera multimodal benchmark designed for robust long-range gait recognition across diverse outdoor distances and environments. We further propose \textbf{EMGaitNet}, an end-to-end framework tailored for long-range multimodal gait recognition. To bridge the modality gap between RGB images and point clouds, we introduce a semantic-guided fusion pipeline. A CLIP-based Semantic Mining (SeMi) module first extracts human body-part-aware semantic cues, which are then employed to align 2D and 3D features via a Semantic-Guided Alignment (SGA) module within a unified embedding space. A Symmetric Cross-Attention Fusion (SCAF) module hierarchically integrates visual contours and 3D geometric features, and a Spatio-Temporal (ST) module captures global gait dynamics. Extensive experiments on various gait datasets validate the effectiveness of our method.

Zhiyang Lu, Wen Jiang, Tianren Wu, Zhichao Wang, Changwang Zhang, Siqi Shen, Ming Cheng• 2026

Related benchmarks

TaskDatasetResultRank
Gait RecognitionSUSTech1K (test)
Rank-1 Accuracy (Clothing)81.7
32
Gait RecognitionFreeGait
Rank-1 Score85.2
7
Gait RecognitionLRGait
Rank-1 Accuracy (D-20)88.5
7
Showing 3 of 3 rows

Other info

Follow for update