Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Spiking Transformer with Spatial-Temporal Attention

About

Spike-based Transformer presents a compelling and energy-efficient alternative to traditional Artificial Neural Network (ANN)-based Transformers, achieving impressive results through sparse binary computations. However, existing spike-based transformers predominantly focus on spatial attention while neglecting crucial temporal dependencies inherent in spike-based processing, leading to suboptimal feature representation and limited performance. To address this limitation, we propose Spiking Transformer with Spatial-Temporal Attention (STAtten), a simple and straightforward architecture that efficiently integrates both spatial and temporal information in the self-attention mechanism. STAtten introduces a block-wise computation strategy that processes information in spatial-temporal chunks, enabling comprehensive feature capture while maintaining the same computational complexity as previous spatial-only approaches. Our method can be seamlessly integrated into existing spike-based transformers without architectural overhaul. Extensive experiments demonstrate that STAtten significantly improves the performance of existing spike-based transformers across both static and neuromorphic datasets, including CIFAR10/100, ImageNet, CIFAR10-DVS, and N-Caltech101. The code is available at https://github.com/Intelligent-Computing-Lab-Yale/STAtten

Donghyun Lee, Yuhang Li, Youngeun Kim, Shiting Xiao, Priyadarshini Panda• 2024

Related benchmarks

TaskDatasetResultRank
Skeleton-based Action RecognitionNTU RGB+D (Cross-View)
Accuracy79.7
213
Skeleton-based Action RecognitionNTU RGB+D 120 Cross-Subject
Top-1 Accuracy60.3
143
Skeleton-based Action RecognitionNTU-RGB+D 120 (Cross-setup)
Accuracy61.7
136
Skeleton-based Action RecognitionNTU RGB+D (Cross-subject)
Accuracy72.8
123
Skeleton-based Action RecognitionNW-UCLA
Accuracy86.9
44
Showing 5 of 5 rows

Other info

Follow for update