Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

PENGUIN: Enhancing Transformer with Periodic-Nested Group Attention for Long-term Time Series Forecasting

About

Despite advances in the Transformer architecture, their effectiveness for long-term time series forecasting (LTSF) remains controversial. In this paper, we investigate the potential of integrating explicit periodicity modeling into the self-attention mechanism to enhance the performance of Transformer-based architectures for LTSF. Specifically, we propose PENGUIN, a simple yet effective periodic-nested group attention mechanism. Our approach introduces a periodic-aware relative attention bias to directly capture periodic structures and a grouped multi-query attention mechanism to handle multiple coexisting periodicities (e.g., daily and weekly cycles) within time series data. Extensive experiments across diverse benchmarks demonstrate that PENGUIN consistently outperforms both MLP-based and Transformer-based models. Code is available at https://github.com/ysygMhdxw/AISTATS2026_PENGUIN.

Tian Sun, Yuqi Chen, Weiwei Sun• 2025

Related benchmarks

TaskDatasetResultRank
Multivariate long-term forecastingETTh1
MSE0.426
394
Multivariate long-term series forecastingETTh2
MSE0.378
367
Multivariate long-term series forecastingWeather
MSE0.228
359
Multivariate long-term series forecastingETTm1
MSE0.378
305
Multivariate long-term series forecastingWeather (test)
MSE0.15
270
Multivariate long-term forecastingElectricity
MSE0.165
236
Multivariate long-term series forecastingETTm2
MSE0.286
223
Multivariate long-term series forecastingTraffic (test)
MSE0.357
220
Multivariate long-term series forecastingElectricity (test)
MSE0.137
170
Multivariate long-term forecastingTraffic
MSE0.44
165
Showing 10 of 27 rows

Other info

Follow for update