Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Unlocking the Power of LSTM for Long Term Time Series Forecasting

About

Traditional recurrent neural network architectures, such as long short-term memory neural networks (LSTM), have historically held a prominent role in time series forecasting (TSF) tasks. While the recently introduced sLSTM for Natural Language Processing (NLP) introduces exponential gating and memory mixing that are beneficial for long term sequential learning, its potential short memory issue is a barrier to applying sLSTM directly in TSF. To address this, we propose a simple yet efficient algorithm named P-sLSTM, which is built upon sLSTM by incorporating patching and channel independence. These modifications substantially enhance sLSTM's performance in TSF, achieving state-of-the-art results. Furthermore, we provide theoretical justifications for our design, and conduct extensive comparative and analytical experiments to fully validate the efficiency and superior performance of our model.

Yaxuan Kong, Zepu Wang, Yuqi Nie, Tian Zhou, Stefan Zohren, Yuxuan Liang, Peng Sun, Qingsong Wen• 2024

Related benchmarks

TaskDatasetResultRank
Time Series ForecastingETTh2
MSE0.349
561
Long-term forecastingETTh1
MSE0.438
365
Time Series ForecastingECL
MSE0.171
211
ForecastingTraffic
MSE0.417
68
Time Series ForecastingETTm2
MSE0.269
53
Time Series ForecastingETTm1
MSE0.374
29
Forecastingsolar
MAE0.261
28
ForecastingWeather
MAE0.256
26
GPP predictionFLUXNET
RMSE1.94
19
CH4 predictionFLUXNET CH4
RMSE62.84
14
Showing 10 of 20 rows

Other info

Follow for update