Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Generative Pretrained Hierarchical Transformer for Time Series Forecasting

About

Recent efforts have been dedicated to enhancing time series forecasting accuracy by introducing advanced network architectures and self-supervised pretraining strategies. Nevertheless, existing approaches still exhibit two critical drawbacks. Firstly, these methods often rely on a single dataset for training, limiting the model's generalizability due to the restricted scale of the training data. Secondly, the one-step generation schema is widely followed, which necessitates a customized forecasting head and overlooks the temporal dependencies in the output series, and also leads to increased training costs under different horizon length settings. To address these issues, we propose a novel generative pretrained hierarchical transformer architecture for forecasting, named \textbf{GPHT}. There are two aspects of key designs in GPHT. On the one hand, we advocate for constructing a mixed dataset under the channel-independent assumption for pretraining our model, comprising various datasets from diverse data scenarios. This approach significantly expands the scale of training data, allowing our model to uncover commonalities in time series data and facilitating improved transfer to specific datasets. On the other hand, GPHT employs an auto-regressive forecasting approach, effectively modeling temporal dependencies in the output series. Importantly, no customized forecasting head is required, enabling \textit{a single model to forecast at arbitrary horizon settings.} We conduct sufficient experiments on eight datasets with mainstream self-supervised pretraining models and supervised models. The results demonstrated that GPHT surpasses the baseline models across various fine-tuning and zero/few-shot learning settings in the traditional long-term forecasting task. We make our codes publicly available\footnote{https://github.com/icantnamemyself/GPHT}.

Zhiding Liu, Jiqian Yang, Mingyue Cheng, Yucong Luo, Zhi Li• 2024

Related benchmarks

TaskDatasetResultRank
Multivariate ForecastingETTh1
MSE0.363
686
Multivariate Time-series ForecastingETTm1
MSE0.291
466
Multivariate Time-series ForecastingETTm2
MSE0.17
389
Multivariate long-term series forecastingETTh2
MSE0.296
367
Multivariate ForecastingETTh2
MSE0.298
350
Multivariate Time-series ForecastingWeather
MSE0.154
340
Time Series ForecastingTraffic (test)
MSE0.411
251
Time Series ForecastingWeather (test)
MSE0.202
200
Multivariate Time-series ForecastingExchange
MAE0.207
181
Multivariate ForecastingTraffic
MSE0.346
112
Showing 10 of 12 rows

Other info

Code

Follow for update