Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Generative Pretrained Hierarchical Transformer for Time Series Forecasting

About

Recent efforts have been dedicated to enhancing time series forecasting accuracy by introducing advanced network architectures and self-supervised pretraining strategies. Nevertheless, existing approaches still exhibit two critical drawbacks. Firstly, these methods often rely on a single dataset for training, limiting the model's generalizability due to the restricted scale of the training data. Secondly, the one-step generation schema is widely followed, which necessitates a customized forecasting head and overlooks the temporal dependencies in the output series, and also leads to increased training costs under different horizon length settings. To address these issues, we propose a novel generative pretrained hierarchical transformer architecture for forecasting, named \textbf{GPHT}. There are two aspects of key designs in GPHT. On the one hand, we advocate for constructing a mixed dataset under the channel-independent assumption for pretraining our model, comprising various datasets from diverse data scenarios. This approach significantly expands the scale of training data, allowing our model to uncover commonalities in time series data and facilitating improved transfer to specific datasets. On the other hand, GPHT employs an auto-regressive forecasting approach, effectively modeling temporal dependencies in the output series. Importantly, no customized forecasting head is required, enabling \textit{a single model to forecast at arbitrary horizon settings.} We conduct sufficient experiments on eight datasets with mainstream self-supervised pretraining models and supervised models. The results demonstrated that GPHT surpasses the baseline models across various fine-tuning and zero/few-shot learning settings in the traditional long-term forecasting task. We make our codes publicly available\footnote{https://github.com/icantnamemyself/GPHT}.

Zhiding Liu, Jiqian Yang, Mingyue Cheng, Yucong Luo, Zhi Li• 2024

Related benchmarks

TaskDatasetResultRank
Multivariate ForecastingETTh1
MSE0.363
645
Multivariate Time-series ForecastingETTm1
MSE0.291
433
Multivariate ForecastingETTh2
MSE0.298
341
Multivariate Time-series ForecastingETTm2
MSE0.17
334
Multivariate long-term series forecastingETTh2
MSE0.296
319
Multivariate Time-series ForecastingWeather
MSE0.154
276
Time Series ForecastingTraffic (test)
MSE0.411
192
Multivariate Time-series ForecastingExchange
MAE0.207
165
Multivariate ForecastingTraffic
MSE0.346
110
Time Series ForecastingWeather (test)
MSE0.202
110
Showing 10 of 12 rows

Other info

Code

Follow for update