Unified Training of Universal Time Series Forecasting Transformers
About
Deep learning for time series forecasting has traditionally operated within a one-model-per-dataset framework, limiting its potential to leverage the game-changing impact of large pre-trained models. The concept of universal forecasting, emerging from pre-training on a vast collection of time series datasets, envisions a single Large Time Series Model capable of addressing diverse downstream forecasting tasks. However, constructing such a model poses unique challenges specific to time series data: i) cross-frequency learning, ii) accommodating an arbitrary number of variates for multivariate time series, and iii) addressing the varying distributional properties inherent in large-scale data. To address these challenges, we present novel enhancements to the conventional time series Transformer architecture, resulting in our proposed Masked Encoder-based Universal Time Series Forecasting Transformer (Moirai). Trained on our newly introduced Large-scale Open Time Series Archive (LOTSA) featuring over 27B observations across nine domains, Moirai achieves competitive or superior performance as a zero-shot forecaster when compared to full-shot models. Code, data, and model weights can be found at https://github.com/SalesforceAIResearch/uni2ts.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Time Series Forecasting | ETTh1 | MSE0.1458 | 601 | |
| Time Series Forecasting | ETTh2 | MSE0.379 | 438 | |
| Time Series Forecasting | ETTm2 | MSE0.343 | 382 | |
| Long-term time-series forecasting | ETTh1 | MAE0.424 | 351 | |
| Long-term time-series forecasting | Weather | MSE0.21 | 348 | |
| Multivariate long-term forecasting | ETTh1 | MSE0.4 | 344 | |
| Time Series Forecasting | ETTm1 | MSE0.714 | 334 | |
| Long-term time-series forecasting | ETTh2 | MSE0.341 | 327 | |
| Multivariate long-term series forecasting | ETTh2 | MSE0.341 | 319 | |
| Long-term time-series forecasting | ETTm2 | MSE0.272 | 305 |