Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ChatTime: A Unified Multimodal Time Series Foundation Model Bridging Numerical and Textual Data

About

Human experts typically integrate numerical and textual multimodal information to analyze time series. However, most traditional deep learning predictors rely solely on unimodal numerical data, using a fixed-length window for training and prediction on a single dataset, and cannot adapt to different scenarios. The powered pre-trained large language model has introduced new opportunities for time series analysis. Yet, existing methods are either inefficient in training, incapable of handling textual information, or lack zero-shot forecasting capability. In this paper, we innovatively model time series as a foreign language and construct ChatTime, a unified framework for time series and text processing. As an out-of-the-box multimodal time series foundation model, ChatTime provides zero-shot forecasting capability and supports bimodal input/output for both time series and text. We design a series of experiments to verify the superior performance of ChatTime across multiple tasks and scenarios, and create four multimodal datasets to address data gaps. The experimental results demonstrate the potential and utility of ChatTime.

Chengsen Wang, Qi Qi, Jingyu Wang, Haifeng Sun, Zirui Zhuang, Jinming Wu, Lei Zhang, Jianxin Liao• 2024

Related benchmarks

TaskDatasetResultRank
Time Series ReasoningTSUR Reasoning (test)
Inductive Accuracy49.18
19
Time Series ForecastingGIFT-Eval Short-term
nMASE0.983
9
Time Series ForecastingTMQA OPE
RMSE1.19e+4
8
Time Series ImputationTMQA OPE
RMSE2.79e+3
8
Anomaly DetectionTMQA OPE
Accuracy48.9
8
Multiple-choice Question AnsweringTMQA
Accuracy54.6
8
Time-series classificationTMQA OPE
Acc43.9
8
True/False Question AnsweringTMQA
Accuracy0.619
8
Multiple-choice Question AnsweringCTQA
Accuracy36.6
8
Time Series ForecastingGIFT-Eval Med-term
nMASE1.439
6
Showing 10 of 11 rows

Other info

Follow for update