Latent ODEs for Irregularly-Sampled Time Series
About
Time series with non-uniform intervals occur in many applications, and are difficult to model using standard recurrent neural networks (RNNs). We generalize RNNs to have continuous-time hidden dynamics defined by ordinary differential equations (ODEs), a model we call ODE-RNNs. Furthermore, we use ODE-RNNs to replace the recognition network of the recently-proposed Latent ODE model. Both ODE-RNNs and Latent ODEs can naturally handle arbitrary time gaps between observations, and can explicitly model the probability of observation times using Poisson processes. We show experimentally that these ODE-based models outperform their RNN-based counterparts on irregularly-sampled data.
Yulia Rubanova, Ricky T. Q. Chen, David Duvenaud• 2019
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Time Series Reconstruction | MuJoCo (test) | MSE0.285 | 51 | |
| Event Prediction | StackOverflow | RMSE0.952 | 42 | |
| Classification | Activity | Accuracy78.5 | 34 | |
| Classification | PhysioNet | AUC Score0.781 | 23 | |
| Mortality Prediction | P-Mortality P12 (test) | AUPRC50.7 | 19 | |
| Per time-step regression | Walker2D | Squared Error1.051 | 19 | |
| Sequence Classification | Bit-stream XOR Event-based (irregular) encoding (test) | Accuracy75.53 | 18 | |
| Sequence Classification | Bit-stream XOR Equidistant encoding (test) | Accuracy60.28 | 18 | |
| Event Prediction | MIMIC | Accuracy82.7 | 15 | |
| Binary sequence classification | Synthetic Equidistant encoding | Accuracy50.47 | 13 |
Showing 10 of 34 rows