Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

STEER: Simple Temporal Regularization For Neural ODEs

About

Training Neural Ordinary Differential Equations (ODEs) is often computationally expensive. Indeed, computing the forward pass of such models involves solving an ODE which can become arbitrarily complex during training. Recent works have shown that regularizing the dynamics of the ODE can partially alleviate this. In this paper we propose a new regularization technique: randomly sampling the end time of the ODE during training. The proposed regularization is simple to implement, has negligible overhead and is effective across a wide variety of tasks. Further, the technique is orthogonal to several other methods proposed to regularize the dynamics of ODEs and as such can be used in conjunction with them. We show through experiments on normalizing flows, time series models and image recognition that the proposed regularization can significantly decrease training time and even improve performance over baseline models.

Arnab Ghosh, Harkirat Singh Behl, Emilien Dupont, Philip H. S. Torr, Vinay Namboodiri• 2020

Related benchmarks

TaskDatasetResultRank
Density EstimationCIFAR-10 (test)
Bits/dim3.397
134
Density EstimationImageNet 32x32 (test)
Bits per Sub-pixel2.35
66
Density EstimationMNIST (test)
NLL (bits/dim)1.024
56
Time Series ReconstructionMuJoCo (test)
MSE8.403
51
Unconditional Image GenerationCIFAR10
BPD3.397
33
Unconditional Image GenerationImageNet-32
BPD3.35
31
Density EstimationFashion (test)
NLL (bits/dim)2.803
27
Mortality PredictionPhysionet (test)
AUC0.777
14
Activity ClassificationActivity (test)
Accuracy75.6
12
Time Series ReconstructionPhysionet (test)
MSE4.833
7
Showing 10 of 11 rows

Other info

Follow for update