Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Tuning the burn-in phase in training recurrent neural networks improves their performance

About

Training recurrent neural networks (RNNs) with standard backpropagation through time (BPTT) can be challenging, especially in the presence of long input sequences. A practical alternative to reduce computational and memory overhead is to perform BPTT repeatedly over shorter segments of the training data set, corresponding to truncated BPTT. In this paper, we examine the training of RNNs when using such a truncated learning approach for time series tasks. Specifically, we establish theoretical bounds on the accuracy and performance loss when optimizing over subsequences instead of the full data sequence. This reveals that the burn-in phase of the RNN is an important tuning knob in its training, with significant impact on the performance guarantees. We validate our theoretical results through experiments on standard benchmarks from the fields of system identification and time series forecasting. In all experiments, we observe a strong influence of the burn-in phase on the training process, and proper tuning can lead to a reduction of the prediction error on the training and test data of more than 60% in some cases.

Julian D. Schiller, Malte Heinrich, Victor G. Lopez, Matthias A. M\"uller• 2026

Related benchmarks

TaskDatasetResultRank
Time Series ForecastingTraffic (test)
MSE0.276
192
System IdentificationSilver-Box (test)
MSE (x10^2)0.112
16
System IdentificationW-H (test)
MSE (x10^2)0.372
16
System IdentificationSilver-Box (train)
MSE1.2
16
System IdentificationRLC (train)
MSE (x10^2)0.176
16
System IdentificationRLC (test)
MSE (x10^2)0.228
16
System IdentificationW-H (train)
MSE (x10^2)0.116
16
Time Series ForecastingElectr. (train)
MSE0.083
12
Time Series ForecastingElectr. (test)
MSE0.144
12
Time Series ForecastingTraffic (train)
MSE0.205
12
Showing 10 of 11 rows

Other info

Follow for update