Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Continual Learning for non-stationary regression via Memory-Efficient Replay

About

Data streams are rarely static in dynamic environments like Industry 4.0. Instead, they constantly change, making traditional offline models outdated unless they can quickly adjust to the new data. This need can be adequately addressed by continual learning (CL), which allows systems to gradually acquire knowledge without incurring the prohibitive costs of retraining them from scratch. Most research on continual learning focuses on classification problems, while very few studies address regression tasks. We propose the first prototype-based generative replay framework designed for online task-free continual regression. Our approach defines an adaptive output-space discretization model, enabling prototype-based generative replay for continual regression without storing raw data. Evidence obtained from several benchmark datasets shows that our framework reduces forgetting and provides more stable performance than other state-of-the-art solutions.

Pablo Garc\'ia-Santaclara, Bruno Fern\'andez-Castro, RebecaP.D\'iaz-Redondo, Mart\'in Alonso-Gamarra• 2026

Related benchmarks

TaskDatasetResultRank
RegressionCalifornia Housing
MSE0.421
71
Regressiondiamonds
MSE0.066
6
Continual Learning RegressionEurope Wind Farm WF1
Forgetting Ratio0.00e+0
5
Continual Learning RegressionEurope Wind Farm WF2
Forgetting Ratio0.00e+0
5
Continual Learning RegressionEurope Wind Farm WF3
Forgetting Ratio0.021
5
Continual Learning RegressionEurope Wind Farm WF4
Forgetting Ratio0.00e+0
5
Continual Learning RegressionEurope Wind Farm WF6
Forgetting Ratio0.00e+0
5
Continual Learning RegressionEurope Wind Farm WF7
Forgetting Ratio0.00e+0
5
Continual Learning RegressionEurope Wind Farm WF8
Forgetting Ratio0.00e+0
5
Continual Learning RegressionEurope Wind Farm WF9
Forgetting Ratio0.034
5
Showing 10 of 20 rows

Other info

Follow for update