Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Recurrent Attentive Neural Process for Sequential Data

About

Neural processes (NPs) learn stochastic processes and predict the distribution of target output adaptively conditioned on a context set of observed input-output pairs. Furthermore, Attentive Neural Process (ANP) improved the prediction accuracy of NPs by incorporating attention mechanism among contexts and targets. In a number of real-world applications such as robotics, finance, speech, and biology, it is critical to learn the temporal order and recurrent structure from sequential data. However, the capability of NPs capturing these properties is limited due to its permutation invariance instinct. In this paper, we proposed the Recurrent Attentive Neural Process (RANP), or alternatively, Attentive Neural Process-RecurrentNeural Network(ANP-RNN), in which the ANP is incorporated into a recurrent neural network. The proposed model encapsulates both the inductive biases of recurrent neural networks and also the strength of NPs for modelling uncertainty. We demonstrate that RANP can effectively model sequential data and outperforms NPs and LSTMs remarkably in a 1D regression toy example as well as autonomous-driving applications.

Shenghao Qin, Jiacheng Zhu, Jimmy Qin, Wenshuo Wang, Ding Zhao• 2019

Related benchmarks

TaskDatasetResultRank
Epidemic ForecastingUS National wILI (seasons 2014/15-2019/20)
RMSE0.61
40
Epidemic Forecasting (4-week ahead)wILI HHS regions (average) 2014-15 to 2019-20 seasons
RMSE1.17
10
Epidemic Forecasting (2-week ahead)wILI HHS regions (average) 2014-15 to 2019-20 seasons
RMSE0.87
10
Showing 3 of 3 rows

Other info

Follow for update