Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

A Recurrent Latent Variable Model for Sequential Data

About

In this paper, we explore the inclusion of latent random variables into the dynamic hidden state of a recurrent neural network (RNN) by combining elements of the variational autoencoder. We argue that through the use of high-level latent random variables, the variational RNN (VRNN)1 can model the kind of variability observed in highly structured sequential data such as natural speech. We empirically evaluate the proposed model against related sequential models on four speech datasets and one handwriting dataset. Our results show the important roles that latent random variables can play in the RNN dynamic hidden state.

Junyoung Chung, Kyle Kastner, Laurent Dinh, Kratarth Goel, Aaron Courville, Yoshua Bengio• 2015

Related benchmarks

TaskDatasetResultRank
Trajectory PredictionETH
minADE201.35
35
Trajectory PredictionETH-UCY ZARA2
minADE (20 steps)1.06
21
Trajectory PredictionT2FPV-ETH (UNIV fold)
minADE@201.27
11
Trajectory PredictionT2FPV-ETH (HOTEL fold)
minADE@201.06
11
Trajectory PredictionEgoTraj TBD
minADE@200.68
11
Trajectory PredictionT2FPV-ETH (ZARA1 fold)
minADE@201.14
11
Showing 6 of 6 rows

Other info

Follow for update