Deep Temporal Sigmoid Belief Networks for Sequence Modeling
About
Deep dynamic generative models are developed to learn sequential dependencies in time-series data. The multi-layered model is designed by constructing a hierarchy of temporal sigmoid belief networks (TSBNs), defined as a sequential stack of sigmoid belief networks (SBNs). Each SBN has a contextual hidden state, inherited from the previous SBNs in the sequence, and is used to regulate its hidden bias. Scalable learning and inference algorithms are derived by introducing a recognition model that yields fast sampling from the variational posterior. This recognition model is trained jointly with the generative model, by maximizing its variational lower bound on the log-likelihood. Experimental results on bouncing balls, polyphonic music, motion capture, and text streams show that the proposed approach achieves state-of-the-art predictive performance, and has the capacity to synthesize various sequences.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Polyphonic music modeling | JSB Chorales | Negative Log-Likelihood (nats)7.48 | 14 | |
| Polyphonic music modeling | Nottingham (Nott) | NLL (nats)3.67 | 14 | |
| Polyphonic music modeling | MuseData (Muse) | Negative Log-Likelihood (nats)6.81 | 12 | |
| Polyphonic music modeling | Piano-midi.de | NLL (nats)7.89 | 12 | |
| Polyphonic Music Generation | Nottingham (test) | -- | 11 | |
| Future Frame Prediction | CMU mocap Dataset 1 (test) | Mean MSE80.21 | 7 | |
| Future Frame Prediction | CMU mocap Dataset 2 (test) | Mean MSE34.86 | 5 | |
| Polyphonic Music Generation | JSB (test) | -- | 5 | |
| Polyphonic Music Generation | Piano (test) | -- | 5 | |
| Polyphonic Music Generation | Musedata (test) | -- | 5 |