Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling

About

For most deep learning practitioners, sequence modeling is synonymous with recurrent networks. Yet recent results indicate that convolutional architectures can outperform recurrent networks on tasks such as audio synthesis and machine translation. Given a new sequence modeling task or dataset, which architecture should one use? We conduct a systematic evaluation of generic convolutional and recurrent architectures for sequence modeling. The models are evaluated across a broad range of standard tasks that are commonly used to benchmark recurrent networks. Our results indicate that a simple convolutional architecture outperforms canonical recurrent networks such as LSTMs across a diverse range of tasks and datasets, while demonstrating longer effective memory. We conclude that the common association between sequence modeling and recurrent networks should be reconsidered, and convolutional networks should be regarded as a natural starting point for sequence modeling tasks. To assist related work, we have made code available at http://github.com/locuslab/TCN .

Shaojie Bai, J. Zico Kolter, Vladlen Koltun• 2018

Related benchmarks

TaskDatasetResultRank
Multivariate ForecastingETTh1
MSE0.511
645
Language ModelingWikiText-103 (test)
Perplexity45.2
524
Multivariate Time-series ForecastingETTm1
MSE0.282
433
Language ModelingPenn Treebank (test)
Perplexity78.93
411
Multivariate long-term series forecastingETTh2
MSE0.444
319
Traffic speed forecastingMETR-LA (test)
MAE2.74
195
Long-range sequence modelingLong Range Arena (LRA)
Text Accuracy60.54
164
Character-level Language ModelingPenn Treebank (test)
BPC1.31
113
Language ModelingLAMBADA
Perplexity1.28e+3
99
Traffic speed forecastingPEMS-BAY (test)
MAE1.45
98
Showing 10 of 124 rows
...

Other info

Code

Follow for update