Learning Generic Sentence Representations Using Convolutional Neural Networks
About
We propose a new encoder-decoder approach to learn distributed sentence representations that are applicable to multiple purposes. The model is learned by using a convolutional neural network as an encoder to map an input sentence into a continuous vector, and using a long short-term memory recurrent neural network as a decoder. Several tasks are considered, including sentence reconstruction and future sentence prediction. Further, a hierarchical encoder-decoder model is proposed to encode a sentence to predict multiple future sentences. By training our models on a large collection of novels, we obtain a highly generic convolutional sentence encoder that performs well in practice. Experimental results on several benchmark datasets, and across a broad range of applications, demonstrate the superiority of the proposed model over competing methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Subjectivity Classification | Subj | Accuracy93.6 | 266 | |
| Text Classification | TREC | Accuracy92.6 | 179 | |
| Sentiment Classification | CR | Accuracy82 | 142 | |
| Text Classification | MR | Accuracy77.8 | 93 | |
| Text Classification | MPQA | Accuracy89.4 | 25 |