Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning to Generate Reviews and Discovering Sentiment

About

We explore the properties of byte-level recurrent language models. When given sufficient amounts of capacity, training data, and compute time, the representations learned by these models include disentangled features corresponding to high-level concepts. Specifically, we find a single unit which performs sentiment analysis. These representations, learned in an unsupervised manner, achieve state of the art on the binary subset of the Stanford Sentiment Treebank. They are also very data efficient. When using only a handful of labeled examples, our approach matches the performance of strong baselines trained on full datasets. We also demonstrate the sentiment unit has a direct influence on the generative process of the model. Simply fixing its value to be positive or negative generates samples with the corresponding positive or negative sentiment.

Alec Radford, Rafal Jozefowicz, Ilya Sutskever• 2017

Related benchmarks

TaskDatasetResultRank
Subjectivity ClassificationSubj
Accuracy94.7
266
Sentiment AnalysisIMDB (test)
Accuracy92.9
248
Text ClassificationSST-2 (test)
Accuracy91.8
185
Text ClassificationTREC
Accuracy90.4
179
Opinion Polarity DetectionMPQA
Accuracy88.5
154
Sentiment ClassificationIMDB (test)
Error Rate7.12
144
Sentiment ClassificationCR
Accuracy91.4
142
Text ClassificationIMDB
Accuracy92.2
107
Sentiment ClassificationStanford Sentiment Treebank SST-2 (test)
Accuracy91.8
99
Text ClassificationMR (test)
Accuracy86.9
99
Showing 10 of 18 rows

Other info

Follow for update