Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

A Structured Self-attentive Sentence Embedding

About

This paper proposes a new model for extracting an interpretable sentence embedding by introducing self-attention. Instead of using a vector, we use a 2-D matrix to represent the embedding, with each row of the matrix attending on a different part of the sentence. We also propose a self-attention mechanism and a special regularization term for the model. As a side effect, the embedding comes with an easy way of visualizing what specific parts of the sentence are encoded into the embedding. We evaluate our model on 3 different tasks: author profiling, sentiment classification, and textual entailment. Results show that our model yields a significant performance gain compared to other sentence embedding methods in all of the 3 tasks.

Zhouhan Lin, Minwei Feng, Cicero Nogueira dos Santos, Mo Yu, Bing Xiang, Bowen Zhou, Yoshua Bengio• 2017

Related benchmarks

TaskDatasetResultRank
Natural Language InferenceSNLI (test)
Accuracy84.2
681
Text ClassificationAG News (test)--
210
Text ClassificationSST-2 (test)
Accuracy86.4
185
Sentiment ClassificationIMDB (test)
Error Rate6.79
144
Sentiment ClassificationMR (test)
Accuracy81.7
142
Text ClassificationYahoo! Answers (test)--
133
Text ClassificationTREC (test)--
113
Text ClassificationIMDB (test)
CA43.3
79
Review Sentiment ClassificationYelp 2014 (test)
Accuracy61.5
41
Text ClassificationDBPedia (test)
Test Error Rate0.007
40
Showing 10 of 18 rows

Other info

Follow for update