Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Dynamic Meta-Embeddings for Improved Sentence Representations

About

While one of the first steps in many NLP systems is selecting what pre-trained word embeddings to use, we argue that such a step is better left for neural networks to figure out by themselves. To that end, we introduce dynamic meta-embeddings, a simple yet effective method for the supervised learning of embedding ensembles, which leads to state-of-the-art performance within the same model class on a variety of tasks. We subsequently show how the technique can be used to shed new light on the usage of word embeddings in NLP systems.

Douwe Kiela, Changhan Wang, Kyunghyun Cho• 2018

Related benchmarks

TaskDatasetResultRank
Natural Language InferenceSNLI (test)
Accuracy86.7
681
Image RetrievalFlickr30K
R@136.5
144
Natural Language InferenceMultiNLI Mismatched
Accuracy74.9
60
Text ClassificationSST binary
Accuracy89.8
29
Caption RetrievalFlickr30K
R@149.7
23
Showing 5 of 5 rows

Other info

Follow for update