Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Word Representations via Gaussian Embedding

About

Current work in lexical distributed representations maps each word to a point vector in low-dimensional space. Mapping instead to a density provides many interesting advantages, including better capturing uncertainty about a representation and its relationships, expressing asymmetries more naturally than dot product or cosine similarity, and enabling more expressive parameterization of decision boundaries. This paper advocates for density-based distributed embeddings and presents a method for learning representations in the space of Gaussian distributions. We compare performance on various word embedding benchmarks, investigate the ability of these embeddings to model entailment and other asymmetric relationships, and explore novel properties of the representation.

Luke Vilnis, Andrew McCallum• 2014

Related benchmarks

TaskDatasetResultRank
Word SimilarityRG-65
Spearman Correlation0.69
41
Word SimilaritySimLex-999
Spearman Correlation25
31
Word SimilarityMechanical Turk-771
Spearman ρ0.57
8
Word SimilarityRW-STANFORD
Spearman Correlation0.4
6
Word SimilarityWS-353 REL
Spearman Correlation0.61
6
Word SimilarityMC-30
Spearman Correlation0.59
6
Word SimilarityWS-353 ALL
Spearman Correlation0.53
6
Word SimilarityWS-YP-130
Spearman Correlation0.37
6
Word SimilarityMEN 3k (train)
Spearman Correlation0.65
6
Word SimilarityMTurk-287
Spearman Correlation0.61
6
Showing 10 of 12 rows

Other info

Follow for update