Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Incorporating Syntactic and Semantic Information in Word Embeddings using Graph Convolutional Networks

About

Word embeddings have been widely adopted across several NLP applications. Most existing word embedding methods utilize sequential context of a word to learn its embedding. While there have been some attempts at utilizing syntactic context of a word, such methods result in an explosion of the vocabulary size. In this paper, we overcome this problem by proposing SynGCN, a flexible Graph Convolution based method for learning word embeddings. SynGCN utilizes the dependency context of a word without increasing the vocabulary size. Word embeddings learned by SynGCN outperform existing methods on various intrinsic and extrinsic tasks and provide an advantage when used with ELMo. We also propose SemGCN, an effective framework for incorporating diverse semantic knowledge for further enhancing learned word representations. We make the source code of both models available to encourage reproducible research.

Shikhar Vashishth, Manik Bhandari, Prateek Yadav, Piyush Rai, Chiranjib Bhattacharyya, Partha Talukdar• 2018

Related benchmarks

TaskDatasetResultRank
ChunkingCoNLL 2000 (test)
F1 Score91.23
88
Named Entity RecognitionOntoNotes 4.0 (test)
F1 Score88.75
55
Word SimilarityWS-353
Spearman Correlation (WS-353)0.732
54
Part-of-Speech TaggingWSJ (test)
Accuracy96.71
51
Word SimilarityWS-353 REL (test)
Spearman Correlation0.457
28
Word SimilaritySimLex-999
Spearman Correlation45.5
23
Word Concept CategorizationAP, Battig, ESSLI (test)
AP Score63.4
11
Word AnalogySemEval 2012
Spearman Correlation23.4
5
Showing 8 of 8 rows

Other info

Follow for update