Incorporating Syntactic and Semantic Information in Word Embeddings using Graph Convolutional Networks
About
Word embeddings have been widely adopted across several NLP applications. Most existing word embedding methods utilize sequential context of a word to learn its embedding. While there have been some attempts at utilizing syntactic context of a word, such methods result in an explosion of the vocabulary size. In this paper, we overcome this problem by proposing SynGCN, a flexible Graph Convolution based method for learning word embeddings. SynGCN utilizes the dependency context of a word without increasing the vocabulary size. Word embeddings learned by SynGCN outperform existing methods on various intrinsic and extrinsic tasks and provide an advantage when used with ELMo. We also propose SemGCN, an effective framework for incorporating diverse semantic knowledge for further enhancing learned word representations. We make the source code of both models available to encourage reproducible research.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Chunking | CoNLL 2000 (test) | F1 Score91.23 | 88 | |
| Named Entity Recognition | OntoNotes 4.0 (test) | F1 Score88.75 | 55 | |
| Word Similarity | WS-353 | Spearman Correlation (WS-353)0.732 | 54 | |
| Part-of-Speech Tagging | WSJ (test) | Accuracy96.71 | 51 | |
| Word Similarity | WS-353 REL (test) | Spearman Correlation0.457 | 28 | |
| Word Similarity | SimLex-999 | Spearman Correlation45.5 | 23 | |
| Word Concept Categorization | AP, Battig, ESSLI (test) | AP Score63.4 | 11 | |
| Word Analogy | SemEval 2012 | Spearman Correlation23.4 | 5 |