Simple and Accurate Dependency Parsing Using Bidirectional LSTM Feature Representations
About
We present a simple and effective scheme for dependency parsing which is based on bidirectional-LSTMs (BiLSTMs). Each sentence token is associated with a BiLSTM vector representing the token in its sentential context, and feature vectors are constructed by concatenating a few BiLSTM vectors. The BiLSTM is trained jointly with the parser objective, resulting in very effective feature extractors for parsing. We demonstrate the effectiveness of the approach by applying it to a greedy transition-based parser as well as to a globally optimized graph-based parser. The resulting parsers have very simple architectures, and match or surpass the state-of-the-art accuracies on English and Chinese.
Eliyahu Kiperwasser, Yoav Goldberg• 2016
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Dependency Parsing | Chinese Treebank (CTB) (test) | UAS87.6 | 99 | |
| Dependency Parsing | English PTB Stanford Dependencies (test) | UAS93.9 | 76 | |
| Dependency Parsing | Penn Treebank (PTB) Section 23 v2.2 (test) | UAS93.9 | 17 | |
| POS Tagging | Penn Treebank (PTB) Section 23 v2.2 (test) | POS Accuracy97.3 | 15 | |
| Dependency Parsing | WSJ section 23 (test) | UAS93.9 | 10 | |
| Parsing | English PTB-SD 3.3.0 (test) | UAS93.9 | 7 | |
| Dependency Parsing | English PTB Yamada and Matsumoto head rules (test) | -- | 6 | |
| Parsing | Chinese PTB 5.1 (test) | UAS87.6 | 4 |
Showing 8 of 8 rows