Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Recursive Non-Autoregressive Graph-to-Graph Transformer for Dependency Parsing with Iterative Refinement

About

We propose the Recursive Non-autoregressive Graph-to-Graph Transformer architecture (RNGTr) for the iterative refinement of arbitrary graphs through the recursive application of a non-autoregressive Graph-to-Graph Transformer and apply it to syntactic dependency parsing. We demonstrate the power and effectiveness of RNGTr on several dependency corpora, using a refinement model pre-trained with BERT. We also introduce Syntactic Transformer (SynTr), a non-recursive parser similar to our refinement model. RNGTr can improve the accuracy of a variety of initial parsers on 13 languages from the Universal Dependencies Treebanks, English and Chinese Penn Treebanks, and the German CoNLL2009 corpus, even improving over the new state-of-the-art results achieved by SynTr, significantly improving the state-of-the-art for all corpora tested.

Alireza Mohammadshahi, James Henderson• 2020

Related benchmarks

TaskDatasetResultRank
Dependency ParsingChinese Treebank (CTB) (test)
UAS92.98
99
Dependency ParsingPenn Treebank (PTB) (test)
LAS95.01
80
Dependency ParsingUD Treebank Arabic (test)
LAS86.31
11
Dependency ParsingUD Treebank Basque (test)
LAS88.2
7
Dependency ParsingUD Treebank Chinese (test)
LAS90.48
7
Dependency ParsingUD Treebank (English) (test)
LAS91.52
7
Dependency ParsingUD Treebank Finnish (test)
LAS91.92
7
Dependency ParsingUD Treebank Hebrew (test)
LAS91.32
7
Dependency ParsingUD Treebank (Hindi) (test)
LAS94.21
7
Dependency ParsingUD Treebank (Italian) (test)
LAS95.16
7
Showing 10 of 16 rows

Other info

Code

Follow for update