AMR Parsing as Sequence-to-Graph Transduction
About
We propose an attention-based model that treats AMR parsing as sequence-to-graph transduction. Unlike most AMR parsers that rely on pre-trained aligners, external semantic resources, or data augmentation, our proposed parser is aligner-free, and it can be effectively trained with limited amounts of labeled AMR data. Our experimental results outperform all previously reported SMATCH scores, on both AMR 2.0 (76.3% F1 on LDC2017T10) and AMR 1.0 (70.2% F1 on LDC2014T12).
Sheng Zhang, Xutai Ma, Kevin Duh, Benjamin Van Durme• 2019
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| AMR parsing | LDC2017T10 AMR 2.0 (test) | Smatch77 | 168 | |
| AMR parsing | AMR 1.0 (test) | Smatch70.2 | 45 | |
| AMR parsing | AMR 1.0 LDC2014T12 (test) | SMATCH F170.2 | 23 |
Showing 3 of 3 rows