Broad-Coverage Semantic Parsing as Transduction
About
We unify different broad-coverage semantic parsing tasks under a transduction paradigm, and propose an attention-based neural framework that incrementally builds a meaning representation via a sequence of semantic relations. By leveraging multiple attention mechanisms, the transducer can be effectively trained without relying on a pre-trained aligner. Experiments conducted on three separate broad-coverage semantic parsing tasks -- AMR, SDP and UCCA -- demonstrate that our attention-based neural transducer improves the state of the art on both AMR and UCCA, and is competitive with the state of the art on SDP.
Sheng Zhang, Xutai Ma, Kevin Duh, Benjamin Van Durme• 2019
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| AMR parsing | LDC2017T10 AMR 2.0 (test) | Smatch77 | 168 | |
| AMR parsing | AMR 1.0 (test) | Smatch71.3 | 45 | |
| AMR parsing | AMR 1.0 LDC2014T12 (test) | SMATCH F171.3 | 23 | |
| Semantic Dependency Parsing | SemEval Task 18 2015 (WSJ ID) | Avg (LF1)92.2 | 17 | |
| Semantic Dependency Parsing | SemEval Task 18 Brown corpus OOD 2015 | Average LF187.1 | 17 | |
| Semantic Parsing | UCCA Wiki in-domain (test) | Primary Labeled F-score76.6 | 14 |
Showing 6 of 6 rows