Graph Pre-training for AMR Parsing and Generation
About
Abstract meaning representation (AMR) highlights the core semantic information of text in a graph structure. Recently, pre-trained language models (PLMs) have advanced tasks of AMR parsing and AMR-to-text generation, respectively. However, PLMs are typically pre-trained on textual data, thus are sub-optimal for modeling structural knowledge. To this end, we investigate graph self-supervised training to improve the structure awareness of PLMs over AMR graphs. In particular, we introduce two graph auto-encoding strategies for graph-to-graph pre-training and four tasks to integrate text and graph information during pre-training. We further design a unified framework to bridge the gap between pre-training and fine-tuning tasks. Experiments on both AMR parsing and AMR-to-text generation show the superiority of our model. To our knowledge, we are the first to consider pre-training on semantic graphs.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| AMR parsing | LDC2017T10 AMR 2.0 (test) | Smatch85.4 | 168 | |
| AMR parsing | AMR 3.0 (test) | SMATCH84.2 | 45 | |
| AMR parsing | AMR 3.0 LDC2020T02 (test) | Smatch Labeled84.2 | 14 | |
| AMR-to-text generation | AMR 2.0 (test) | BLEU49.8 | 10 | |
| Text-to-UMR Parsing | UMR English sentences v2.0 | AnCast0.817 | 6 | |
| Scene Graph Parsing | Random (test) | Set Match28.45 | 6 | |
| Scene Graph Parsing | Length (test) | Set Match1.22e+3 | 6 | |
| AMR-to-text generation | AMR 3.0 (test) | BLEU49.2 | 5 | |
| Scene Graph Parsing | FACTUAL (test) | Completeness0.31 | 5 |