Deep Semantic Role Labeling with Self-Attention
About
Semantic Role Labeling (SRL) is believed to be a crucial step towards natural language understanding and has been widely studied. Recent years, end-to-end SRL with recurrent neural networks (RNN) has gained increasing attention. However, it remains a major challenge for RNNs to handle structural information and long range dependencies. In this paper, we present a simple and effective architecture for SRL which aims to address these problems. Our model is based on self-attention which can directly capture the relationships between two tokens regardless of their distance. Our single model achieves F$_1=83.4$ on the CoNLL-2005 shared task dataset and F$_1=82.7$ on the CoNLL-2012 shared task dataset, which outperforms the previous state-of-the-art results by $1.8$ and $1.0$ F$_1$ score respectively. Besides, our model is computationally efficient, and the parsing speed is 50K tokens per second on a single Titan X GPU.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Semantic Role Labeling | CoNLL 2012 (test) | F1 Score83.9 | 49 | |
| Semantic Role Labeling | CoNLL 2005 (WSJ) | F1 Score86.1 | 41 | |
| Span-based Semantic Role Labeling | CoNLL 2005 (Out-of-domain (Brown)) | F1 Score74.1 | 41 | |
| Semantic Role Labeling | CoNLL Brown 2005 (test) | F174.8 | 40 | |
| Semantic Role Labeling | CoNLL 2005 (Brown) | F1 Score74.8 | 31 | |
| Semantic Role Labeling | CoNLL WSJ 2005 (test) | Precision85.9 | 29 | |
| Span Semantic Role Labeling | CoNLL-2012 (OntoNotes) v5.0 (test) | F1 Score83.9 | 25 | |
| Semantic Role Labeling | CoNLL 2005 (WSJ (in-domain)) | F1 Score84.8 | 24 | |
| Semantic Role Labeling | CoNLL 2012 (dev) | F184.1 | 23 | |
| Semantic Role Labeling | CoNLL 2005 (dev) | F1 Score0.846 | 22 |