Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Deep Semantic Role Labeling with Self-Attention

About

Semantic Role Labeling (SRL) is believed to be a crucial step towards natural language understanding and has been widely studied. Recent years, end-to-end SRL with recurrent neural networks (RNN) has gained increasing attention. However, it remains a major challenge for RNNs to handle structural information and long range dependencies. In this paper, we present a simple and effective architecture for SRL which aims to address these problems. Our model is based on self-attention which can directly capture the relationships between two tokens regardless of their distance. Our single model achieves F$_1=83.4$ on the CoNLL-2005 shared task dataset and F$_1=82.7$ on the CoNLL-2012 shared task dataset, which outperforms the previous state-of-the-art results by $1.8$ and $1.0$ F$_1$ score respectively. Besides, our model is computationally efficient, and the parsing speed is 50K tokens per second on a single Titan X GPU.

Zhixing Tan, Mingxuan Wang, Jun Xie, Yidong Chen, Xiaodong Shi• 2017

Related benchmarks

TaskDatasetResultRank
Semantic Role LabelingCoNLL 2012 (test)
F1 Score83.9
49
Semantic Role LabelingCoNLL 2005 (WSJ)
F1 Score86.1
41
Span-based Semantic Role LabelingCoNLL 2005 (Out-of-domain (Brown))
F1 Score74.1
41
Semantic Role LabelingCoNLL Brown 2005 (test)
F174.8
40
Semantic Role LabelingCoNLL 2005 (Brown)
F1 Score74.8
31
Semantic Role LabelingCoNLL WSJ 2005 (test)
Precision85.9
29
Span Semantic Role LabelingCoNLL-2012 (OntoNotes) v5.0 (test)
F1 Score83.9
25
Semantic Role LabelingCoNLL 2005 (WSJ (in-domain))
F1 Score84.8
24
Semantic Role LabelingCoNLL 2012 (dev)
F184.1
23
Semantic Role LabelingCoNLL 2005 (dev)
F1 Score0.846
22
Showing 10 of 18 rows

Other info

Code

Follow for update