Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Conll

Benchmarks

Task NameDataset NameSOTA ResultTrend
Named Entity RecognitionCoNLL 2003 (test)
F1 Score98.31
539
Grammatical Error CorrectionCoNLL 2014 (test)
F0.5 Score72.2
207
Named Entity RecognitionCoNLL English 2003 (test)
F1 Score94.6
135
Coreference ResolutionCoNLL English 2012 (test)
MUC F1 Score88
114
Named Entity RecognitionCoNLL 03
F1 (Entity)94.6
102
Named Entity RecognitionCoNLL Spanish NER 2002 (test)
F1 Score95.9
98
ChunkingCoNLL 2000 (test)
F1 Score97.3
88
Named Entity RecognitionCoNLL Dutch 2002 (test)
F1 Score95.7
87
Named Entity RecognitionConll 2003
F1 Score96.77
86
Named Entity RecognitionCoNLL German 2003 (test)
F1 Score88.34
78
Relation ExtractionCONLL04
Relation Strict F178.84
52
Semantic Role LabelingCoNLL 2012 (test)
F1 Score88.59
49
Span-based Semantic Role LabelingCoNLL 2005 (Out-of-domain (Brown))
F1 Score85.1
41
Semantic Role LabelingCoNLL 2005 (WSJ)
F1 Score95.5
41
Named Entity RecognitionCoNLL 2003 (dev)
F1 Score97.21
40
Semantic Role LabelingCoNLL Brown 2005 (test)
F193.7
40
Relation ExtractionCoNLL 04
F179.73
39
Grammatical Error CorrectionCoNLL 2014
F0.565.2
39
Joint Entity and Relation ExtractionCONLL04
Entity F193.26
33
Dependency Semantic Role LabelingCoNLL 2009 (test)
F1 Score93.03
32
Semantic Role LabelingCoNLL WSJ English benchmark 2009 (test)
F1 Score92.83
31
Semantic Role LabelingCoNLL 2005 (Brown)
F1 Score92.1
31
Semantic Role LabelingCoNLL WSJ 2005 (test)
Precision87.13
29
Relation ExtractionCoNLL04 (test)
F1 Score75.8
28
Semantic Role LabelingCoNLL English Brown 2009 (test)
F1 Score86.05
28
Showing 25 of 168 rows