Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Simplify the Usage of Lexicon in Chinese NER

About

Recently, many works have tried to augment the performance of Chinese named entity recognition (NER) using word lexicons. As a representative, Lattice-LSTM (Zhang and Yang, 2018) has achieved new benchmark results on several public Chinese NER datasets. However, Lattice-LSTM has a complex model architecture. This limits its application in many industrial areas where real-time NER responses are needed. In this work, we propose a simple but effective method for incorporating the word lexicon into the character representations. This method avoids designing a complicated sequence modeling architecture, and for any neural NER model, it requires only subtle adjustment of the character representation layer to introduce the lexicon information. Experimental studies on four benchmark Chinese NER datasets show that our method achieves an inference speed up to 6.15 times faster than those of state-ofthe-art methods, along with a better performance. The experimental results also show that the proposed method can be easily incorporated with pre-trained models like BERT.

Ruotian Ma, Minlong Peng, Qi Zhang, Xuanjing Huang• 2019

Related benchmarks

TaskDatasetResultRank
Named Entity RecognitionMSRA (test)
F1 Score95.42
63
Named Entity RecognitionOntoNotes 4.0 (test)
F1 Score82.81
55
Named Entity RecognitionRESUME
F1 Score96.11
52
Named Entity RecognitionWeibo (test)
Overall Score70.5
50
Named Entity RecognitionMSRA
F1 Score94.06
29
Named Entity RecognitionResume (test)
F1 Score96.11
28
Named Entity RecognitionWeiboNER
F1 Score70.5
27
Named Entity RecognitionChinese OntoNotes 4.0 (test)
F1 Score82.81
19
Named Entity RecognitionOntoNotes 4.0
F1 Score76.16
18
Named Entity RecognitionChinese MSRA (test)
F1 Score95.42
18
Showing 10 of 18 rows

Other info

Code

Follow for update