Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Exploration and Exploitation: Two Ways to Improve Chinese Spelling Correction Models

About

A sequence-to-sequence learning with neural networks has empirically proven to be an effective framework for Chinese Spelling Correction (CSC), which takes a sentence with some spelling errors as input and outputs the corrected one. However, CSC models may fail to correct spelling errors covered by the confusion sets, and also will encounter unseen ones. We propose a method, which continually identifies the weak spots of a model to generate more valuable training instances, and apply a task-specific pre-training strategy to enhance the model. The generated adversarial examples are gradually added to the training set. Experimental results show that such an adversarial training method combined with the pretraining strategy can improve both the generalization and robustness of multiple CSC models across three different datasets, achieving stateof-the-art performance for CSC task.

Chong Li, Cenyuan Zhang, Xiaoqing Zheng, Xuanjing Huang• 2021

Related benchmarks

TaskDatasetResultRank
Chinese Spelling CheckSIGHAN15 (test)--
53
Chinese Spelling CheckSIGHAN14 (test)
Correction F168.6
28
Chinese Spelling CheckSIGHAN13 (test)
Detection F184.9
16
Chinese Spell CheckingSIGHAN 2013 (ATTACK)
Detection F150.8
9
Chinese Spell CheckingSIGHAN ATTACK 2014
Detection F154.9
9
Chinese Spell CheckingSIGHAN 2015 (ATTACK)
Detection F145.9
9
Showing 6 of 6 rows

Other info

Code

Follow for update