PSDSpell: Pre-Training with Self-Distillation Learning for Chinese Spelling Correction

Li He, Xiaowu Zhang, Jianyong Duan, Hao Wang,Xin Li, Liang Zhao

IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS(2024)

引用 0|浏览1
暂无评分
摘要
Chinese spelling correction (CSC) models detect and correct a text typo based on the misspelled character and its context. Recently, Bert -based models have dominated the research of Chinese spelling correction . However, these methods only focus on the semantic information of the text during the pretraining stage, neglecting the learning of correcting spelling errors. Moreover, when multiple incorrect characters are in the text, the context introduces noisy information, making it difficult for the model to accurately detect the positions of the incorrect characters, leading to false corrections. To address these limitations, we apply the multimodal pretrained language model ChineseBert to the task of spelling correction. We propose a self-distillation learning-based pretraining strategy, where a confusion set is used to construct text containing erroneous characters, allowing the model to jointly learns how to understand language and correct spelling errors. Additionally, we introduce a single-channel masking mechanism to mitigate the noise caused by the incorrect characters. This mechanism masks the semantic encoding channel while preserving the phonetic and glyph encoding channels, reducing the noise introduced by incorrect characters during the prediction process. Finally, experiments are conducted on widely used benchmarks. Our model achieves superior performance against state -of -the -art methods by a remarkable gain.
更多
查看译文
关键词
spelling correction,ChineseBert,self-distillation,multimodal information
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要