Chinese Spelling Correction as Rephrasing Language Model

Authors

  • Linfeng Liu Shanghai Jiao Tong University
  • Hongqiu Wu Shanghai Jiao Tong University
  • Hai Zhao Shanghai Jiao Tong University

DOI:

https://doi.org/10.1609/aaai.v38i17.29829

Keywords:

NLP: Applications

Abstract

This paper studies Chinese Spelling Correction (CSC), which aims to detect and correct potential spelling errors in a given sentence. Current state-of-the-art methods regard CSC as a sequence tagging task and fine-tune BERT-based models on sentence pairs. However, we note a critical flaw in the process of tagging one character to another, that the correction is excessively conditioned on the error. This is opposite from human mindset, where individuals rephrase the complete sentence based on its semantics, rather than solely on the error patterns memorized before. Such a counter-intuitive learning process results in the bottleneck of generalizability and transferability of machine spelling correction. To address this, we propose Rephrasing Language Modeling (ReLM), where the model is trained to rephrase the entire sentence by infilling additional slots, instead of character-to-character tagging. This novel training paradigm achieves the new state-of-theart results across fine-tuned and zero-shot CSC benchmarks, outperforming previous counterparts by a large margin. Our method also learns transferable language representation when CSC is jointly trained with other tasks.

Published

2024-03-24

How to Cite

Liu, L., Wu, H., & Zhao, H. (2024). Chinese Spelling Correction as Rephrasing Language Model. Proceedings of the AAAI Conference on Artificial Intelligence, 38(17), 18662-18670. https://doi.org/10.1609/aaai.v38i17.29829

Issue

Section

AAAI Technical Track on Natural Language Processing II