Denoising Pre-training for Machine Translation Quality Estimation with Curriculum Learning

Authors

  • Xiang Geng National Key Laboratory for Novel Software Technology, Nanjing University
  • Yu Zhang National Key Laboratory for Novel Software Technology, Nanjing University
  • Jiahuan Li National Key Laboratory for Novel Software Technology, Nanjing University
  • Shujian Huang National Key Laboratory for Novel Software Technology, Nanjing University
  • Hao Yang Huawei
  • Shimin Tao Huawei
  • Yimeng Chen Huawei
  • Ning Xie Huawei
  • Jiajun Chen National Key Laboratory for Novel Software Technology, Nanjing University

DOI:

https://doi.org/10.1609/aaai.v37i11.26508

Keywords:

SNLP: Machine Translation & Multilinguality

Abstract

Quality estimation (QE) aims to assess the quality of machine translations when reference translations are unavailable. QE plays a crucial role in many real-world applications of machine translation. Because labeled QE data are usually limited in scale, recent research, such as DirectQE, pre-trains QE models with pseudo QE data and obtains remarkable performance. However, there tends to be inevitable noise in the pseudo data, hindering models from learning QE accurately. Our study shows that the noise mainly comes from the differences between pseudo and real translation outputs. To handle this problem, we propose CLQE, a denoising pre-training framework for QE based on curriculum learning. More specifically, we propose to measure the degree of noise in the pseudo QE data with some metrics based on statistical or distributional features. With the guidance of these metrics, CLQE gradually pre-trains the QE model using data from cleaner to noisier. Experiments on various benchmarks reveal that CLQE outperforms DirectQE and other strong baselines. We also show that with our framework, pre-training converges faster than directly using the pseudo data. We make our CLQE code available (https://github.com/NJUNLP/njuqe).

Downloads

Published

2023-06-26

How to Cite

Geng, X., Zhang, Y., Li, J., Huang, S., Yang, H., Tao, S., Chen, Y., Xie, N., & Chen, J. (2023). Denoising Pre-training for Machine Translation Quality Estimation with Curriculum Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 37(11), 12827-12835. https://doi.org/10.1609/aaai.v37i11.26508

Issue

Section

AAAI Technical Track on Speech & Natural Language Processing