Rephrasing the Reference for Non-autoregressive Machine Translation

Authors

  • Chenze Shao Institute of Computing Technology, Chinese Academy of Sciences University of Chinese Academy of Sciences
  • Jinchao Zhang Tencent
  • Jie Zhou Tencent
  • Yang Feng Institute of Computing Technology, Chinese Academy of Sciences University of Chinese Academy of Sciences

DOI:

https://doi.org/10.1609/aaai.v37i11.26587

Keywords:

SNLP: Machine Translation & Multilinguality, SNLP: Generation

Abstract

Non-autoregressive neural machine translation (NAT) models suffer from the multi-modality problem that there may exist multiple possible translations of a source sentence, so the reference sentence may be inappropriate for the training when the NAT output is closer to other translations. In response to this problem, we introduce a rephraser to provide a better training target for NAT by rephrasing the reference sentence according to the NAT output. As we train NAT based on the rephraser output rather than the reference sentence, the rephraser output should fit well with the NAT output and not deviate too far from the reference, which can be quantified as reward functions and optimized by reinforcement learning. Experiments on major WMT benchmarks and NAT baselines show that our approach consistently improves the translation quality of NAT. Specifically, our best variant achieves comparable performance to the autoregressive Transformer, while being 14.7 times more efficient in inference.

Downloads

Published

2023-06-26

How to Cite

Shao, C., Zhang, J., Zhou, J., & Feng, Y. (2023). Rephrasing the Reference for Non-autoregressive Machine Translation. Proceedings of the AAAI Conference on Artificial Intelligence, 37(11), 13538-13546. https://doi.org/10.1609/aaai.v37i11.26587

Issue

Section

AAAI Technical Track on Speech & Natural Language Processing