Rephrasing the Reference for Non-autoregressive Machine Translation
DOI:
https://doi.org/10.1609/aaai.v37i11.26587Keywords:
SNLP: Machine Translation & Multilinguality, SNLP: GenerationAbstract
Non-autoregressive neural machine translation (NAT) models suffer from the multi-modality problem that there may exist multiple possible translations of a source sentence, so the reference sentence may be inappropriate for the training when the NAT output is closer to other translations. In response to this problem, we introduce a rephraser to provide a better training target for NAT by rephrasing the reference sentence according to the NAT output. As we train NAT based on the rephraser output rather than the reference sentence, the rephraser output should fit well with the NAT output and not deviate too far from the reference, which can be quantified as reward functions and optimized by reinforcement learning. Experiments on major WMT benchmarks and NAT baselines show that our approach consistently improves the translation quality of NAT. Specifically, our best variant achieves comparable performance to the autoregressive Transformer, while being 14.7 times more efficient in inference.Downloads
Published
2023-06-26
How to Cite
Shao, C., Zhang, J., Zhou, J., & Feng, Y. (2023). Rephrasing the Reference for Non-autoregressive Machine Translation. Proceedings of the AAAI Conference on Artificial Intelligence, 37(11), 13538-13546. https://doi.org/10.1609/aaai.v37i11.26587
Issue
Section
AAAI Technical Track on Speech & Natural Language Processing