Alignment-Enhanced Transformer for Constraining NMT with Pre-Specified Translations

Authors

  • Kai Song Alibaba Group
  • Kun Wang Soochow University
  • Heng Yu Alibaba Group
  • Yue Zhang Westlake University
  • Zhongqiang Huang Alibaba Group
  • Weihua Luo Alibaba Group
  • Xiangyu Duan Soochow University
  • Min Zhang Soochow University

DOI:

https://doi.org/10.1609/aaai.v34i05.6418

Abstract

We investigate the task of constraining NMT with pre-specified translations, which has practical significance for a number of research and industrial applications. Existing works impose pre-specified translations as lexical constraints during decoding, which are based on word alignments derived from target-to-source attention weights. However, multiple recent studies have found that word alignment derived from generic attention heads in the Transformer is unreliable. We address this problem by introducing a dedicated head in the multi-head Transformer architecture to capture external supervision signals. Results on five language pairs show that our method is highly effective in constraining NMT with pre-specified translations, consistently outperforming previous methods in translation quality.

Downloads

Published

2020-04-03

How to Cite

Song, K., Wang, K., Yu, H., Zhang, Y., Huang, Z., Luo, W., Duan, X., & Zhang, M. (2020). Alignment-Enhanced Transformer for Constraining NMT with Pre-Specified Translations. Proceedings of the AAAI Conference on Artificial Intelligence, 34(05), 8886-8893. https://doi.org/10.1609/aaai.v34i05.6418

Issue

Section

AAAI Technical Track: Natural Language Processing