CAR-Transformer: Cross-Attention Reinforcement Transformer for Cross-Lingual Summarization

Authors

  • Yuang Cai Beijing University of Posts and Telecommunications Key Laboratory of Trustworthy Distributed Computing and Service (BUPT), Ministry of Education
  • Yuyu Yuan Beijing University of Posts and Telecommunications Key Laboratory of Trustworthy Distributed Computing and Service (BUPT), Ministry of Education

DOI:

https://doi.org/10.1609/aaai.v38i16.29724

Keywords:

NLP: Summarization

Abstract

Cross-Lingual Summarization (CLS) involves generating a summary for a given document in another language. Most of the existing approaches adopt multi-task training and knowledge distillation, which increases the training cost and improves the performance of CLS tasks intuitively but unexplainably. In this work, we propose Cross-Attention Reinforcement (CAR) module and incorporate the module into the transformer backbone to formulate the CAR-Transformer. The CAR module formulates a pseudo summarization policy parameterized by the cross-attention weights reinforced by the ground-truth monolingual summary without introducing extra model parameters. Our approach demonstrates more consistent improvement across CLS tasks compared to traditional multi-task training methods and outperforms the fine-tuned vanilla mBART by 3.67 and the best-performing multi-task training approach by 1.48 in ROUGE-L F1 score on the WikiLingua Korean-to-English CLS task.

Published

2024-03-24

How to Cite

Cai, Y., & Yuan, Y. (2024). CAR-Transformer: Cross-Attention Reinforcement Transformer for Cross-Lingual Summarization. Proceedings of the AAAI Conference on Artificial Intelligence, 38(16), 17718-17726. https://doi.org/10.1609/aaai.v38i16.29724

Issue

Section

AAAI Technical Track on Natural Language Processing I