CAR-Transformer: Cross-Attention Reinforcement Transformer for Cross-Lingual Summarization
DOI:
https://doi.org/10.1609/aaai.v38i16.29724Keywords:
NLP: SummarizationAbstract
Cross-Lingual Summarization (CLS) involves generating a summary for a given document in another language. Most of the existing approaches adopt multi-task training and knowledge distillation, which increases the training cost and improves the performance of CLS tasks intuitively but unexplainably. In this work, we propose Cross-Attention Reinforcement (CAR) module and incorporate the module into the transformer backbone to formulate the CAR-Transformer. The CAR module formulates a pseudo summarization policy parameterized by the cross-attention weights reinforced by the ground-truth monolingual summary without introducing extra model parameters. Our approach demonstrates more consistent improvement across CLS tasks compared to traditional multi-task training methods and outperforms the fine-tuned vanilla mBART by 3.67 and the best-performing multi-task training approach by 1.48 in ROUGE-L F1 score on the WikiLingua Korean-to-English CLS task.Downloads
Published
2024-03-24
How to Cite
Cai, Y., & Yuan, Y. (2024). CAR-Transformer: Cross-Attention Reinforcement Transformer for Cross-Lingual Summarization. Proceedings of the AAAI Conference on Artificial Intelligence, 38(16), 17718-17726. https://doi.org/10.1609/aaai.v38i16.29724
Issue
Section
AAAI Technical Track on Natural Language Processing I