Sequence Level Contrastive Learning for Text Summarization

Authors

  • Shusheng Xu IIIS, Tsinghua University
  • Xingxing Zhang Microsoft Research Asia
  • Yi Wu IIIS, Tsinghua University Shanghai Qi Zhi Institute
  • Furu Wei Microsoft Research Asia

DOI:

https://doi.org/10.1609/aaai.v36i10.21409

Keywords:

Speech & Natural Language Processing (SNLP)

Abstract

Contrastive learning models have achieved great success in unsupervised visual representation learning, which maximize the similarities between feature representations of different views of the same image, while minimize the similarities between feature representations of views of different images. In text summarization, the output summary is a shorter form of the input document and they have similar meanings. In this paper, we propose a contrastive learning model for supervised abstractive text summarization, where we view a document, its gold summary and its model generated summaries as different views of the same mean representation and maximize the similarities between them during training. We improve over a strong sequence-to-sequence text generation model (i.e., BART) on three different summarization datasets. Human evaluation also shows that our model achieves better faithfulness ratings compared to its counterpart without contrastive objectives. We release our code at https://github.com/xssstory/SeqCo.

Downloads

Published

2022-06-28

How to Cite

Xu, S., Zhang, X., Wu, Y., & Wei, F. (2022). Sequence Level Contrastive Learning for Text Summarization. Proceedings of the AAAI Conference on Artificial Intelligence, 36(10), 11556-11565. https://doi.org/10.1609/aaai.v36i10.21409

Issue

Section

AAAI Technical Track on Speech and Natural Language Processing