Compressed Heterogeneous Graph for Abstractive Multi-Document Summarization

Authors

  • Miao Li School of Computing and Information Systems, The University of Melbourne
  • Jianzhong Qi School of Computing and Information Systems, The University of Melbourne
  • Jey Han Lau School of Computing and Information Systems, The University of Melbourne

DOI:

https://doi.org/10.1609/aaai.v37i11.26537

Keywords:

SNLP: Summarization, ML: Graph-based Machine Learning, ML: Transfer, Domain Adaptation, Multi-Task Learning, SNLP: Generation

Abstract

Multi-document summarization (MDS) aims to generate a summary for a number of related documents. We propose HGSum — an MDS model that extends an encoder-decoder architecture to incorporate a heterogeneous graph to represent different semantic units (e.g., words and sentences) of the documents. This contrasts with existing MDS models which do not consider different edge types of graphs and as such do not capture the diversity of relationships in the documents. To preserve only key information and relationships of the documents in the heterogeneous graph, HGSum uses graph pooling to compress the input graph. And to guide HGSum to learn the compression, we introduce an additional objective that maximizes the similarity between the compressed graph and the graph constructed from the ground-truth summary during training. HGSum is trained end-to-end with the graph similarity and standard cross-entropy objectives. Experimental results over Multi-News, WCEP-100, and Arxiv show that HGSum outperforms state-of-the-art MDS models. The code for our model and experiments is available at: https://github.com/oaimli/HGSum.

Downloads

Published

2023-06-26

How to Cite

Li, M., Qi, J., & Lau, J. H. (2023). Compressed Heterogeneous Graph for Abstractive Multi-Document Summarization. Proceedings of the AAAI Conference on Artificial Intelligence, 37(11), 13085-13093. https://doi.org/10.1609/aaai.v37i11.26537

Issue

Section

AAAI Technical Track on Speech & Natural Language Processing