Multi-Task Learning for Abstractive and Extractive Summarization

Yangbin Chen*, Yun Ma, Xudong Mao, Qing Li

*Corresponding author for this work

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

46 Citations (Scopus)
88 Downloads (CityUHK Scholars)

Abstract

The abstractive method and extractive method are two main approaches for automatic document summarization. In this paper, to fully integrate the relatedness and advantages of both approaches, we propose a general unified framework for abstractive summarization which incorporates extractive summarization as an auxiliary task. In particular, our framework is composed of a shared hierarchical document encoder, a hierarchical attention mechanism-based decoder, and an extractor. We adopt multi-task learning method to train these two tasks jointly, which enables the shared encoder to better capture the semantics of the document. Moreover, as our main task is abstractive summarization, we constrain the attention learned in the abstractive task with the labels of the extractive task to strengthen the consistency between the two tasks. Experiments on the CNN/DailyMail dataset demonstrate that both the auxiliary task and the attention constraint contribute to improve the performance significantly, and our model is comparable to the state-of-the-art abstractive models. In addition, we cut half number of labels of the extractive task, pretrain the extractor, and jointly train the two tasks using the estimated sentence salience of the extractive task to constrain the attention of the abstractive task. The results do not decrease much compared with using full-labeled data of the auxiliary task.
Original languageEnglish
Pages (from-to)14-23
JournalData Science and Engineering
Volume4
Issue number1
DOIs
Publication statusPublished - Mar 2019

Bibliographical note

Full text of this publication does not contain sufficient affiliation information. With consent from the author(s) concerned, the Research Unit(s) information for this record is based on the existing academic department affiliation of the author(s).

Research Keywords

  • Attention mechanism
  • Multi-task learning
  • Automatic document summarization

Publisher's Copyright Statement

  • This full text is made available under CC-BY 4.0. https://creativecommons.org/licenses/by/4.0/

Fingerprint

Dive into the research topics of 'Multi-Task Learning for Abstractive and Extractive Summarization'. Together they form a unique fingerprint.
  • DON: Sentiment-preserving Text Summarization

    LI, Q. (Principal Investigator / Project Coordinator)

    1/09/1726/06/18

    Project: Research

Cite this