Controllable Abstractive Summarization

Angela Fan, David Grangier, Michael Auli


Abstract
Current models for document summarization disregard user preferences such as the desired length, style, the entities that the user might be interested in, or how much of the document the user has already read. We present a neural summarization model with a simple but effective mechanism to enable users to specify these high level attributes in order to control the shape of the final summaries to better suit their needs. With user input, our system can produce high quality summaries that follow user preferences. Without user input, we set the control variables automatically – on the full text CNN-Dailymail dataset, we outperform state of the art abstractive systems (both in terms of F1-ROUGE1 40.38 vs. 39.53 F1-ROUGE and human evaluation.
Anthology ID:
W18-2706
Volume:
Proceedings of the 2nd Workshop on Neural Machine Translation and Generation
Month:
July
Year:
2018
Address:
Melbourne, Australia
Editors:
Alexandra Birch, Andrew Finch, Thang Luong, Graham Neubig, Yusuke Oda
Venue:
NGT
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
45–54
Language:
URL:
https://preview.aclanthology.org/build-pipeline-with-new-library/W18-2706/
DOI:
10.18653/v1/W18-2706
Bibkey:
Cite (ACL):
Angela Fan, David Grangier, and Michael Auli. 2018. Controllable Abstractive Summarization. In Proceedings of the 2nd Workshop on Neural Machine Translation and Generation, pages 45–54, Melbourne, Australia. Association for Computational Linguistics.
Cite (Informal):
Controllable Abstractive Summarization (Fan et al., NGT 2018)
Copy Citation:
PDF:
https://preview.aclanthology.org/build-pipeline-with-new-library/W18-2706.pdf
Data
CNN/Daily Mail