@inproceedings{bae-etal-2019-summary,
    title = "Summary Level Training of Sentence Rewriting for Abstractive Summarization",
    author = "Bae, Sanghwan  and
      Kim, Taeuk  and
      Kim, Jihoon  and
      Lee, Sang-goo",
    editor = "Wang, Lu  and
      Cheung, Jackie Chi Kit  and
      Carenini, Giuseppe  and
      Liu, Fei",
    booktitle = "Proceedings of the 2nd Workshop on New Frontiers in Summarization",
    month = nov,
    year = "2019",
    address = "Hong Kong, China",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/D19-5402",
    doi = "10.18653/v1/D19-5402",
    pages = "10--20",
    abstract = "As an attempt to combine extractive and abstractive summarization, Sentence Rewriting models adopt the strategy of extracting salient sentences from a document first and then paraphrasing the selected ones to generate a summary. However, the existing models in this framework mostly rely on sentence-level rewards or suboptimal labels, causing a mismatch between a training objective and evaluation metric. In this paper, we present a novel training signal that directly maximizes summary-level ROUGE scores through reinforcement learning. In addition, we incorporate BERT into our model, making good use of its ability on natural language understanding. In extensive experiments, we show that a combination of our proposed model and training procedure obtains new state-of-the-art performance on both CNN/Daily Mail and New York Times datasets. We also demonstrate that it generalizes better on DUC-2002 test set.",
}
Markdown (Informal)
[Summary Level Training of Sentence Rewriting for Abstractive Summarization](https://aclanthology.org/D19-5402) (Bae et al., 2019)
ACL