@inproceedings{kwon-etal-2023-abstractive,
title = "Abstractive Document Summarization with Summary-length Prediction",
author = "Kwon, Jingun and
Kamigaito, Hidetaka and
Okumura, Manabu",
editor = "Vlachos, Andreas and
Augenstein, Isabelle",
booktitle = "Findings of the Association for Computational Linguistics: EACL 2023",
month = may,
year = "2023",
address = "Dubrovnik, Croatia",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.findings-eacl.45/",
doi = "10.18653/v1/2023.findings-eacl.45",
pages = "618--624",
abstract = "Recently, we can obtain a practical abstractive document summarization model by fine-tuning a pre-trained language model (PLM). Since the pre-training for PLMs does not consider summarization-specific information such as the target summary length, there is a gap between the pre-training and fine-tuning for PLMs in summarization tasks. To fill the gap, we propose a method for enabling the model to understand the summarization-specific information by predicting the summary length in the encoder and generating a summary of the predicted length in the decoder in fine-tuning. Experimental results on the WikiHow, NYT, and CNN/DM datasets showed that our methods improve ROUGE scores from BART by generating summaries of appropriate lengths. Further, we observed about 3.0, 1,5, and 3.1 point improvements for ROUGE-1, -2, and -L, respectively, from GSum on the WikiHow dataset. Human evaluation results also showed that our methods improve the informativeness and conciseness of summaries."
}
Markdown (Informal)
[Abstractive Document Summarization with Summary-length Prediction](https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.findings-eacl.45/) (Kwon et al., Findings 2023)
ACL