Abstract
Query Focused Abstractive Summarization (QFAS) represents an abstractive summary from the source document based on a given query. To measure the performance of abstractive summarization tasks, different datasets have been broadly used. However, for QFAS tasks, only a limited number of datasets have been used, which are comparatively small and provide single sentence summaries. This paper presents a query generation approach, where we considered most similar words between documents and summaries for generating queries. By implementing our query generation approach, we prepared two relatively large datasets, namely CNN/DailyMail and Newsroom which contain multiple sentence summaries and can be used for future QFAS tasks. We also implemented a pre-processing approach to perform QFAS tasks using a pretrained language model, BERTSUM. In our pre-processing approach, we sorted the sentences of the documents from the most query-related sentences to the less query-related sentences. Then, we fine-tuned the BERTSUM model for generating the abstractive summaries. We also experimented on one of the largely used datasets, Debatepedia, to compare our QFAS approach with other models. The experimental results show that our approach outperforms the state-of-the-art models on three ROUGE scores.- Anthology ID:
- 2020.inlg-1.11
- Volume:
- Proceedings of the 13th International Conference on Natural Language Generation
- Month:
- December
- Year:
- 2020
- Address:
- Dublin, Ireland
- Editors:
- Brian Davis, Yvette Graham, John Kelleher, Yaji Sripada
- Venue:
- INLG
- SIG:
- SIGGEN
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 80–85
- Language:
- URL:
- https://aclanthology.org/2020.inlg-1.11
- DOI:
- 10.18653/v1/2020.inlg-1.11
- Cite (ACL):
- Deen Mohammad Abdullah and Yllias Chali. 2020. Towards Generating Query to Perform Query Focused Abstractive Summarization using Pre-trained Model. In Proceedings of the 13th International Conference on Natural Language Generation, pages 80–85, Dublin, Ireland. Association for Computational Linguistics.
- Cite (Informal):
- Towards Generating Query to Perform Query Focused Abstractive Summarization using Pre-trained Model (Abdullah & Chali, INLG 2020)
- PDF:
- https://preview.aclanthology.org/nschneid-patch-2/2020.inlg-1.11.pdf
- Code
- deen-abdullah/QABSBERT
- Data
- CNN/Daily Mail, NEWSROOM