Towards Document-Level Paraphrase Generation with Sentence Rewriting and Reordering

Zhe Lin, Yitao Cai, Xiaojun Wan


Abstract
Paraphrase generation is an important task in natural language processing. Previous works focus on sentence-level paraphrase generation, while ignoring document-level paraphrase generation, which is a more challenging and valuable task. In this paper, we explore the task of document-level paraphrase generation for the first time and focus on the inter-sentence diversity by considering sentence rewriting and reordering. We propose CoRPG (Coherence Relationship guided Paraphrase Generation), which leverages graph GRU to encode the coherence relationship graph and get the coherence-aware representation for each sentence, which can be used for re-arranging the multiple (possibly modified) input sentences. We create a pseudo document-level paraphrase dataset for training CoRPG. Automatic evaluation results show CoRPG outperforms several strong baseline models on the BERTScore and diversity scores. Human evaluation also shows our model can generate document paraphrase with more diversity and semantic preservation.
Anthology ID:
2021.findings-emnlp.89
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2021
Month:
November
Year:
2021
Address:
Punta Cana, Dominican Republic
Editors:
Marie-Francine Moens, Xuanjing Huang, Lucia Specia, Scott Wen-tau Yih
Venue:
Findings
SIG:
SIGDAT
Publisher:
Association for Computational Linguistics
Note:
Pages:
1033–1044
Language:
URL:
https://aclanthology.org/2021.findings-emnlp.89
DOI:
10.18653/v1/2021.findings-emnlp.89
Bibkey:
Cite (ACL):
Zhe Lin, Yitao Cai, and Xiaojun Wan. 2021. Towards Document-Level Paraphrase Generation with Sentence Rewriting and Reordering. In Findings of the Association for Computational Linguistics: EMNLP 2021, pages 1033–1044, Punta Cana, Dominican Republic. Association for Computational Linguistics.
Cite (Informal):
Towards Document-Level Paraphrase Generation with Sentence Rewriting and Reordering (Lin et al., Findings 2021)
Copy Citation:
PDF:
https://preview.aclanthology.org/emnlp-22-attachments/2021.findings-emnlp.89.pdf
Code
 l-zhe/corpg