Perspective-driven Preference Optimization with Entropy Maximization for Diverse Argument Generation

Yilin Cao, Ruike Zhang, Penghui Wei, Qingchao Kong, Wenji Mao


Abstract
In subjective natural language generation tasks, generating diverse perspectives is essential for fostering balanced discourse and mitigating bias. Argument generation with diverse perspectives plays a vital role in advancing the understanding of controversial claims. Despite the strong generative capabilities of large language models (LLMs), the diversity of perspectives remains insufficiently explored within argument generation task. Moreover, there remains a significant research gap in developing methods that explicitly generate multi-perspective arguments under the quality control of claim-stance alignment constraints. In this paper, we propose POEM, a Perspective-aware Preference Optimization with Entropy Maximization framework for diverse argument generation. It enhances perspective diversity through preference optimization based on the constructed preference dataset via perspective mining and diversity measuring. It further introduces entropy maximization to promote perspective diversity by encouraging dispersed semantic representations among the generated arguments. Experimental results on claim-stance argument generation benchmarks show that POEM is capable of generating diverse arguments while maintaining comparable performances in claim and stance controllability as well as text quality compared to the state-of-the-art baselines and human evaluation.
Anthology ID:
2025.findings-emnlp.1223
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2025
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
22479–22496
Language:
URL:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.1223/
DOI:
10.18653/v1/2025.findings-emnlp.1223
Bibkey:
Cite (ACL):
Yilin Cao, Ruike Zhang, Penghui Wei, Qingchao Kong, and Wenji Mao. 2025. Perspective-driven Preference Optimization with Entropy Maximization for Diverse Argument Generation. In Findings of the Association for Computational Linguistics: EMNLP 2025, pages 22479–22496, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Perspective-driven Preference Optimization with Entropy Maximization for Diverse Argument Generation (Cao et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.1223.pdf
Checklist:
 2025.findings-emnlp.1223.checklist.pdf