CLEAR: A Comprehensive Linguistic Evaluation of Argument Rewriting by Large Language Models

Thomas Huber, Christina Niklaus


Abstract
While LLMs have been extensively studied on general text generation tasks, there is less research on text rewriting, a task related to general text generation, and particularly on the behavior of models on this task. In this paper we analyze what changes LLMs make in a text rewriting setting. We focus specifically on argumentative texts and their improvement, a task named Argument Improvement (ArgImp). We present CLEAR: an evaluation pipeline consisting of 57 metrics mapped to four linguistic levels: lexical, syntactic, semantic and pragmatic. This pipeline is used to examine the qualities of LLM-rewritten arguments on a broad set of argumentation corpora and compare the behavior of different LLMs on this task and analyze the behavior of different LLMs on this task in terms of linguistic levels. By taking all four linguistic levels into consideration, we find that the models perform ArgImp by shortening the texts while simultaneously increasing average word length and merging sentences. Overall we note an increase in the persuasion and coherence dimensions.
Anthology ID:
2025.findings-emnlp.1065
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2025
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
19548–19568
Language:
URL:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.1065/
DOI:
10.18653/v1/2025.findings-emnlp.1065
Bibkey:
Cite (ACL):
Thomas Huber and Christina Niklaus. 2025. CLEAR: A Comprehensive Linguistic Evaluation of Argument Rewriting by Large Language Models. In Findings of the Association for Computational Linguistics: EMNLP 2025, pages 19548–19568, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
CLEAR: A Comprehensive Linguistic Evaluation of Argument Rewriting by Large Language Models (Huber & Niklaus, Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.1065.pdf
Checklist:
 2025.findings-emnlp.1065.checklist.pdf