@inproceedings{kaur-2025-echoes,
title = "Echoes of Agreement: Argument Driven Sycophancy in Large Language models",
author = "Kaur, Avneet",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.1241/",
doi = "10.18653/v1/2025.findings-emnlp.1241",
pages = "22803--22812",
ISBN = "979-8-89176-335-7",
abstract = "Existing evaluation of political biases in Large Language Models (LLMs) outline the high sensitivity to prompt formulation. Furthermore, Large Language Models are known to exhibit sycophancy, a tendency to align their outputs with a user{'}s stated belief, which is often attributed to human feedback during fine-tuning. However, such bias in the presence of explicit argumentation within a prompt remains underexplored. This paper investigates how argumentative prompts induce sycophantic behaviour in LLMs in a political context. Through a series of experiments, we demonstrate that models consistently alter their responses to mirror the stance present expressed by the user. This sycophantic behaviour is observed in both single and multi-turn interactions, and its intensity correlates with argument strength. Our findings establish a link between user stance and model sycophancy, revealing a critical vulnerability that impacts model reliability. Thus has significant implications for models being deployed in real-world settings and calls for developing robust evaluations and mitigations against manipulative or biased interactions."
}Markdown (Informal)
[Echoes of Agreement: Argument Driven Sycophancy in Large Language models](https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.1241/) (Kaur, Findings 2025)
ACL