Reading Between the Prompts: How Stereotypes Shape LLM’s Implicit Personalization

Vera Neplenbroek, Arianna Bisazza, Raquel Fernández


Abstract
Generative Large Language Models (LLMs) infer user’s demographic information from subtle cues in the conversation — a phenomenon called implicit personalization. Prior work has shown that such inferences can lead to lower quality responses for users assumed to be from minority groups, even when no demographic information is explicitly provided. In this work, we systematically explore how LLMs respond to stereotypical cues using controlled synthetic conversations, by analyzing the models’ latent user representations through both model internals and generated answers to targeted user questions. Our findings reveal that LLMs do infer demographic attributes based on these stereotypical signals, which for a number of groups even persists when the user explicitly identifies with a different demographic group. Finally, we show that this form of stereotype-driven implicit personalization can be effectively mitigated by intervening on the model’s internal representations using a trained linear probe to steer them toward the explicitly stated identity. Our results highlight the need for greater transparency and control in how LLMs represent user identity.
Anthology ID:
2025.emnlp-main.1029
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
20378–20411
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1029/
DOI:
Bibkey:
Cite (ACL):
Vera Neplenbroek, Arianna Bisazza, and Raquel Fernández. 2025. Reading Between the Prompts: How Stereotypes Shape LLM’s Implicit Personalization. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 20378–20411, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Reading Between the Prompts: How Stereotypes Shape LLM’s Implicit Personalization (Neplenbroek et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1029.pdf
Checklist:
 2025.emnlp-main.1029.checklist.pdf