Guardrails, not Guidance: Understanding Responses to LGBTQ+ Language in Large Language Models

Joshua Tint


Abstract
Language models have integrated themselves into many aspects of digital life, shaping everything from social media to translation. This paper investigates how large language models (LLMs) respond to LGBTQ+ slang and heteronormative language. Through two experiments, the study assesses the emotional content and the impact of queer slang on responses from models including GPT-3.5, GPT-4o, Llama2, Llama3, Gemma and Mistral. The findings reveal that heteronormative prompts can trigger safety mechanisms, leading to neutral or corrective responses, while LGBTQ+ slang elicits more negative emotions. These insights punctuate the need to provide equitable outcomes for minority slangs and argots, in addition to eliminating explicit bigotry from language models.
Anthology ID:
2025.queerinai-main.2
Volume:
Proceedings of the Queer in AI Workshop
Month:
May
Year:
2025
Address:
Hybrid format (in-person and virtual)
Editors:
A Pranav, Alissa Valentine, Shaily Bhatt, Yanan Long, Arjun Subramonian, Amanda Bertsch, Anne Lauscher, Ankush Gupta
Venues:
QueerInAI | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
6–16
Language:
URL:
https://preview.aclanthology.org/fix-sig-urls/2025.queerinai-main.2/
DOI:
Bibkey:
Cite (ACL):
Joshua Tint. 2025. Guardrails, not Guidance: Understanding Responses to LGBTQ+ Language in Large Language Models. In Proceedings of the Queer in AI Workshop, pages 6–16, Hybrid format (in-person and virtual). Association for Computational Linguistics.
Cite (Informal):
Guardrails, not Guidance: Understanding Responses to LGBTQ+ Language in Large Language Models (Tint, QueerInAI 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/fix-sig-urls/2025.queerinai-main.2.pdf