Leveraging Large Language Models in Detecting Anti-LGBTQIA+ User-generated Texts

Quoc-Toan Nguyen, Josh Nguyen, Tuan Pham, William John Teahan


Abstract
Anti-LGBTQIA+ texts in user-generated content pose significant risks to online safety and inclusivity. This study investigates the capabilities and limitations of five widely adopted Large Language Models (LLMs)—DeepSeek-V3, GPT-4o, GPT-4o-mini, GPT-o1-mini, and Llama3.3-70B—in detecting such harmful content. Our findings reveal that while LLMs demonstrate potential in identifying offensive language, their effectiveness varies across models and metrics, with notable shortcomings in calibration. Furthermore, linguistic analysis exposes deeply embedded patterns of discrimination, reinforcing the urgency for improved detection mechanisms for this marginalised population. In summary, this study demonstrates the significant potential of LLMs for practical application in detecting anti-LGBTQIA+ user-generated texts and provides valuable insights from text analysis that can inform topic modelling. These findings contribute to developing safer digital platforms and enhancing protection for LGBTQIA+ individuals.
Anthology ID:
2025.queerinai-main.4
Volume:
Proceedings of the Queer in AI Workshop
Month:
May
Year:
2025
Address:
Hybrid format (in-person and virtual)
Editors:
A Pranav, Alissa Valentine, Shaily Bhatt, Yanan Long, Arjun Subramonian, Amanda Bertsch, Anne Lauscher, Ankush Gupta
Venues:
QueerInAI | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
26–34
Language:
URL:
https://preview.aclanthology.org/fix-sig-urls/2025.queerinai-main.4/
DOI:
Bibkey:
Cite (ACL):
Quoc-Toan Nguyen, Josh Nguyen, Tuan Pham, and William John Teahan. 2025. Leveraging Large Language Models in Detecting Anti-LGBTQIA+ User-generated Texts. In Proceedings of the Queer in AI Workshop, pages 26–34, Hybrid format (in-person and virtual). Association for Computational Linguistics.
Cite (Informal):
Leveraging Large Language Models in Detecting Anti-LGBTQIA+ User-generated Texts (Nguyen et al., QueerInAI 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/fix-sig-urls/2025.queerinai-main.4.pdf