@inproceedings{proebsting-poliak-2025-biases,
title = "Biases in Large Language Model-Elicited Text: A Case Study in Natural Language Inference",
author = "Proebsting, Grace and
Poliak, Adam",
editor = "Rambow, Owen and
Wanner, Leo and
Apidianaki, Marianna and
Al-Khalifa, Hend and
Eugenio, Barbara Di and
Schockaert, Steven",
booktitle = "Proceedings of the 31st International Conference on Computational Linguistics",
month = jan,
year = "2025",
address = "Abu Dhabi, UAE",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2025.coling-main.389/",
pages = "5836--5851",
abstract = "We test whether NLP datasets created with Large Language Models (LLMs) contain annotation artifacts and social biases like NLP datasets elicited from crowd-source workers. We recreate a portion of the Stanford Natural Language Inference corpus using GPT-4, Llama-2 70b for Chat, and Mistral 7b Instruct. We train hypothesis-only classifiers to determine whether LLM-elicited NLI datasets contain annotation artifacts. Next, we use point-wise mutual information to identify the words in each dataset that are associated with gender, race, and age-related terms. On our LLM-generated NLI datasets, fine-tuned BERT hypothesis-only classifiers achieve between 86-96{\%} accuracy. Our analyses further characterize the annotation artifacts and stereotypical biases in LLM-generated datasets."
}
Markdown (Informal)
[Biases in Large Language Model-Elicited Text: A Case Study in Natural Language Inference](https://preview.aclanthology.org/fix-sig-urls/2025.coling-main.389/) (Proebsting & Poliak, COLING 2025)
ACL