2025
pdf
bib
abs
SHADES: Towards a Multilingual Assessment of Stereotypes in Large Language Models
Margaret Mitchell
|
Giuseppe Attanasio
|
Ioana Baldini
|
Miruna Clinciu
|
Jordan Clive
|
Pieter Delobelle
|
Manan Dey
|
Sil Hamilton
|
Timm Dill
|
Jad Doughman
|
Ritam Dutt
|
Avijit Ghosh
|
Jessica Zosa Forde
|
Carolin Holtermann
|
Lucie-Aimée Kaffee
|
Tanmay Laud
|
Anne Lauscher
|
Roberto L Lopez-Davila
|
Maraim Masoud
|
Nikita Nangia
|
Anaelia Ovalle
|
Giada Pistilli
|
Dragomir Radev
|
Beatrice Savoldi
|
Vipul Raheja
|
Jeremy Qin
|
Esther Ploeger
|
Arjun Subramonian
|
Kaustubh Dhole
|
Kaiser Sun
|
Amirbek Djanibekov
|
Jonibek Mansurov
|
Kayo Yin
|
Emilio Villa Cueva
|
Sagnik Mukherjee
|
Jerry Huang
|
Xudong Shen
|
Jay Gala
|
Hamdan Al-Ali
|
Tair Djanibekov
|
Nurdaulet Mukhituly
|
Shangrui Nie
|
Shanya Sharma
|
Karolina Stanczak
|
Eliza Szczechla
|
Tiago Timponi Torrent
|
Deepak Tunuguntla
|
Marcelo Viridiano
|
Oskar Van Der Wal
|
Adina Yakefu
|
Aurélie Névéol
|
Mike Zhang
|
Sydney Zink
|
Zeerak Talat
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Large Language Models (LLMs) reproduce and exacerbate the social biases present in their training data, and resources to quantify this issue are limited. While research has attempted to identify and mitigate such biases, most efforts have been concentrated around English, lagging the rapid advancement of LLMs in multilingual settings. In this paper, we introduce a new multilingual parallel dataset SHADES to help address this issue, designed for examining culturally-specific stereotypes that may be learned by LLMs. The dataset includes stereotypes from 20 regions around the world and 16 languages, spanning multiple identity categories subject to discrimination worldwide. We demonstrate its utility in a series of exploratory evaluations for both “base” and “instruction-tuned” language models. Our results suggest that stereotypes are consistently reflected across models and languages, with some languages and models indicating much stronger stereotype biases than others.
pdf
bib
abs
A City of Millions: Mapping Literary Social Networks At Scale
Sil Hamilton
|
Rebecca Hicke
|
David Mimno
|
Matthew Wilkens
Proceedings of the 5th International Conference on Natural Language Processing for Digital Humanities
We release 70,509 high-quality social networks extracted from multilingual fiction and nonfiction narratives. We additionally provide metadata for ~30,000 of these texts (73% nonfiction and 27% fiction) written between 1800 and 1999 in 58 languages. This dataset provides information on historical social worlds at an unprecedented scale, including data for 2,510,021 individuals in 2,805,482 pair-wise relationships annotated for affinity and relationship type. We achieve this scale by automating previously manual methods of extracting social networks; specifically, we adapt an existing annotation task as a language model prompt, ensuring consistency at scale with the use of structured output. This dataset serves as a unique resource for humanities and social science research by providing data on cognitive models of social realities.
2024
pdf
bib
abs
Detecting Mode Collapse in Language Models via Narration
Sil Hamilton
Proceedings of the First edition of the Workshop on the Scaling Behavior of Large Language Models (SCALE-LLM 2024)
No two authors write alike. Personal flourishes invoked in written narratives, from lexicon to rhetorical devices, imply a particular author—what literary theorists label the implied or virtual author; distinct from the real author or narrator of a text. Early large language models trained on unfiltered training sets drawn from a variety of discordant sources yielded incoherent personalities, problematic for conversational tasks but proving useful for sampling literature from multiple perspectives. Successes in alignment research in recent years have allowed researchers to impose subjectively consistent personae on language models via instruction tuning and reinforcement learning from human feedback (RLHF), but whether aligned models retain the ability to model an arbitrary virtual author has received little scrutiny. By studying 4,374 stories sampled from three OpenAI language models, we show successive versions of GPT-3 suffer from increasing degrees of “mode collapse” whereby overfitting the model during alignment constrains it from generalizing over authorship: models suffering from mode collapse become unable to assume a multiplicity of perspectives. Our method and results are significant for researchers seeking to employ language models in sociological simulations.
2023
pdf
bib
abs
Mrs. Dalloway Said She Would Segment the Chapters Herself
Peiqi Sui
|
Lin Wang
|
Sil Hamilton
|
Thorsten Ries
|
Kelvin Wong
|
Stephen Wong
Proceedings of the 5th Workshop on Narrative Understanding
This paper proposes a sentiment-centric pipeline to perform unsupervised plot extraction on non-linear novels like Virginia Woolf’s Mrs. Dalloway, a novel widely considered to be “plotless. Combining transformer-based sentiment analysis models with statistical testing, we model sentiment’s rate-of-change and correspondingly segment the novel into emotionally self-contained units qualitatively evaluated to be meaningful surrogate pseudo-chapters. We validate our findings by evaluating our pipeline as a fully unsupervised text segmentation model, achieving a F-1 score of 0.643 (regional) and 0.214 (exact) in chapter break prediction on a validation set of linear novels with existing chapter structures. In addition, we observe notable differences between the distributions of predicted chapter lengths in linear and non-linear fictional narratives, with the latter exhibiting significantly greater variability. Our results hold significance for narrative researchers appraising methods for extracting plots from non-linear novels.
2022
pdf
bib
abs
The COVID That Wasn’t: Counterfactual Journalism Using GPT
Sil Hamilton
|
Andrew Piper
Proceedings of the 6th Joint SIGHUM Workshop on Computational Linguistics for Cultural Heritage, Social Sciences, Humanities and Literature
In this paper, we explore the use of large language models to assess human interpretations of real world events. To do so, we use a language model trained prior to 2020 to artificially generate news articles concerning COVID-19 given the headlines of actual articles written during the pandemic. We then compare stylistic qualities of our artificially generated corpus with a news corpus, in this case 5,082 articles produced by CBC News between January 23 and May 5, 2020. We find our artificially generated articles exhibits a considerably more negative attitude towards COVID and a significantly lower reliance on geopolitical framing. Our methods and results hold importance for researchers seeking to simulate large scale cultural processes via recent breakthroughs in text generation.