@inproceedings{nguyen-etal-2026-llms,
title = "How Do {LLM}s Generate Contrastive Sentiments? A Mechanistic Perspective",
author = {Nguyen, Van Bach and
Schl{\"o}tterer, J{\"o}rg and
Seifert, Christin},
editor = "Demberg, Vera and
Inui, Kentaro and
Marquez, Llu{\'i}s",
booktitle = "Proceedings of the 19th Conference of the {E}uropean Chapter of the {A}ssociation for {C}omputational {L}inguistics (Volume 1: Long Papers)",
month = mar,
year = "2026",
address = "Rabat, Morocco",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-eacl/2026.eacl-long.311/",
pages = "6619--6635",
ISBN = "979-8-89176-380-7",
abstract = "This paper presents a mechanistic investigation of how large language models (LLMs) generate contrastive sentiments. We define this task as transforming the sentiment of a given text (e.g., from positive to negative) while making minimal changes to its content. We identify two core mechanisms: (1) a preservation mechanism that maintains the sentiment of the input text, primarily mediated by specific attention heads, and (2) a sentiment transformation mechanism, which integrates a representation of the target sentiment label with the original valenced words using a circuit containing both MLP and attention layers. Building on these findings, we propose and validate a novel mechanistic intervention. By modifying key attention heads, we steer the LLM toward more effective contrastive generation, increasing the sentiment flip rate without sacrificing the minimality of changes. Our work not only deepens the understanding of the mechanisms underlying contrastive sentiment generation in LLMs, but also introduces a promising new direction to steer LLM behavior via targeted, mechanistic interventions."
}Markdown (Informal)
[How Do LLMs Generate Contrastive Sentiments? A Mechanistic Perspective](https://preview.aclanthology.org/ingest-eacl/2026.eacl-long.311/) (Nguyen et al., EACL 2026)
ACL