@inproceedings{lee-etal-2025-small,
title = "Small Changes, Big Impact: How Manipulating a Few Neurons Can Drastically Alter {LLM} Aggression",
author = "Lee, Jaewook and
Jang, Junseo and
Kwon, Oh-Woog and
Kim, Harksoo",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingestion-acl-25/2025.acl-long.1144/",
pages = "23478--23505",
ISBN = "979-8-89176-251-0",
abstract = "Recent remarkable advances in Large Language Models (LLMs) have led to innovations in various domains such as education, healthcare, and finance, while also raising serious concerns that they can be easily misused for malicious purposes. Most previous research has focused primarily on observing how jailbreak attack techniques bypass safety mechanisms like Reinforcement Learning through Human Feedback (RLHF). However, whether there are neurons within LLMs that directly govern aggression has not been sufficiently investigated. To fill this gap, this study identifies specific neurons ({``}aggression neurons'') closely related to the expression of aggression and systematically analyzes how manipulating them affects the model{'}s overall aggression. Specifically, using a large-scale synthetic text corpus (aggressive and non-aggressive), we measure the activation frequency of each neuron, then apply masking and activation techniques to quantitatively evaluate changes in aggression by layer and by manipulation ratio. Experimental results show that, in all models, manipulating only a small number of neurons can increase aggression by up to 33{\%}, and the effect is even more extreme when aggression neurons are concentrated in certain layers. Moreover, even models of the same scale exhibit nonlinear changes in aggression patterns, suggesting that simple external safety measures alone may not be sufficient for complete defense."
}
Markdown (Informal)
[Small Changes, Big Impact: How Manipulating a Few Neurons Can Drastically Alter LLM Aggression](https://preview.aclanthology.org/ingestion-acl-25/2025.acl-long.1144/) (Lee et al., ACL 2025)
ACL