@inproceedings{gu-etal-2025-neuronmerge,
title = "{N}euron{M}erge: Merging Models via Functional Neuron Groups",
author = "Gu, Wangyun and
Gao, Qianghua and
Li-Xin, Zhang and
Shen, Xu and
Ye, Jieping",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2025",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/display_plenaries/2025.findings-acl.471/",
pages = "9015--9037",
ISBN = "979-8-89176-256-5",
abstract = "Model merging techniques like task arithmetic, which combines model parameters through weighted averaging, have proven effective. However, the success of task arithmetic relies on the linearity between model weight differences and output feature changes, which is often lacking in conventional fine-tuned models. In this work, we employ neuron description methods to analyze and classify neurons based on their functionalities. We theoretically demonstrate that grouping Multi-Layer Perceptron (MLP) neurons by functionality enhances model linearity. Building on this, we propose a neuron-based task arithmetic merging method that consistently improves performance across various tasks and model scales. Our approach is complementary to existing merging techniques, achieving superior results in merging models fine-tuned on fundamental tasks like Math, Code and Translation."
}
Markdown (Informal)
[NeuronMerge: Merging Models via Functional Neuron Groups](https://preview.aclanthology.org/display_plenaries/2025.findings-acl.471/) (Gu et al., Findings 2025)
ACL