@inproceedings{chang-etal-2024-parts,
title = "When Parts Are Greater Than Sums: Individual {LLM} Components Can Outperform Full Models",
author = "Chang, Ting-Yun and
Thomason, Jesse and
Jia, Robin",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.emnlp-main.574/",
doi = "10.18653/v1/2024.emnlp-main.574",
pages = "10280--10299",
abstract = "This paper studies in-context learning by decomposing the output of large language models into the individual contributions of attention heads and MLPs (components). We observe curious components: good-performing ones that individually do well on a classification task, even when the model performs poorly; bad-performing ones that do much worse than chance; and label-biased components that always predict the same label. We find that component accuracies are well-correlated across different demonstration sets and perturbations of prompt templates. Based on our findings, we propose component reweighting, which learns to linearly re-scale the component activations from a few labeled examples. Given 24 labeled examples, our method improves by an average of 6.0{\%} accuracy points over 24-shot ICL across 8 tasks on Llama-2-7B. Overall, this paper both enriches our understanding of ICL and provides a practical method for improvement by examining model internals."
}
Markdown (Informal)
[When Parts Are Greater Than Sums: Individual LLM Components Can Outperform Full Models](https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.emnlp-main.574/) (Chang et al., EMNLP 2024)
ACL