@inproceedings{yan-2025-addition,
title = "Addition in Four Movements: Mapping Layer-wise Information Trajectories in {LLM}s",
author = "Yan, Yao",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.397/",
doi = "10.18653/v1/2025.findings-emnlp.397",
pages = "7518--7532",
ISBN = "979-8-89176-335-7",
abstract = "Arithmetic offers a compact test of whether large language models compute or memorize. We study multi-digit addition in LLaMA-3-8B-Instruct using linear probes and the Logit Lens, and find a consistent four-stage, layer-wise ordering of probe-decodable signal types across depth: (1) early layers encode formula structure (operand/operator layout) while the gold next token is still far from top-1; (2) mid layers expose digit-wise sums and carry indicators; (3) deeper layers express result-level numerical abstractions that support near-perfect digit decoding from hidden states; and (4) near the output, representations align with final sequence generation, with the correct next token reliably ranked first. Across experiments, each signal family becomes linearly decodable with high accuracy (stage-wise peaks typically $\geq$95{\%} on in-domain multi-digit addition, and up to 99{\%}). Taken together, these observations{---}\textit{in our setting}{---}are consistent with a hierarchical, computation-first account rather than rote pattern matching, and help explain why Logit Lens inspection is most informative mainly in later layers. Code and data are available at https://github.com/YaoToolChest/addition-in-four-movements.git."
}Markdown (Informal)
[Addition in Four Movements: Mapping Layer-wise Information Trajectories in LLMs](https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.397/) (Yan, Findings 2025)
ACL