@inproceedings{zhang-etal-2025-finite,
title = "Finite State Automata Inside Transformers with Chain-of-Thought: A Mechanistic Study on State Tracking",
author = "Zhang, Yifan and
Du, Wenyu and
Jin, Dongming and
Fu, Jie and
Jin, Zhi",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/landing_page/2025.acl-long.668/",
pages = "13603--13621",
ISBN = "979-8-89176-251-0",
abstract = "Chain-of-thought (CoT) significantly enhances the performance of large language models (LLMs) across a wide range of tasks, and prior research shows that CoT can theoretically increase expressiveness. However, there is limited mechanistic understanding of the algorithms that Transformer+CoT can learn. Our key contributions are: (1) We evaluate the state tracking capabilities of Transformer+CoT and its variants, confirming the effectiveness of CoT. (2) Next, we identify the circuit (a subset of model components, responsible for tracking the world state), indicating that late-layer MLP neurons play a key role. We propose two metrics, compression and distinction, and show that the neuron sets for each state achieve nearly 100{\%} accuracy, providing evidence of an implicit finite state automaton (FSA) embedded within the model. (3) Additionally, we explore three challenging settings: skipping intermediate steps, introducing data noises, and testing length generalization. Our results demonstrate that Transformer+CoT learns robust algorithms (FSAs), highlighting its resilience in challenging scenarios. Our code is available at https://github.com/IvanChangPKU/FSA."
}
Markdown (Informal)
[Finite State Automata Inside Transformers with Chain-of-Thought: A Mechanistic Study on State Tracking](https://preview.aclanthology.org/landing_page/2025.acl-long.668/) (Zhang et al., ACL 2025)
ACL