Revealing Redundant Syntax in Large Language Models through Multi-Hop Dependency Paths
Masaki Sashida, Takeshi Kojima, Yusuke Iwasawa, Yutaka Matsuo
Abstract
Prior work on attention–syntax alignment has largely focused on single-hop Universal Dependency edges (DPs). In this paper, we treat short multi-hop dependency paths (MDPs) (e.g., “obl+case”) as first-class units and analyze them alongside DPs. Across three pretrained autoregressive LMs (GPT-2 XL, Llama 3 8B, Qwen3-8B) and one encoder baseline (BERT-large), we extract 2–3 hop MDPs from UD-parsed English and quantify head–relation alignment with an Unlabeled Attachment Score (UAS)–style metric modified for causal masking in decoder-only models. Rank visualizations reveal both overlap and specialization: we observe heads that align with both DPs and MDPs, as well as heads that appear specialized for one route. To test functional relevance, we first identify heads by UAS and then apply an undifferentiated (uniform) attention ablation to those heads; we evaluate the impact on BLiMP and LAMBADA. Ablating the top 10% of all heads shows that MDP-selected heads induce larger drops than DP-selected heads and that the union (“Mix”) of DP- and MDP-selected heads yields the largest drops. For GPT-2 XL, the observed drops are (BLiMP: 𝛥DP = 1.35 pp, 𝛥MDP = 4.81 pp, 𝛥Mix = 7.11 pp; LAMBADA: 𝛥DP = 4.70 pp, 𝛥MDP = 25.17 pp, 𝛥Mix = 32.99 pp), all exceeding size-matched random controls. These results indicate that models can route information consistent with syntactic dependencies via both DP and MDP pathways, with MDPs playing a distinct and measurable role in some settings under our interventions.- Anthology ID:
- 2026.findings-eacl.214
- Volume:
- Findings of the Association for Computational Linguistics: EACL 2026
- Month:
- March
- Year:
- 2026
- Address:
- Rabat, Morocco
- Editors:
- Vera Demberg, Kentaro Inui, Lluís Marquez
- Venue:
- Findings
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 4114–4137
- Language:
- URL:
- https://preview.aclanthology.org/ingest-eacl/2026.findings-eacl.214/
- DOI:
- Cite (ACL):
- Masaki Sashida, Takeshi Kojima, Yusuke Iwasawa, and Yutaka Matsuo. 2026. Revealing Redundant Syntax in Large Language Models through Multi-Hop Dependency Paths. In Findings of the Association for Computational Linguistics: EACL 2026, pages 4114–4137, Rabat, Morocco. Association for Computational Linguistics.
- Cite (Informal):
- Revealing Redundant Syntax in Large Language Models through Multi-Hop Dependency Paths (Sashida et al., Findings 2026)
- PDF:
- https://preview.aclanthology.org/ingest-eacl/2026.findings-eacl.214.pdf