@inproceedings{saritas-yildiz-2025-reproduction,
title = "A Reproduction Study: The Kernel {PCA} Interpretation of Self-Attention Fails Under Scrutiny",
author = "Sar{\i}ta{\c{s}}, Karahan and
Y{\i}ld{\i}z, {\c{C}}a{\u{g}}atay",
editor = "Zhao, Jin and
Wang, Mingyang and
Liu, Zhu",
booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop)",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/landing_page/2025.acl-srw.11/",
pages = "173--185",
ISBN = "979-8-89176-254-1",
abstract = "In this reproduction study, we revisit recent claims that self-attention implements kernel principal component analysis (KPCA) (Teo and Nguyen, 2024), positing that (i) value vectors $V$ capture the eigenvectors of the Gram matrix of the keys, and (ii) that self-attention projects queries onto the principal component axes of the key matrix $K$ in a feature space. Our analysis reveals three critical inconsistencies: (1) No alignment exists between learned self-attention value vectors and what is proposed in the KPCA perspective, with average similarity metrics (optimal cosine similarity $\leq 0.32$, linear CKA (Centered Kernel Alignment) $\leq 0.11$, kernel CKA $\leq 0.32$) indicating negligible correspondence; (2) Reported decreases in reconstruction loss $J_\text{proj}$, arguably justifying the claim that the self-attentionminimizes the projection error of KPCA, are misinterpreted, as the quantities involved differ by orders of magnitude ($\sim 10^3$); (3) Gram matrix eigenvalue statistics, introduced to justify that $V$ captures the eigenvector of the gram matrix, are irreproducible without undocumented implementation-specific adjustments. Across 10 transformer architectures, we conclude that the KPCA interpretation of self-attention lacks empirical support."
}
Markdown (Informal)
[A Reproduction Study: The Kernel PCA Interpretation of Self-Attention Fails Under Scrutiny](https://preview.aclanthology.org/landing_page/2025.acl-srw.11/) (Sarıtaş & Yıldız, ACL 2025)
ACL