Layered Insights: Generalizable Analysis of Human Authorial Style by Leveraging All Transformer Layers

Milad Alshomary, Nikhil Reddy Varimalla, Vishal Anand, Smaranda Muresan, Kathleen McKeown


Abstract
We propose a new approach for the authorship attribution task that leverages the various linguistic representations learned at different layers of pre-trained transformer-based models. We evaluate our approach on two popular authorship attribution models and three evaluation datasets, in in-domain and out-of-domain scenarios. We found that utilizing various transformer layers improves the robustness of authorship attribution models when tested on out-of-domain data, resulting in a much stronger performance. Our analysis gives further insights into how our model’s different layers get specialized in representing certain linguistic aspects that we believe benefit the model when tested out of the domain.
Anthology ID:
2025.emnlp-main.521
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
10290–10303
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.521/
DOI:
Bibkey:
Cite (ACL):
Milad Alshomary, Nikhil Reddy Varimalla, Vishal Anand, Smaranda Muresan, and Kathleen McKeown. 2025. Layered Insights: Generalizable Analysis of Human Authorial Style by Leveraging All Transformer Layers. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 10290–10303, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Layered Insights: Generalizable Analysis of Human Authorial Style by Leveraging All Transformer Layers (Alshomary et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.521.pdf
Checklist:
 2025.emnlp-main.521.checklist.pdf