Layered Insights: Generalizable Analysis of Human Authorial Style by Leveraging All Transformer Layers
Milad Alshomary, Nikhil Reddy Varimalla, Vishal Anand, Smaranda Muresan, Kathleen McKeown
Abstract
We propose a new approach for the authorship attribution task that leverages the various linguistic representations learned at different layers of pre-trained transformer-based models. We evaluate our approach on two popular authorship attribution models and three evaluation datasets, in in-domain and out-of-domain scenarios. We found that utilizing various transformer layers improves the robustness of authorship attribution models when tested on out-of-domain data, resulting in a much stronger performance. Our analysis gives further insights into how our model’s different layers get specialized in representing certain linguistic aspects that we believe benefit the model when tested out of the domain.- Anthology ID:
- 2025.emnlp-main.521
- Volume:
- Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
- Month:
- November
- Year:
- 2025
- Address:
- Suzhou, China
- Editors:
- Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
- Venue:
- EMNLP
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 10290–10303
- Language:
- URL:
- https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.521/
- DOI:
- Cite (ACL):
- Milad Alshomary, Nikhil Reddy Varimalla, Vishal Anand, Smaranda Muresan, and Kathleen McKeown. 2025. Layered Insights: Generalizable Analysis of Human Authorial Style by Leveraging All Transformer Layers. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 10290–10303, Suzhou, China. Association for Computational Linguistics.
- Cite (Informal):
- Layered Insights: Generalizable Analysis of Human Authorial Style by Leveraging All Transformer Layers (Alshomary et al., EMNLP 2025)
- PDF:
- https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.521.pdf