Who Taught You That? Tracing Teachers in Model Distillation

Somin Wadhwa, Chantal Shaib, Silvio Amir, Byron C Wallace


Abstract
Model distillation – using outputs from a large teacher model to teach a small student model – is a practical means of creating efficient models for a particular task. We ask: Can we identify a students’ teacher based on its outputs? Such “footprints” left by teacher LLMs would be interesting artifacts. Beyond this, reliable teacher inference may have practical implications as actors seek to distill specific capabilities of massive proprietary LLMs into deployed smaller LMs, potentially violating terms of service. We consider practical task distillation targets including summarization, question answering, and instruction-following. We assume a finite set of candidate teacher models, which we treat as blackboxes. We design discriminative models that operate over lexical features. We find that n-gram similarity alone is unreliable for identifying teachers, but part-of-speech (PoS) templates preferred by student models mimic those of their teachers.
Anthology ID:
2025.findings-acl.173
Volume:
Findings of the Association for Computational Linguistics: ACL 2025
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Wanxiang Che, Joyce Nabende, Ekaterina Shutova, Mohammad Taher Pilehvar
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
3307–3315
Language:
URL:
https://preview.aclanthology.org/mtsummit-25-ingestion/2025.findings-acl.173/
DOI:
10.18653/v1/2025.findings-acl.173
Bibkey:
Cite (ACL):
Somin Wadhwa, Chantal Shaib, Silvio Amir, and Byron C Wallace. 2025. Who Taught You That? Tracing Teachers in Model Distillation. In Findings of the Association for Computational Linguistics: ACL 2025, pages 3307–3315, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Who Taught You That? Tracing Teachers in Model Distillation (Wadhwa et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/mtsummit-25-ingestion/2025.findings-acl.173.pdf