@inproceedings{li-caragea-2023-distilling,
title = "Distilling Calibrated Knowledge for Stance Detection",
author = "Li, Yingjie and
Caragea, Cornelia",
editor = "Rogers, Anna and
Boyd-Graber, Jordan and
Okazaki, Naoaki",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2023",
month = jul,
year = "2023",
address = "Toronto, Canada",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.findings-acl.393/",
doi = "10.18653/v1/2023.findings-acl.393",
pages = "6316--6329",
abstract = "Stance detection aims to determine the position of an author toward a target and provides insights into people`s views on controversial topics such as marijuana legalization. Despite recent progress in this task, most existing approaches use hard labels (one-hot vectors) during training, which ignores meaningful signals among categories offered by soft labels. In this work, we explore knowledge distillation for stance detection and present a comprehensive analysis. Our contributions are: 1) we propose to use knowledge distillation over multiple generations in which a student is taken as a new teacher to transfer knowledge to a new fresh student; 2) we propose a novel dynamic temperature scaling for knowledge distillation to calibrate teacher predictions in each generation step. Extensive results on three stance detection datasets show that knowledge distillation benefits stance detection and a teacher is able to transfer knowledge to a student more smoothly via calibrated guiding signals. We publicly release our code to facilitate future research."
}
Markdown (Informal)
[Distilling Calibrated Knowledge for Stance Detection](https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.findings-acl.393/) (Li & Caragea, Findings 2023)
ACL