@inproceedings{shaozhen-etal-2024-choosy,
title = "Choosy Babies Need One Coach: Inducing Mode-Seeking Behavior in {B}aby{L}lama with Reverse {KL} Divergence",
author = "Shi, Shaozhen and
Matusevych, Yevgen and
Nissim, Malvina",
editor = "Hu, Michael Y. and
Mueller, Aaron and
Ross, Candace and
Williams, Adina and
Linzen, Tal and
Zhuang, Chengxu and
Choshen, Leshem and
Cotterell, Ryan and
Warstadt, Alex and
Wilcox, Ethan Gotlieb",
booktitle = "The 2nd BabyLM Challenge at the 28th Conference on Computational Natural Language Learning",
month = nov,
year = "2024",
address = "Miami, FL, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.conll-babylm.8/",
pages = "95--105",
abstract = "This study presents our submission to the Strict-Small Track of the 2nd BabyLM Challenge. We use a teacher-student distillation setup with the BabyLLaMa model (Timiryasov and Tastet, 2023) as a backbone. To make the student`s learning process more focused, we replace the objective function with a reverse Kullback-Leibler divergence, known to cause mode-seeking (rather than mode-averaging) behaviour in computational learners. We further experiment with having a single teacher (instead of an ensemble of two teachers) and implement additional optimization strategies to improve the distillation process. Our experiments show that under reverse KL divergence, a single-teacher model often outperforms or matches multiple-teacher models across most tasks. Additionally, incorporating advanced optimization techniques further enhances model performance, demonstrating the effectiveness and robustness of our proposed approach. These findings support our idea that {\textquotedblleft}choosy babies need one coach{\textquotedblright}."
}
Markdown (Informal)
[Choosy Babies Need One Coach: Inducing Mode-Seeking Behavior in BabyLlama with Reverse KL Divergence](https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.conll-babylm.8/) (Shi et al., CoNLL-BabyLM 2024)
ACL