@inproceedings{kamzela-etal-2025-llm,
    title = "You are an {LLM} teaching a smaller model everything you know: Multi-task pretraining of language models with {LLM}-designed study plans",
    author = "Kamzela, Wiktor  and
      Lango, Mateusz  and
      Dusek, Ondrej",
    editor = "Charpentier, Lucas  and
      Choshen, Leshem  and
      Cotterell, Ryan  and
      Gul, Mustafa Omer  and
      Hu, Michael Y.  and
      Liu, Jing  and
      Jumelet, Jaap  and
      Linzen, Tal  and
      Mueller, Aaron  and
      Ross, Candace  and
      Shah, Raj Sanjay  and
      Warstadt, Alex  and
      Wilcox, Ethan Gotlieb  and
      Williams, Adina",
    booktitle = "Proceedings of the First BabyLM Workshop",
    month = nov,
    year = "2025",
    address = "Suzhou, China",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/2025.babylm-main.33/",
    pages = "469--487",
    ISBN = "TODO",
    abstract = "This paper proposes a multi-task pre-training of language models without any text corpora.The method leverages an existing Large Language Model (LLM) to generate a diverse corpus containing training data for 56 automatically designed tasks and uses generated labels to enhance the training signal.The method does not rely on hidden states or even output distributions of the teacher model, so may be employed in scenarios when the teacher LLM is available only through an API.The conducted experiments show that models trained on the proposed synthetic corpora achieve competitive or superior performance compared to those trained on same-sized human-written texts."
}Markdown (Informal)
[You are an LLM teaching a smaller model everything you know: Multi-task pretraining of language models with LLM-designed study plans](https://preview.aclanthology.org/ingest-emnlp/2025.babylm-main.33/) (Kamzela et al., BabyLM 2025)
ACL