@inproceedings{xu-etal-2023-baize,
title = "Baize: An Open-Source Chat Model with Parameter-Efficient Tuning on Self-Chat Data",
author = "Xu, Canwen and
Guo, Daya and
Duan, Nan and
McAuley, Julian",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2023.emnlp-main.385/",
doi = "10.18653/v1/2023.emnlp-main.385",
pages = "6268--6278",
abstract = "Chat models, such as ChatGPT, have shown impressive capabilities and have been rapidly adopted across numerous domains. However, these models are only accessible through a restricted API, creating barriers for new research and progress in the field. We propose a pipeline that can automatically generate a high-quality multi-turn chat corpus by leveraging ChatGPT to engage in a conversation with itself. Subsequently, we employ parameter-efficient tuning to enhance LLaMA, an open-source large language model. The resulting model, named Baize, demonstrates good performance in multi-turn dialogues with guardrails that minimize potential risks. Additionally, we propose a new technique called Self-Distill with Feedback, to further improve the performance of the Baize models with feedback from ChatGPT."
}
Markdown (Informal)
[Baize: An Open-Source Chat Model with Parameter-Efficient Tuning on Self-Chat Data](https://preview.aclanthology.org/add-emnlp-2024-awards/2023.emnlp-main.385/) (Xu et al., EMNLP 2023)
ACL