@inproceedings{huang-etal-2023-learning-preference,
title = "Learning Preference Model for {LLM}s via Automatic Preference Data Generation",
author = "Huang, Shijia and
Zhao, Jianqiao and
Li, Yanyang and
Wang, Liwei",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2023.emnlp-main.570/",
doi = "10.18653/v1/2023.emnlp-main.570",
pages = "9187--9199",
abstract = "Despite the advanced capacities of the state-of-the-art large language models (LLMs), they suffer from issues of hallucination, stereotype, etc. Preference models play an important role in LLM alignment, yet training preference models predominantly rely on human-annotated data. This reliance limits their versatility and scalability. In this paper, we propose learning the preference model for LLMs via automatic preference data generation (AutoPM). Our approach involves both In-Breadth Data Generation, which elicits pairwise preference data from LLMs following the helpful-honest-harmless (HHH) criteria, and In-Depth Data Generation, which enriches the dataset with responses spanning a wide quality range. With HHH-guided preference data, our approach simultaneously enables the LLMs to learn human preferences and align with human values. Quantitative assessments on five benchmark datasets demonstrate the reliability and potential of AutoPM, pointing out a more general and scalable way to improve LLM performance."
}
Markdown (Informal)
[Learning Preference Model for LLMs via Automatic Preference Data Generation](https://preview.aclanthology.org/add-emnlp-2024-awards/2023.emnlp-main.570/) (Huang et al., EMNLP 2023)
ACL