@inproceedings{saeidi-etal-2025-insights,
title = "Insights into Alignment: Evaluating {DPO} and its Variants Across Multiple Tasks",
author = "Saeidi, Amir and
Verma, Shivanshu and
Uddin, Md Nayem and
Baral, Chitta",
editor = "Zhao, Jin and
Wang, Mingyang and
Liu, Zhu",
booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop)",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/landing_page/2025.acl-srw.26/",
pages = "409--421",
ISBN = "979-8-89176-254-1",
abstract = "This study evaluates Direct Preference Optimization (DPO) and its variants for aligning Large Language Models (LLMs) with human preferences, testing three configurations: (1) with Supervised Fine-Tuning (SFT), (2) without SFT, and (3) without SFT but using an instruction-tuned model. We further investigate how training set size influences model performance. Our evaluation spans 13 benchmarks{---}covering dialogue, reasoning, mathematical problem-solving, question answering, truthfulness, MT-Bench, Big Bench, and the Open LLM Leaderboard. We find that: (1) alignment methods often achieve near-optimal performance even with smaller subsets of training data; (2) although they offer limited improvements on complex reasoning tasks, they enhance mathematical problem-solving; and (3) using an instruction-tuned model improves truthfulness. These insights highlight the conditions under which alignment methods excel, as well as their limitations."
}
Markdown (Informal)
[Insights into Alignment: Evaluating DPO and its Variants Across Multiple Tasks](https://preview.aclanthology.org/landing_page/2025.acl-srw.26/) (Saeidi et al., ACL 2025)
ACL