@inproceedings{liu-etal-2025-focalpo,
title = "{F}ocal{PO}: Enhancing Preference Optimizing by Focusing on Correct Preference Rankings",
author = "Liu, Tong and
Yu, Xiao and
Zhou, Wenxuan and
Gu, Jindong and
Tresp, Volker",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingestion-acl-25/2025.acl-short.21/",
pages = "256--267",
ISBN = "979-8-89176-252-7",
abstract = "Efficient preference optimization algorithms such as Direct Preference Optimization (DPO) have become a popular approach in aligning large language models (LLMs) with human preferences. These algorithms implicitly treat the LLM as a reward model, and focus on training it to correct misranked preference pairs. However, recent work (CITATION) empirically finds that DPO training \textit{rarely improves these misranked preference pairs}, despite its gradient emphasizing on these cases. We introduce FocalPO, a DPO variant that instead \textit{down-weighs} misranked preference pairs and prioritizes enhancing the model{'}s understanding of pairs that it can already rank correctly. Inspired by Focal Loss used in vision tasks, FocalPO achieves this by adding a modulating factor to dynamically scale DPO loss. Our experiment demonstrates that FocalPO surpasses DPO and its variants on popular benchmarks like Alpaca Eval 2.0 and Arena-Hard using Mistral-Base-7B and Llama-3-Instruct-8B, with the introduced hyperparameter fixed. Additionally, we empirically reveals how FocalPO affects training on correct and incorrect sample groups, further underscoring its effectiveness."
}
Markdown (Informal)
[FocalPO: Enhancing Preference Optimizing by Focusing on Correct Preference Rankings](https://preview.aclanthology.org/ingestion-acl-25/2025.acl-short.21/) (Liu et al., ACL 2025)
ACL