@inproceedings{zhang-etal-2025-persona,
title = "Persona-judge: Personalized Alignment of Large Language Models via Token-level Self-judgment",
author = "Zhang, Xiaotian and
Chen, Ruizhe and
Feng, Yang and
Liu, Zuozhu",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2025",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/acl25-workshop-ingestion/2025.findings-acl.260/",
pages = "5037--5049",
ISBN = "979-8-89176-256-5",
abstract = "Aligning language models with human preferences presents significant challenges, particularly in achieving personalization without incurring excessive computational costs. Existing methods rely on reward signals and additional annotated data, limiting their scalability and adaptability to diverse human values. To address these challenges, we introduce Persona-judge, a novel discriminative paradigm that enables training-free personalized alignment with unseen preferences. Instead of optimizing policy parameters through external reward feedback, Persona-judge leverages the intrinsic preference judgment capabilities of the model. Specifically, a draft model generates candidate tokens conditioned on a given preference, while a judge model, embodying another preference, cross-validates the predicted tokens whether to be accepted. Experimental results demonstrate that Persona-judge, using the inherent preference evaluation mechanisms of the model, offers a scalable and computationally efficient solution to personalized alignment, paving the way for more adaptive customized alignment. Our code is available here."
}
Markdown (Informal)
[Persona-judge: Personalized Alignment of Large Language Models via Token-level Self-judgment](https://preview.aclanthology.org/acl25-workshop-ingestion/2025.findings-acl.260/) (Zhang et al., Findings 2025)
ACL