@inproceedings{weng-etal-2025-adversary,
title = "Adversary-Aware {DPO}: Enhancing Safety Alignment in Vision Language Models via Adversarial Training",
author = "Weng, Fenghua and
Lou, Jian and
Feng, Jun and
Huang, Minlie and
Wang, Wenjie",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.735/",
doi = "10.18653/v1/2025.findings-emnlp.735",
pages = "13644--13657",
ISBN = "979-8-89176-335-7",
abstract = "Safety alignment is critical in pre-trained large language models (LLMs) to generate responses aligned with human values and refuse harmful queries. Unlike LLM, the current safety alignment of VLMs is often achieved with post-hoc safety fine-tuning. However, these methods are less effective to white-box attacks. To address this, we propose $\textit{Adversary-aware DPO (ADPO)}$, a novel training framework that explicitly considers adversary. $\textit{Adversary-aware DPO (ADPO)}$ integrates adversarial training into DPO to enhance the safety alignment of VLMs under worst-case adversarial perturbations. $\textit{ADPO}$ introduces two key components: (1) an adversarial-trained reference model that generates human-preferred responses under worst-case perturbations, and (2) an adversary-aware DPO loss that generates winner-loser pairs accounting for adversarial distortions. By combining these innovations, $\textit{ADPO}$ ensures that VLMs remain robust and reliable even in the presence of sophisticated jailbreak attacks. Extensive experiments demonstrate that $\textit{ADPO}$ outperforms baselines in terms of both safety alignment and general utility of VLMs."
}Markdown (Informal)
[Adversary-Aware DPO: Enhancing Safety Alignment in Vision Language Models via Adversarial Training](https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.735/) (Weng et al., Findings 2025)
ACL