@inproceedings{lan-etal-2025-phi,
title = "Phi: Preference Hijacking in Multi-modal Large Language Models at Inference Time",
author = "Lan, Yifan and
Cao, Yuanpu and
Zhang, Weitong and
Lin, Lu and
Chen, Jinghui",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.901/",
pages = "17851--17876",
ISBN = "979-8-89176-332-6",
abstract = "Recently, Multimodal Large Language Models (MLLMs) have gained significant attention across various domains. However, their widespread adoption has also raised serious safety concerns.In this paper, we uncover a new safety risk of MLLMs: the output preference of MLLMs can be arbitrarily manipulated by carefully optimized images. Such attacks often generate contextually relevant yet biased responses that are neither overtly harmful nor unethical, making them difficult to detect. Specifically, we introduce a novel method, **P**reference **Hi**jacking (**Phi**), for manipulating the MLLM response preferences using a preference hijacked image. Our method works at inference time and requires no model modifications. Additionally, we introduce a universal hijacking perturbation {--} a transferable component that can be embedded into different images to hijack MLLM responses toward any attacker-specified preferences. Experimental results across various tasks demonstrate the effectiveness of our approach. The code for Phi is accessible at https://github.com/Yifan-Lan/Phi."
}Markdown (Informal)
[Phi: Preference Hijacking in Multi-modal Large Language Models at Inference Time](https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.901/) (Lan et al., EMNLP 2025)
ACL