@inproceedings{lu-etal-2025-x,
title = "{X}-Boundary: Establishing Exact Safety Boundary to Shield {LLM}s from Jailbreak Attacks without Compromising Usability",
author = "Lu, Xiaoya and
Liu, Dongrui and
Yu, Yi and
Xu, Luxin and
Shao, Jing",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.282/",
doi = "10.18653/v1/2025.findings-emnlp.282",
pages = "5247--5272",
ISBN = "979-8-89176-335-7",
abstract = "With the widespread application of large language models (LLMs) across various domains, techniques for enhancing their security have progressed rapidly. In this paper, we reveal that although existing defense methods can improve the robustness of LLMs against jailbreaks, they compromise usability, i.e., reducing general capabilities or causing the over-refusal problem. From the perspective of LLM mechanism interpretability, we discover that these methods fail to establish a boundary that exactly distinguishes safe and harmful feature representations. Therefore, boundary-safe representations close to harmful representations are inevitably disrupted, leading to a decline in usability. To address this issue, we propose X-Boundary to push harmful representations away from boundary-safe representations and obtain an exact distinction boundary. In this way, harmful representations can be precisely erased without disrupting safe ones. Experimental results show that X-Boundary achieves state-of-the-art defense performance against both single-turn and multi-turn jailbreak attacks, while reducing the over-refusal rate by about 20{\%} and maintaining nearly complete general capability. Furthermore, we theoretically prove and empirically verify that X-Boundary can accelerate the convergence process during training."
}Markdown (Informal)
[X-Boundary: Establishing Exact Safety Boundary to Shield LLMs from Jailbreak Attacks without Compromising Usability](https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.282/) (Lu et al., Findings 2025)
ACL