@inproceedings{li-etal-2025-know,
title = "Know the Unknown: An Uncertainty-Sensitive Method for {LLM} Instruction Tuning",
author = "Li, Jiaqi and
Tang, Yixuan and
Yang, Yi",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2025",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/display_plenaries/2025.findings-acl.153/",
pages = "2972--2989",
ISBN = "979-8-89176-256-5",
abstract = "Large language models (LLMs) demonstrate remarkable capabilities but face challenges from hallucinations, which typically arise from insufficient knowledge or context. While instructing LLMs to acknowledge knowledge limitations by responding with ``I don{'}t know'' appears promising, we find that models consistently struggle with admitting knowledge gaps. This challenge may originate from current instruction datasets that emphasise answer generation over knowledge boundary awareness. To address this limitation, we introduce **U**ncertainty-and-**S**ensitivity-Aware Tuning **(US-Tuning)**, a novel two-stage approach for contextual question answering (QA). The first stage enhances LLMs' ability to recognise their knowledge boundaries, while the second stage reinforces instruction adherence through carefully designed causal prompts. Our experimental results demonstrate that US-Tuning not only significantly reduces incorrect answers in contextual QA but also improves models' faithfulness to their parametric knowledge, mitigating hallucinations in general QA tasks. Our fine-tuned Llama2-7B model achieves up to a 34.7{\%} improvement in handling out-of-knowledge questions and outperforms GPT-4 by 4.2{\%} in overall performance."
}
Markdown (Informal)
[Know the Unknown: An Uncertainty-Sensitive Method for LLM Instruction Tuning](https://preview.aclanthology.org/display_plenaries/2025.findings-acl.153/) (Li et al., Findings 2025)
ACL