@inproceedings{deng-etal-2024-dont,
    title = "Don{'}t Just Say ``{I} don{'}t know''! Self-aligning Large Language Models for Responding to Unknown Questions with Explanations",
    author = "Deng, Yang  and
      Zhao, Yong  and
      Li, Moxin  and
      Ng, See-Kiong  and
      Chua, Tat-Seng",
    editor = "Al-Onaizan, Yaser  and
      Bansal, Mohit  and
      Chen, Yun-Nung",
    booktitle = "Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing",
    month = nov,
    year = "2024",
    address = "Miami, Florida, USA",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/2024.emnlp-main.757/",
    doi = "10.18653/v1/2024.emnlp-main.757",
    pages = "13652--13673",
    abstract = "Despite the remarkable abilities of Large Language Models (LLMs) to answer questions, they often display a considerable level of overconfidence even when the question does not have a definitive answer. To avoid providing hallucinated answers to these unknown questions, existing studies typically investigate approaches to refusing to answer these questions. In this work, we propose a novel and scalable self-alignment method to utilize the LLM itself to enhance its response-ability to different types of unknown questions, being capable of not just refusing to answer but further proactively providing explanations to the unanswerability of unknown questions. Specifically, the Self-Align method first employ a two-stage class-aware self-augmentation approach to generate a large amount of unknown question-response data. Then we conduct disparity-driven self-curation to select qualified data for fine-tuning the LLM itself for aligning the responses to unknown questions as desired. Experimental results on two datasets across four types of unknown questions validate the superiority of the Self-Aligned method over existing baselines in terms of three types of task formulation."
}Markdown (Informal)
[Don’t Just Say “I don’t know”! Self-aligning Large Language Models for Responding to Unknown Questions with Explanations](https://preview.aclanthology.org/ingest-emnlp/2024.emnlp-main.757/) (Deng et al., EMNLP 2024)
ACL