@inproceedings{chen-etal-2024-xplainllm,
title = "{X}plain{LLM}: A Knowledge-Augmented Dataset for Reliable Grounded Explanations in {LLM}s",
author = "Chen, Zichen and
Chen, Jianda and
Singh, Ambuj and
Sra, Misha",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.emnlp-main.432/",
doi = "10.18653/v1/2024.emnlp-main.432",
pages = "7578--7596",
abstract = "Large Language Models (LLMs) have achieved remarkable success in natural language tasks, yet understanding their reasoning processes remains a significant challenge. We address this by introducing XplainLLM, a dataset accompanying an explanation framework designed to enhance LLM transparency and reliability. Our dataset comprises 24,204 instances where each instance interprets the LLM`s reasoning behavior using knowledge graphs (KGs) and graph attention networks (GAT), and includes explanations of LLMs such as the decoder-only Llama-3 and the encoder-only RoBERTa. XplainLLM also features a framework for generating grounded explanations and the \textit{debugger-scores} for multidimensional quality analysis. Our explanations include \textit{why-choose} and \textit{why-not-choose} components, \textit{reason-elements}, and \textit{debugger-scores} that collectively illuminate the LLM`s reasoning behavior. Our evaluations demonstrate XplainLLM`s potential to reduce hallucinations and improve grounded explanation generation in LLMs. XplainLLM is a resource for researchers and practitioners to build trust and verify the reliability of LLM outputs. Our code and dataset are publicly available."
}
Markdown (Informal)
[XplainLLM: A Knowledge-Augmented Dataset for Reliable Grounded Explanations in LLMs](https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.emnlp-main.432/) (Chen et al., EMNLP 2024)
ACL