@inproceedings{li-etal-2025-alleviating,
title = "Alleviating Hallucinations in Large Language Models via Truthfulness-driven Rank-adaptive {L}o{RA}",
author = "Li, Jiahao and
Mao, Zhendong and
Wang, Quan",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2025",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/display_plenaries/2025.findings-acl.103/",
pages = "2020--2031",
ISBN = "979-8-89176-256-5",
abstract = "Improving the truthfulness of LLMs to alleviate hallucinations has become critical for promoting the practical deployment of LLMs. Current fine-tuning-based methods ignore the intrinsic discrepancy in the truthfulness correlations across LLM internal modules, and instead treat them equally, which may potentially decrease the performance of truthfulness improvement. In this paper, we propose a truthfulness-driven rank-adaptive LoRA method to improve LLM truthfulness (RaLFiT), which adaptively allocates the ranks in LoRA training according to the truthfulness correlations of modules within LLM. Specifically, it first measures the truthfulness correlation of each LLM module by a probing process, and allocates higher ranks to strongly correlated modules, which means a larger update subspace during training. Experimental results on TruthfulQA show that RaLFiT consistently outperforms previous state-of-the-art methods across the Llama LLM family, verifying its effectiveness and superiority, and for the first time makes the performance of 7B Llama LLMs exceed GPT-4."
}
Markdown (Informal)
[Alleviating Hallucinations in Large Language Models via Truthfulness-driven Rank-adaptive LoRA](https://preview.aclanthology.org/display_plenaries/2025.findings-acl.103/) (Li et al., Findings 2025)
ACL