@inproceedings{wu-etal-2024-synchronous,
title = "Synchronous Faithfulness Monitoring for Trustworthy Retrieval-Augmented Generation",
author = "Wu, Di and
Gu, Jia-Chen and
Yin, Fan and
Peng, Nanyun and
Chang, Kai-Wei",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2024.emnlp-main.527/",
doi = "10.18653/v1/2024.emnlp-main.527",
pages = "9390--9406",
abstract = "Retrieval-augmented language models (RALMs) have shown strong performance and wide applicability in knowledge-intensive tasks. However, there are significant trustworthiness concerns as RALMs are prone to generating unfaithful outputs, including baseless information or contradictions with the retrieved context. This paper proposes SynCheck, a lightweight monitor that leverages fine-grained decoding dynamics including sequence likelihood, uncertainty quantification, context influence, and semantic alignment to synchronously detect unfaithful sentences. By integrating efficiently measurable and complementary signals, SynCheck enables accurate and immediate feedback and intervention. Experiments show that SynCheck significantly outperforms existing faithfulness detection baselines, achieving over 0.85 AUROC across a suite of six long-form retrieval-augmented generation tasks. Leveraging SynCheck, we further introduce FOD, a faithfulness-oriented decoding algorithm guided by beam search for long-form retrieval-augmented generation. Empirical results demonstrate that FOD outperforms traditional strategies such as abstention, reranking, or contrastive decoding significantly in terms of faithfulness, achieving over 10{\%} improvement across six datasets."
}
Markdown (Informal)
[Synchronous Faithfulness Monitoring for Trustworthy Retrieval-Augmented Generation](https://preview.aclanthology.org/add-emnlp-2024-awards/2024.emnlp-main.527/) (Wu et al., EMNLP 2024)
ACL