@inproceedings{lauriola-etal-2025-analyzing,
title = "Analyzing and Improving Coherence of Large Language Models in Question Answering",
author = "Lauriola, Ivano and
Campese, Stefano and
Moschitti, Alessandro",
editor = "Chiruzzo, Luis and
Ritter, Alan and
Wang, Lu",
booktitle = "Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)",
month = apr,
year = "2025",
address = "Albuquerque, New Mexico",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2025.naacl-long.588/",
pages = "11740--11755",
ISBN = "979-8-89176-189-6",
abstract = "Large language models (LLMs) have recently revolutionized natural language processing. These models, however, often suffer from instability or lack of coherence, that is the ability of the models to generate semantically equivalent outputs when receiving diverse yet semantically equivalent input variations. In this work, we analyze the behavior of multiple LLMs, including Mixtral-8x7B, Llama2-70b, Smaug-72b, and Phi-3, when dealing with multiple lexical variations of the same info-seeking questions. Our results suggest that various LLMs struggle to consistently answer diverse equivalent queries. To address this issue, we show how redundant information encoded as a prompt can increase the coherence of these models. In addition, we introduce a Retrieval-Augmented Generation (RAG) technique that supplements LLMs with the top-$k$ most similar questions from a question retrieval engine. This knowledge-augmentation leads to 4-8 percentage point improvement in end-to-end performance in factual question answering tasks. These findings underscore the need to enhance LLM stability and coherence through semantic awareness."
}
Markdown (Informal)
[Analyzing and Improving Coherence of Large Language Models in Question Answering](https://preview.aclanthology.org/fix-sig-urls/2025.naacl-long.588/) (Lauriola et al., NAACL 2025)
ACL
- Ivano Lauriola, Stefano Campese, and Alessandro Moschitti. 2025. Analyzing and Improving Coherence of Large Language Models in Question Answering. In Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), pages 11740–11755, Albuquerque, New Mexico. Association for Computational Linguistics.