Adam Elwood
2025
Small Encoders Can Rival Large Decoders in Detecting Groundedness
Istabrak Abbes
|
Gabriele Prato
|
Quentin Fournier
|
Fernando Rodriguez
|
Alaa Boukhary
|
Adam Elwood
|
Sarath Chandar
Findings of the Association for Computational Linguistics: ACL 2025
Augmenting large language models (LLMs) with external context significantly improves their performance in natural language processing (NLP) tasks. However, LLMs struggle to answer queries reliably when the provided context lacks information, often resorting to ungrounded speculation or internal knowledge. Groundedness – generating responses strictly supported by the context – is essential for ensuring factual consistency and trustworthiness. This study focuses on detecting whether a given query is grounded in a document provided in context before the costly answer generation by LLMs. Such a detection mechanism can significantly reduce both inference time and resource consumption. We show that lightweight, task-specific encoder models such as RoBERTa and NomicBERT, fine-tuned on curated datasets, can achieve accuracy comparable to state-of-the-art LLMs, such as Llama3 8B and GPT4o, in groundedness detection while reducing inference latency by orders of magnitude.
An LLM-Based Approach for Insight Generation in Data Analysis
Alberto Sánchez Pérez
|
Alaa Boukhary
|
Paolo Papotti
|
Luis Castejón Lozano
|
Adam Elwood
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Generating insightful and actionable information from databases is critical in data analysis. This paper introduces a novel approach using Large Language Models (LLMs) to automatically generate textual insights. Given a multi-table database as input, our method leverages LLMs to produce concise, text-based insights that reflect interesting patterns in the tables. Our framework includes a Hypothesis Generator to formulate domain-relevant questions, a Query Agent to answer such questions by generating SQL queries against a database, and a Summarization module to verbalize the insights. The insights are evaluated for both correctness and subjective insightfulness using a hybrid model of human judgment and automated metrics. Experimental results on public and enterprise databases demonstrate that our approach generates more insightful insights than other approaches while maintaining correctness.
Search
Fix author
Co-authors
- Alaa Boukhary 2
- Istabrak Abbes 1
- Sarath Chandar 1
- Quentin Fournier 1
- Luis Castejón Lozano 1
- show all...