@inproceedings{lattimer-etal-2023-fast,
title = "Fast and Accurate Factual Inconsistency Detection Over Long Documents",
author = "Lattimer, Barrett and
Chen, Patrick H. and
Zhang, Xinyuan and
Yang, Yi",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2023.emnlp-main.105/",
doi = "10.18653/v1/2023.emnlp-main.105",
pages = "1691--1703",
abstract = "Generative AI models exhibit remarkable potential; however, hallucinations across various tasks present a significant challenge, particularly for longer inputs that current approaches struggle to address effectively. We introduce SCALE (Source Chunking Approach for Large-scale inconsistency Evaluation), a task-agnostic model for detecting factual inconsistencies using a novel chunking strategy. Specifically, SCALE is a Natural Language Inference (NLI) based model that uses large text chunks to condition over long texts. This approach achieves state-of-the-art performance in factual inconsistency detection for diverse tasks and long inputs. Additionally, we leverage the chunking mechanism and employ a novel algorithm to explain SCALE`s decisions through relevant source sentence retrieval. Our evaluations reveal that SCALE outperforms existing methods on both standard benchmarks and a new long-form dialogue dataset ScreenEval we constructed. Moreover, SCALE surpasses competitive systems in efficiency and model explanation evaluations. We have released our code and data publicly to GitHub."
}
Markdown (Informal)
[Fast and Accurate Factual Inconsistency Detection Over Long Documents](https://preview.aclanthology.org/add-emnlp-2024-awards/2023.emnlp-main.105/) (Lattimer et al., EMNLP 2023)
ACL