@inproceedings{maghakian-etal-2025-embedding,
title = "Embedding-Free {RAG}",
author = "Maghakian, Jessica and
Sinha, Raunak and
Schettewi, Max and
Kaur, Gunkirat",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/name-variant-enfa-fane/2025.findings-emnlp.1360/",
doi = "10.18653/v1/2025.findings-emnlp.1360",
pages = "24974--24985",
ISBN = "979-8-89176-335-7",
abstract = "Retrieval-Augmented Generation (RAG) is the current state-of-the-art method for mitigating the shortcomings of large language models (LLMs) by incorporating external knowledge sources to provide more relevant and accurate responses to user queries. However building performant RAG systems for real use-cases typically requires heavy investment from NLP experts, such as fine-tuning embedding models for specialized domains, experimenting with text chunking strategies and other niche hyperparameter tunings. We propose Embedding-Free RAG, a model-agnostic approach that enables the deployment of a one-size-fits-all RAG pipeline for user-provided grounding documents. Unlike traditional RAG, which relies on embedding models for information retrieval, Embedding-Free RAG leverages the generalized reasoning abilities of LLMs in a novel algorithmic framework during the retrieval stage. Extensive experiments demonstrate that Embedding-Free RAG outperforms existing state-of-the-art methods, achieving up to 4.6x higher F1 scores and up to 2x better question answering accuracy across a wide range of challenging domains."
}Markdown (Informal)
[Embedding-Free RAG](https://preview.aclanthology.org/name-variant-enfa-fane/2025.findings-emnlp.1360/) (Maghakian et al., Findings 2025)
ACL
- Jessica Maghakian, Raunak Sinha, Max Schettewi, and Gunkirat Kaur. 2025. Embedding-Free RAG. In Findings of the Association for Computational Linguistics: EMNLP 2025, pages 24974–24985, Suzhou, China. Association for Computational Linguistics.