A Comparison of Independent and Joint Fine-tuning Strategies for Retrieval-Augmented Generation

Neal Gregory Lawton, Alfy Samuel, Anoop Kumar, Daben Liu


Abstract
Retrieval augmented generation (RAG) is a popular framework for question answering that is powered by two large language models (LLMs): an embedding model that retrieves context documents from a database that are relevant to a given question, and a generator model that uses the retrieved context to generate an answer to the question. Both the embedding and generator models can be fine-tuned to increase performance of a RAG pipeline on a new task, but multiple fine-tuning strategies exist with different costs and benefits. In this paper, we evaluate and compare several RAG fine-tuning strategies, including independent, joint, and two-phase fine-tuning. In our experiments, we observe that all of these strategies achieve about equal improvement in EM and F1 generation quality metrics, although they have significantly different computational costs. We conclude the optimal fine-tuning strategy to use depends on whether the training dataset includes context labels and whether a grid search over the learning rates for the embedding and generator models is required.
Anthology ID:
2025.findings-emnlp.1247
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2025
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
22896–22904
Language:
URL:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.1247/
DOI:
10.18653/v1/2025.findings-emnlp.1247
Bibkey:
Cite (ACL):
Neal Gregory Lawton, Alfy Samuel, Anoop Kumar, and Daben Liu. 2025. A Comparison of Independent and Joint Fine-tuning Strategies for Retrieval-Augmented Generation. In Findings of the Association for Computational Linguistics: EMNLP 2025, pages 22896–22904, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
A Comparison of Independent and Joint Fine-tuning Strategies for Retrieval-Augmented Generation (Lawton et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.1247.pdf
Checklist:
 2025.findings-emnlp.1247.checklist.pdf