@inproceedings{zhou-etal-2023-enhancing-generative,
title = "Enhancing Generative Retrieval with Reinforcement Learning from Relevance Feedback",
author = "Zhou, Yujia and
Dou, Zhicheng and
Wen, Ji-Rong",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2023.emnlp-main.768/",
doi = "10.18653/v1/2023.emnlp-main.768",
pages = "12481--12490",
abstract = "The recent advent of end-to-end generative retrieval marks a significant shift in document retrieval methods, leveraging differentiable search indexes to directly produce relevant document identifiers (docids) in response to a specific query. Nevertheless, this approach faces two fundamental challenges: (i) a discrepancy between the token-level probabilistic optimization and the broader document-level relevance estimation; (ii) an overemphasis on top-1 results at the expense of overall ranking quality. To tackle these challenges, we propose a generative retrieval model with reinforcement learning from relevance feedback, which aims to align token-level docid generation with document-level relevance estimation. The training process incorporates three stages: supervised fine-tuning, relevance reward model training, and reinforced learning-to-rank from relevance feedback. To train a high-quality reward model, we define {\textquotedblleft}relevance{\textquotedblright} under three progressive scenarios, which collectively offer a comprehensive evaluation of the document relevance. Experiments conducted on two benchmark datasets demonstrate the effectiveness of our proposed approach."
}
Markdown (Informal)
[Enhancing Generative Retrieval with Reinforcement Learning from Relevance Feedback](https://preview.aclanthology.org/add-emnlp-2024-awards/2023.emnlp-main.768/) (Zhou et al., EMNLP 2023)
ACL