REARANK: Reasoning Re-ranking Agent via Reinforcement Learning

Le Zhang, Bo Wang, Xipeng Qiu, Siva Reddy, Aishwarya Agrawal


Abstract
We present REARANK, a large language model (LLM)-based listwise reasoning rerank- ing agent. REARANK explicitly reasons be- fore reranking, significantly improving both performance and interpretability. Leveraging reinforcement learning and data augmentation, REARANK achieves substantial improvements over baseline models across popular informa- tion retrieval benchmarks, notably requiring only 179 annotated samples. Built on top of Qwen2.5-7B, our REARANK-7B demonstrates performance comparable to GPT-4 on both in- domain and out-of-domain benchmarks and even surpasses GPT-4 on reasoning-intensive BRIGHT benchmarks. These results under- score the effectiveness of our approach and highlight how reinforcement learning can en- hance LLM reasoning capabilities in reranking.
Anthology ID:
2025.emnlp-main.125
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
2458–2471
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.125/
DOI:
Bibkey:
Cite (ACL):
Le Zhang, Bo Wang, Xipeng Qiu, Siva Reddy, and Aishwarya Agrawal. 2025. REARANK: Reasoning Re-ranking Agent via Reinforcement Learning. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 2458–2471, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
REARANK: Reasoning Re-ranking Agent via Reinforcement Learning (Zhang et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.125.pdf
Checklist:
 2025.emnlp-main.125.checklist.pdf