CoRet: Improved Retriever for Code Editing
Fabio James Fehr, Prabhu Teja S, Luca Franceschi, Giovanni Zappella
Abstract
In this paper, we introduce CoRet, a dense retrieval model designed for code-editing tasks that integrates code semantics, repository structure, and call-graph dependencies. The model focuses on retrieving relevant portions of a code repository based on natural language queries such as requests to implement new features or fix bugs. These retrieved code chunks can then be presented to an user or to a second code-editing model or agent. To train CoRet, we propose a loss function explicitly designed for repository-level retrieval. On SWE-bench and Long Code Arena’s bug localisation datasets, we show that our model substantially improves retrieval recall by at least 15 percentage points over existing models, and ablate the design choices to show their importance in achieving these results.- Anthology ID:
- 2025.acl-short.62
- Volume:
- Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)
- Month:
- July
- Year:
- 2025
- Address:
- Vienna, Austria
- Editors:
- Wanxiang Che, Joyce Nabende, Ekaterina Shutova, Mohammad Taher Pilehvar
- Venue:
- ACL
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 775–789
- Language:
- URL:
- https://preview.aclanthology.org/landing_page/2025.acl-short.62/
- DOI:
- Cite (ACL):
- Fabio James Fehr, Prabhu Teja S, Luca Franceschi, and Giovanni Zappella. 2025. CoRet: Improved Retriever for Code Editing. In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 775–789, Vienna, Austria. Association for Computational Linguistics.
- Cite (Informal):
- CoRet: Improved Retriever for Code Editing (Fehr et al., ACL 2025)
- PDF:
- https://preview.aclanthology.org/landing_page/2025.acl-short.62.pdf