@inproceedings{han-etal-2023-intervention,
title = "Intervention-Based Alignment of Code Search with Execution Feedback",
author = "Han, Hojae and
Kim, Minsoo and
Hwang, Seung-won and
Duan, Nan and
Lu, Shuai",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2023",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.findings-emnlp.148/",
doi = "10.18653/v1/2023.findings-emnlp.148",
pages = "2241--2263",
abstract = "One of the fundamental goals in code search is to retrieve a functionally correct code for a given natural language query. As annotating for correctness requires executing test cases (i.e. obtaining execution feedback), existing code search training datasets approximate text-code co-occurrences as positive execution feedback. However, this approximation may misalign models' retrieval decisions from ground-truth correctness. To address such limitation, we propose Code Intervention-based Reinforcement Learning (CIRL) that perturbs training code to result in misalignment (i.e. code intervention), then tests models' decisions and corrects them with the execution feedback by reinforcement learning. The first technical contribution of CIRL is to induce the execution feedback from perturbation, without actual execution. Secondly, CIRL introduces structural perturbations using abstract syntax trees, going beyond simple lexical changes. Experimental results on various datasets demonstrate the effectiveness of CIRL compared to conventional approaches."
}
Markdown (Informal)
[Intervention-Based Alignment of Code Search with Execution Feedback](https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.findings-emnlp.148/) (Han et al., Findings 2023)
ACL