@inproceedings{wang-etal-2025-swe,
title = "{SWE}-Dev: Building Software Engineering Agents with Training and Inference Scaling",
author = "Wang, Haoran and
Hou, Zhenyu and
Wei, Yao and
Tang, Jie and
Dong, Yuxiao",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2025",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingestion-acl-25/2025.findings-acl.193/",
pages = "3742--3761",
ISBN = "979-8-89176-256-5",
abstract = "Large language models (LLMs) have advanced rapidly from conversational problem solving to addressing real-world tasks involving tool use, such as software engineering (SWE). Recent LLM-powered toolkits, such as OpenAI Codex and Cursor, have offered end-to-end automation of the software development process. However, building effective SWE agents remains challenging due to the lack of high-quality training data and effective test cases. To address this issue, we present SWE-Dev, an SWE agent built upon open-source LLMs. First, we develop a robust pipeline to synthesize test cases for patch evaluation. Second, we scale up agent trajectories to construct the training data for building SWE-Dev. Experiments on the SWE-bench-Verified benchmark show that the SWE-Dev models can achieve top performance among all open SWE agents. Specifically, the success rates of the SWE-Dev 7B and 32B parameter models reach 23.4{\%} and 36.6{\%}, respectively, outperforming state-of-the-art open-source models. All code, models, and datasets are publicly available at https://github.com/THUDM/SWE-Dev."
}
Markdown (Informal)
[SWE-Dev: Building Software Engineering Agents with Training and Inference Scaling](https://preview.aclanthology.org/ingestion-acl-25/2025.findings-acl.193/) (Wang et al., Findings 2025)
ACL