@inproceedings{sheng-shuai-2025-slm,
title = "{SLM}-{SQL}: An Exploration of Small Language Models for Text-to-{SQL}",
author = "Sheng, Lei and
Shuai, Xu Shuai",
editor = "Inui, Kentaro and
Sakti, Sakriani and
Wang, Haofen and
Wong, Derek F. and
Bhattacharyya, Pushpak and
Banerjee, Biplab and
Ekbal, Asif and
Chakraborty, Tanmoy and
Singh, Dhirendra Pratap",
booktitle = "Proceedings of the 14th International Joint Conference on Natural Language Processing and the 4th Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics",
month = dec,
year = "2025",
address = "Mumbai, India",
publisher = "The Asian Federation of Natural Language Processing and The Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-ijcnlp-aacl/2025.findings-ijcnlp.92/",
pages = "1497--1512",
ISBN = "979-8-89176-303-6",
abstract = "Large language models (LLMs) have demonstrated strong performance in translating natural language questions into SQL queries (Text-to-SQL). In contrast, small language models (SLMs) ranging from 0.5B to 1.5B parameters currently underperform on Text-to-SQL tasks due to their limited logical reasoning capabilities. However, SLMs offer inherent advantages in inference speed and suitability for edge deployment. To explore their potential in Text-to-SQL applications, we leverage recent advancements in post-training techniques. Specifically, we used the open-source SynSQL-2.5M dataset to construct two derived datasets: SynSQL-Think-916K for SQL generation and SynSQL-Merge-Think-310K for SQL merge revision. We then applied supervised fine-tuning and reinforcement learning-based post-training to the SLM, followed by inference using a corrective self-consistency approach. Experimental results validate the effectiveness and generalizability of our method, SLM-SQL. On the BIRD development set, the five evaluated models achieved an average improvement of 31.4 points. Notably, the 0.5B model reached 56.87{\%} execution accuracy (EX), while the 1.5B model achieved 67.08{\%} EX. On the BIRD private test set, our 0.5B model achieves 61.82{\%} EX, while the 1.5B model achieves 70.49{\%}. We will release our dataset, model, and code to github: https://github.com/CycloneBoy/slm{\_}sql."
}Markdown (Informal)
[SLM-SQL: An Exploration of Small Language Models for Text-to-SQL](https://preview.aclanthology.org/ingest-ijcnlp-aacl/2025.findings-ijcnlp.92/) (Sheng & Shuai, Findings 2025)
ACL
- Lei Sheng and Xu Shuai Shuai. 2025. SLM-SQL: An Exploration of Small Language Models for Text-to-SQL. In Proceedings of the 14th International Joint Conference on Natural Language Processing and the 4th Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics, pages 1497–1512, Mumbai, India. The Asian Federation of Natural Language Processing and The Association for Computational Linguistics.