PromptCoT: Synthesizing Olympiad-level Problems for Mathematical Reasoning in Large Language Models

Xueliang Zhao, Wei Wu, Jian Guan, Lingpeng Kong


Abstract
The ability of large language models to solve complex mathematical problems has progressed significantly, particularly for tasks requiring advanced reasoning. However, the scarcity of sufficiently challenging problems, particularly at the Olympiad level, hinders further advancements. In this work, we introduce PromptCoT, a novel approach for automatically generating high-quality Olympiad-level math problems. The proposed method synthesizes complex problems based on mathematical concepts and the rationale behind problem construction, emulating the thought processes of experienced problem designers. We provide a theoretical analysis demonstrating that an optimal rationale should maximize both the likelihood of rationale generation given the associated concepts and the likelihood of problem generation conditioned on both the rationale and the concepts. Our method is evaluated on standard benchmarks including GSM8K, MATH-500, and AIME2024, where it consistently outperforms existing problem generation methods. Furthermore, we demonstrate that PromptCoT exhibits superior scalability, consistently maintaining high performance as the dataset size increases, outperforming the baselines.
Anthology ID:
2025.findings-acl.935
Volume:
Findings of the Association for Computational Linguistics: ACL 2025
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Wanxiang Che, Joyce Nabende, Ekaterina Shutova, Mohammad Taher Pilehvar
Venues:
Findings | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
18167–18188
Language:
URL:
https://preview.aclanthology.org/ingestion-acl-25/2025.findings-acl.935/
DOI:
Bibkey:
Cite (ACL):
Xueliang Zhao, Wei Wu, Jian Guan, and Lingpeng Kong. 2025. PromptCoT: Synthesizing Olympiad-level Problems for Mathematical Reasoning in Large Language Models. In Findings of the Association for Computational Linguistics: ACL 2025, pages 18167–18188, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
PromptCoT: Synthesizing Olympiad-level Problems for Mathematical Reasoning in Large Language Models (Zhao et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingestion-acl-25/2025.findings-acl.935.pdf