Understand, Solve and Translate: Bridging the Multilingual Mathematical Reasoning Gap

Hyunwoo Ko, Guijin Son, Dasol Choi


Abstract
Large language models (LLMs) demonstrate exceptional performance on complex reasoning tasks. However, despite their strong reasoning capabilities in high-resource languages (e.g., English and Chinese), a significant performance gap persists in other languages. To investigate this gap in Korean, we introduce HRM8K, a benchmark comprising 8,011 English-Korean parallel bilingual math problems. Through systematic analysis of model behaviors, we identify a key finding: these performance disparities stem primarily from difficulties in comprehending non-English inputs, rather than limitations in reasoning capabilities. Based on these findings, we propose UST(Understand, Solve, and Translate), a method that strategically uses English as an anchor for reasoning and solution generation. By fine-tuning the model on 130k synthetically generated data points, method achieves a 10.91% improvement on the HRM8K benchmark and reduces the multilingual performance gap from 11.6%% to 0.7%%. Additionally, we show that improvements from method generalize effectively to different Korean domains, demonstrating that capabilities acquired from machine-verifiable content can be generalized to other areas. We publicly release the benchmark, training dataset, and models.
Anthology ID:
2025.mrl-main.6
Volume:
Proceedings of the 5th Workshop on Multilingual Representation Learning (MRL 2025)
Month:
November
Year:
2025
Address:
Suzhuo, China
Editors:
David Ifeoluwa Adelani, Catherine Arnett, Duygu Ataman, Tyler A. Chang, Hila Gonen, Rahul Raja, Fabian Schmidt, David Stap, Jiayi Wang
Venues:
MRL | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
78–95
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.mrl-main.6/
DOI:
Bibkey:
Cite (ACL):
Hyunwoo Ko, Guijin Son, and Dasol Choi. 2025. Understand, Solve and Translate: Bridging the Multilingual Mathematical Reasoning Gap. In Proceedings of the 5th Workshop on Multilingual Representation Learning (MRL 2025), pages 78–95, Suzhuo, China. Association for Computational Linguistics.
Cite (Informal):
Understand, Solve and Translate: Bridging the Multilingual Mathematical Reasoning Gap (Ko et al., MRL 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.mrl-main.6.pdf