Enhanced Data Synthesis for LLM through Reasoning Structures Generated by Hierarchical GFlowNet

Tianpeng Bu, Minying Zhang, Hongtao Duan, Shurui Li, Lulu Hu, Yu Li


Abstract
Large language models (LLMs) excel in problem-solving but require training data with diverse reasoning processes. Existing methods mainly optimize instruction-response pairs but lack a systematic design for the underlying reasoning structure. This paper proposes RSS: a Reasoning Structure driven data Synthesis method. We first proactively develop a hierarchical GFlowNet to construct reasoning structures efficiently through a coarse-to-fine directed acyclic graph (DAG) growth process. Then reasoning DAGs are leveraged to actively guide the instruction generation via an iterative suggester-editor workflow and enhance response quality using a structure-aware strategy. Experiments show that LLMs trained on our synthetic datasets achieve 48.50%, 84.00%, 79.90% for AlpacaEval2, GSM8K and HumanEval, outperforming existing data synthesis methods.
Anthology ID:
2025.findings-acl.821
Volume:
Findings of the Association for Computational Linguistics: ACL 2025
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Wanxiang Che, Joyce Nabende, Ekaterina Shutova, Mohammad Taher Pilehvar
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
15931–15958
Language:
URL:
https://preview.aclanthology.org/display_plenaries/2025.findings-acl.821/
DOI:
Bibkey:
Cite (ACL):
Tianpeng Bu, Minying Zhang, Hongtao Duan, Shurui Li, Lulu Hu, and Yu Li. 2025. Enhanced Data Synthesis for LLM through Reasoning Structures Generated by Hierarchical GFlowNet. In Findings of the Association for Computational Linguistics: ACL 2025, pages 15931–15958, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Enhanced Data Synthesis for LLM through Reasoning Structures Generated by Hierarchical GFlowNet (Bu et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/display_plenaries/2025.findings-acl.821.pdf