Unlocking the Planning Capabilities of Large Language Models with Maximum Diversity Fine-tuning

Wenjun Li, Changyu Chen, Pradeep Varakantham


Abstract
Large language models (LLMs) have demonstrated impressive task-solving capabilities through prompting techniques and system designs, including solving planning tasks (e.g., math proofs, basic travel planning) when sufficient data is available online and used during pre-training. However, for planning tasks with limited prior data (e.g., blocks world, advanced travel planning), the performance of LLMs, including proprietary models like GPT and Gemini, is poor. This paper investigates the impact of fine-tuning on the planning capabilities of LLMs, revealing that LLMs can achieve strong performance in planning through substantial (tens of thousands of specific examples) fine-tuning. Yet, this process incurs high economic, time, and computational costs for each planning problem variation. To address this, we propose Clustering-Based Maximum Diversity Sampling (CMDS), which selects diverse and representative data to enhance sample efficiency and the model’s generalization capability. Extensive evaluations demonstrate that CMDS-l, a baseline method combining CMDS with language embeddings, outperforms random sampling. Furthermore, we introduce a novel algorithm, CMDS-g, which encodes planning task instances with their graph representations into the embedding space. Empirical results show that CMDS-g consistently outperforms baseline methods across various scales and multiple benchmark domains.
Anthology ID:
2025.findings-naacl.183
Volume:
Findings of the Association for Computational Linguistics: NAACL 2025
Month:
April
Year:
2025
Address:
Albuquerque, New Mexico
Editors:
Luis Chiruzzo, Alan Ritter, Lu Wang
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
3318–3340
Language:
URL:
https://preview.aclanthology.org/fix-sig-urls/2025.findings-naacl.183/
DOI:
Bibkey:
Cite (ACL):
Wenjun Li, Changyu Chen, and Pradeep Varakantham. 2025. Unlocking the Planning Capabilities of Large Language Models with Maximum Diversity Fine-tuning. In Findings of the Association for Computational Linguistics: NAACL 2025, pages 3318–3340, Albuquerque, New Mexico. Association for Computational Linguistics.
Cite (Informal):
Unlocking the Planning Capabilities of Large Language Models with Maximum Diversity Fine-tuning (Li et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/fix-sig-urls/2025.findings-naacl.183.pdf