LaMDAgent: An Autonomous Framework for Post-Training Pipeline Optimization via LLM Agents

Taro Yano, Yoichi Ishibashi, Masafumi Oyamada


Abstract
Large Language Models (LLMs) excel across diverse tasks, with post-training methods like Supervised Fine-Tuning (SFT), Preference Learning, and Model Merging enabling effective domain and task adaptation. While outcomes can vary with data orderings or component combinations, yet manual pipeline optimization is costly and labor-intensive. Existing approaches typically rely on manual design or focus narrowly on optimizing individual components, such as data ordering or merging parameters. We propose LaMDAgent, an LLM Agent-driven framework that autonomously constructs and optimizes end-to-end post-training pipelines by exploring various model improving methods, objects, and their applied orderings based on task-based feedback. LaMDAgent achieves a 9.0-point gain in tool-use accuracy without degrading instruction-following, and identifies high-performing strategies overlooked by manual design.We further analyze the impact of data and model scaling to reduce computational costs on the exploration, finding that model size scalings introduces new challenges, whereas scaling data size enables cost-effective pipeline discovery.
Anthology ID:
2025.emnlp-main.1529
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
30066–30083
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1529/
DOI:
Bibkey:
Cite (ACL):
Taro Yano, Yoichi Ishibashi, and Masafumi Oyamada. 2025. LaMDAgent: An Autonomous Framework for Post-Training Pipeline Optimization via LLM Agents. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 30066–30083, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
LaMDAgent: An Autonomous Framework for Post-Training Pipeline Optimization via LLM Agents (Yano et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1529.pdf
Checklist:
 2025.emnlp-main.1529.checklist.pdf