@inproceedings{yano-etal-2025-lamdagent,
title = "{L}a{MDA}gent: An Autonomous Framework for Post-Training Pipeline Optimization via {LLM} Agents",
author = "Yano, Taro and
Ishibashi, Yoichi and
Oyamada, Masafumi",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1529/",
pages = "30066--30083",
ISBN = "979-8-89176-332-6",
abstract = "Large Language Models (LLMs) excel across diverse tasks, with post-training methods like Supervised Fine-Tuning (SFT), Preference Learning, and Model Merging enabling effective domain and task adaptation. While outcomes can vary with data orderings or component combinations, yet manual pipeline optimization is costly and labor-intensive. Existing approaches typically rely on manual design or focus narrowly on optimizing individual components, such as data ordering or merging parameters. We propose LaMDAgent, an LLM Agent-driven framework that autonomously constructs and optimizes end-to-end post-training pipelines by exploring various model improving methods, objects, and their applied orderings based on task-based feedback. LaMDAgent achieves a 9.0-point gain in tool-use accuracy without degrading instruction-following, and identifies high-performing strategies overlooked by manual design.We further analyze the impact of data and model scaling to reduce computational costs on the exploration, finding that model size scalings introduces new challenges, whereas scaling data size enables cost-effective pipeline discovery."
}Markdown (Informal)
[LaMDAgent: An Autonomous Framework for Post-Training Pipeline Optimization via LLM Agents](https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1529/) (Yano et al., EMNLP 2025)
ACL