@inproceedings{wu-etal-2024-toolplanner,
title = "{T}ool{P}lanner: A Tool Augmented {LLM} for Multi Granularity Instructions with Path Planning and Feedback",
author = "Wu, Qinzhuo and
Liu, Wei and
Luan, Jian and
Wang, Bin",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2024.emnlp-main.1018/",
doi = "10.18653/v1/2024.emnlp-main.1018",
pages = "18315--18339",
abstract = "Recently, tool-augmented LLMs have gained increasing attention. Given an instruction, tool-augmented LLMs can interact with various external tools in multiple rounds and provide a final answer. However, previous LLMs were trained on overly detailed instructions, which included API names or parameters, while real users would not explicitly mention these API details. This leads to a gap between trained LLMs and real-world scenarios. In addition, most works ignore whether the interaction process follows the instruction. To address these issues, we constructed a training dataset called MGToolBench, which contains statement and category-level instructions to better reflect real-world scenarios. In addition, we propose ToolPlanner, a two-stage reinforcement learning framework that utilizes path planning and two feedback mechanisms to enhance the LLM`s task completion and instruction-following capabilities. Experimental results show that ToolPlanner significantly improves the Match Rate, Pass Rate and Win Rate by 26.8{\%}, 20.2{\%}, and 5.6{\%} compared to the SOTA model. Human evaluation verifies that the multi-granularity instructions can better align with users' usage habits. Our data and code will be released upon acceptance."
}
Markdown (Informal)
[ToolPlanner: A Tool Augmented LLM for Multi Granularity Instructions with Path Planning and Feedback](https://preview.aclanthology.org/add-emnlp-2024-awards/2024.emnlp-main.1018/) (Wu et al., EMNLP 2024)
ACL