@inproceedings{li-etal-2024-towards,
title = "Towards Autonomous Tool Utilization in Language Models: A Unified, Efficient and Scalable Framework",
author = "Li, Zhi and
Li, Yicheng and
Ye, Hequan and
Zhang, Yin",
editor = "Calzolari, Nicoletta and
Kan, Min-Yen and
Hoste, Veronique and
Lenci, Alessandro and
Sakti, Sakriani and
Xue, Nianwen",
booktitle = "Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)",
month = may,
year = "2024",
address = "Torino, Italia",
publisher = "ELRA and ICCL",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.lrec-main.1427/",
pages = "16422--16432",
abstract = "In recent research, significant advancements have been achieved in tool learning for large language models. Looking towards future advanced studies, the issue of fully autonomous tool utilization is particularly intriguing: given only a query, language models can autonomously decide whether to employ a tool, which specific tool to select, and how to utilize these tools, all without needing any tool-specific prompts within the context. To achieve this, we introduce a unified, efficient, and scalable framework for fine-tuning language models. Based on the degree of tool dependency, we initially categorize queries into three distinct types. By transforming the entire process into a sequential decision-making problem through conditional probability decomposition, our approach unifies the three types and autoregressively generates decision processes. Concurrently, we`ve introduced an {\textquotedblleft}instruct, execute, and reformat{\textquotedblright} strategy specifically designed for efficient data annotation. Through end-to-end training on the annotated dataset comprising 26 diverse APIs, the model demonstrates a level of self-awareness, automatically seeking tool assistance when necessary. It significantly surpasses original instruction-tuned open-source language models and GPT-3.5/4 on multiple evaluation metrics. To address real-world scalability needs, we`ve enhanced our framework with a dynamic rehearsal strategy for continual learning, proven to require minimal new annotations to exhibit remarkable performance."
}
Markdown (Informal)
[Towards Autonomous Tool Utilization in Language Models: A Unified, Efficient and Scalable Framework](https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.lrec-main.1427/) (Li et al., LREC-COLING 2024)
ACL