@inproceedings{lee-etal-2024-orchestrallm,
title = "{O}rchestra{LLM}: Efficient Orchestration of Language Models for Dialogue State Tracking",
author = "Lee, Chia-Hsuan and
Cheng, Hao and
Ostendorf, Mari",
editor = "Duh, Kevin and
Gomez, Helena and
Bethard, Steven",
booktitle = "Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)",
month = jun,
year = "2024",
address = "Mexico City, Mexico",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.naacl-long.79/",
doi = "10.18653/v1/2024.naacl-long.79",
pages = "1434--1445",
abstract = "Large language models (LLMs) have revolutionized the landscape of Natural Language Processing, but are computationally expensive. To reduce the cost without sacrificing performance, previous studies have explored various approaches to harness the potential of Smaller Language Models (SLMs) as cost-effective alternatives to their larger counterparts. Driven by findings that SLMs and LLMs exhibit complementary strengths in a structured knowledge extraction task, this work presents a novel SLM/LLM routing framework designed to improve computational efficiency and enhance task performance. In dialogue state tracking tasks, the proposed routing framework enhances performance substantially compared to relying solely on LLMs, while reducing the computational costs by over 50{\%}."
}
Markdown (Informal)
[OrchestraLLM: Efficient Orchestration of Language Models for Dialogue State Tracking](https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.naacl-long.79/) (Lee et al., NAACL 2024)
ACL