Navigating the Unknown: Intent Classification and Out-of-Distribution Detection Using Large Language Models

Yusuf Sali, Sıtkı Can Toraman


Abstract
Out-of-Distribution (OOD) detection is a challenging task that requires great generalization capability for the practicality and safety of task-oriented dialogue systems (TODS). With the dawn of large language models (LLMs), their enhanced ability to handle diverse patterns and contexts may aid in addressing this challenging task. In this paper, we investigate the current performance of LLMs in the near-OOD setting, where OOD queries belong to the same domain but different intents. To take advantage of out-of-the-shelf capabilities of LLMs, we do not use fine-tuning. We study the performance of one of the leading frontier models, GPT-4o, in 3 well-known public datasets and 3 in-house datasets, using 10 different methods and prompt variations. We study the performance of different prompts and techniques in Gemini 1.5 Flash and Llama 3.1-70b. We investigate the effect of increasing the number of In-Distribution (ID) intents. We propose a novel hybrid method that is cost-efficient, high-performing, highly robust, and versatile enough to be used with smaller LLMs without sacrificing performance. This is achieved by combining ID success of smaller text classification models and high generalization capabilities of LLMs in OOD detection.
Anthology ID:
2025.findings-emnlp.791
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2025
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
14652–14664
Language:
URL:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.791/
DOI:
10.18653/v1/2025.findings-emnlp.791
Bibkey:
Cite (ACL):
Yusuf Sali and Sıtkı Can Toraman. 2025. Navigating the Unknown: Intent Classification and Out-of-Distribution Detection Using Large Language Models. In Findings of the Association for Computational Linguistics: EMNLP 2025, pages 14652–14664, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Navigating the Unknown: Intent Classification and Out-of-Distribution Detection Using Large Language Models (Sali & Toraman, Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.791.pdf
Checklist:
 2025.findings-emnlp.791.checklist.pdf