@inproceedings{hacioglu-etal-2025-speechllms,
title = "{S}peech{LLM}s for Large-scale Contextualized Zero-shot Slot Filling",
author = "Hacioglu, Kadri and
E, Manjunath K and
Stolcke, Andreas",
editor = "Potdar, Saloni and
Rojas-Barahona, Lina and
Montella, Sebastien",
booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track",
month = nov,
year = "2025",
address = "Suzhou (China)",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-industry.49/",
pages = "703--715",
ISBN = "979-8-89176-333-3",
abstract = "Slot filling is a crucial subtask in spoken language understanding (SLU), traditionally implemented as a cascade of speech recognition followed by one or more natural language understanding (NLU) components. The recent advent of speech-based large language models (speechLLMs), which integrate speech and textual foundation models, has opened new avenues for achieving speech understanding tasks in a more unified, generative, and instruction-following manner while promising data and compute efficiency with zero-shot abilities, generalizing to unseen slot labels. We address the slot-filling task by creating an empirical upper bound for the task, identifying performance, robustness, and generalization gaps, and proposing improvements to the training data, architecture, and training strategies to narrow the gap with the upper bound result. We show that each of these measures improve performance substantially, while highlighting practical challenges and providing empirical guidance and insights for harnessing these emerging models."
}Markdown (Informal)
[SpeechLLMs for Large-scale Contextualized Zero-shot Slot Filling](https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-industry.49/) (Hacioglu et al., EMNLP 2025)
ACL