@inproceedings{yin-etal-2025-midlm,
title = "{MIDLM}: Multi-Intent Detection with Bidirectional Large Language Models",
author = "Yin, Shangjian and
Huang, Peijie and
Xu, Yuhong",
editor = "Rambow, Owen and
Wanner, Leo and
Apidianaki, Marianna and
Al-Khalifa, Hend and
Eugenio, Barbara Di and
Schockaert, Steven",
booktitle = "Proceedings of the 31st International Conference on Computational Linguistics",
month = jan,
year = "2025",
address = "Abu Dhabi, UAE",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2025.coling-main.179/",
pages = "2616--2625",
abstract = "Decoder-only Large Language Models (LLMs) have demonstrated exceptional performance in language generation, exhibiting broad capabilities across various tasks. However, the application to label-sensitive language understanding tasks remains challenging due to the limitations of their autoregressive architecture, which restricts the sharing of token information within a sentence. In this paper, we address the Multi-Intent Detection (MID) task and introduce MIDLM, a bidirectional LLM framework that incorporates intent number detection and multi-intent selection. This framework allows autoregressive LLMs to leverage bidirectional information awareness through post-training, eliminating the need for training the models from scratch. Comprehensive evaluations across 8 datasets show that MIDLM consistently outperforms both existing vanilla models and pretrained baselines, demonstrating its superior performance in the MID task."
}
Markdown (Informal)
[MIDLM: Multi-Intent Detection with Bidirectional Large Language Models](https://preview.aclanthology.org/fix-sig-urls/2025.coling-main.179/) (Yin et al., COLING 2025)
ACL