Michael Yang
2025
REIC: RAG-Enhanced Intent Classification at Scale
Ziji Zhang
|
Michael Yang
|
Zhiyu Chen
|
Yingying Zhuang
|
Shu-Ting Pi
|
Qun Liu
|
Rajashekar Maragoud
|
Vy Nguyen
|
Anurag Beniwal
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track
Accurate intent classification is critical for efficient routing in customer service, ensuring customers are connected with the most suitable agents while reducing handling times and operational costs. However, as companies expand their product lines, intent classification faces scalability challenges due to the increasing number of intents and variations in taxonomy across different verticals. In this paper, we introduce REIC, a Retrieval-augmented generation Enhanced Intent Classification approach, which addresses these challenges effectively. REIC leverages retrieval-augmented generation (RAG) to dynamically incorporate relevant knowledge, enabling precise classification without the need for frequent retraining. Through extensive experiments on real-world datasets, we demonstrate that REIC outperforms traditional fine-tuning, zero-shot, and few-shot methods in large-scale customer service settings. Our results highlight its effectiveness in both in-domain and out-of-domain scenarios, demonstrating its potential for real-world deployment in adaptive and large-scale intent classification systems.
2020
Training and Inference Methods for High-Coverage Neural Machine Translation
Michael Yang
|
Yixin Liu
|
Rahul Mayuranath
Proceedings of the Fourth Workshop on Neural Generation and Translation
In this paper, we introduce a system built for the Duolingo Simultaneous Translation And Paraphrase for Language Education (STAPLE) shared task at the 4th Workshop on Neural Generation and Translation (WNGT 2020). We participated in the English-to-Japanese track with a Transformer model pretrained on the JParaCrawl corpus and fine-tuned in two steps on the JESC corpus and then the (smaller) Duolingo training corpus. First, during training, we find it is essential to deliberately expose the model to higher-quality translations more often during training for optimal translation performance. For inference, encouraging a small amount of diversity with Diverse Beam Search to improve translation coverage yielded marginal improvement over regular Beam Search. Finally, using an auxiliary filtering model to filter out unlikely candidates from Beam Search improves performance further. We achieve a weighted F1 score of 27.56% on our own test set, outperforming the STAPLE AWS translations baseline score of 4.31%.
Search
Fix author
Co-authors
- Anurag Beniwal 1
- Zhiyu Chen 1
- Yixin Liu 1
- Qun Liu 1
- Rajashekar Maragoud 1
- show all...