Xuanqing Liu
2025
Learning LLM Preference over Intra-Dialogue Pairs: A Framework for Utterance-level Understandings
Xuanqing Liu
|
Luyang Kong
|
Wei Niu
|
Afshin Khashei
|
Belinda Zeng
|
Steve Johnson
|
Jon Jay
|
Davor Golac
|
Matt Pope
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 3: Industry Track)
Large language models (LLMs) have demonstrated remarkable capabilities in handling complex dialogue tasks without requiring use case-specific fine-tuning. However, analyzing live dialogues in real-time necessitates low-latency processing systems, making it impractical to deploy models with billions of parameters due to latency constraints. As a result, practitioners often prefer smaller models with millions of parameters, trained on high-quality, human-annotated datasets. Yet, curating such datasets is both time-consuming and costly. Consequently, there is a growing need to combine the scalability of LLM-generated labels with the precision of human annotations, enabling fine-tuned smaller models to achieve both higher speed and accuracy comparable to larger models. In this paper, we introduce a simple yet effective framework to address this challenge. Our approach is specifically designed for per-utterance classification problems, which encompass tasks such as intent detection, dialogue state tracking, and more. To mitigate the impact of labeling errors from LLMs – the primary source of inaccuracies in student models – we propose a noise-reduced preference learning loss. Experimental results demonstrate that our method significantly improves accuracy across utterance-level dialogue tasks, including sentiment detection (over 2%), dialogue act classification (over 1.5%), etc.
2024
BPID: A Benchmark for Personal Identity Deduplication
Runhui Wang
|
Yefan Tao
|
Adit Krishnan
|
Luyang Kong
|
Xuanqing Liu
|
Yuqian Deng
|
Yunzhao Yang
|
Henrik Johnson
|
Andrew Borthwick
|
Shobhit Gupta
|
Aditi Sinha Gundlapalli
|
Davor Golac
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing: Industry Track
Data deduplication is a critical task in data management and mining, focused on consolidating duplicate records that refer to the same entity. Personally Identifiable Information (PII) is a critical class of data for deduplication across various industries. Consumer data, stored and generated through various engagement channels, is crucial for marketers, agencies, and publishers. However, a major challenge to PII data deduplication is the lack of open-source benchmark datasets due to stringent privacy concerns, which hinders the research, development, and evaluation of robust solutions.This paper addresses this critical lack of PII deduplication benchmarks by introducing the first open-source, high-quality dataset for this task. We provide two datasets: one with 1,000,000 unlabeled synthetic PII profiles and a subset of 10,000 pairs curated and labeled by trained annotators as matches or non-matches. Our datasets contain synthetic profiles built from publicly available sources that do not represent any real individuals, thus ensuring privacy and ethical compliance. We provide several challenging data variations to evaluate the effectiveness of various deduplication techniques, including traditional supervised methods, deep-learning approaches, and large language models (LLMs). Our work aims to set a new standard for PII deduplication, paving the way for more accurate and secure solutions. We share our data publicly at this link - https://zenodo.org/records/13932202.
Search
Fix data
Co-authors
- Davor Golac 2
- Luyang Kong 2
- Andrew Borthwick 1
- Yuqian Deng 1
- Aditi Sinha Gundlapalli 1
- show all...