Xiaoliang Wang
2025
daDPO: Distribution-Aware DPO for Distilling Conversational Abilities
Zhengze Zhang
|
Shiqi Wang
|
Yiqun Shen
|
Simin Guo
|
Dahua Lin
|
Xiaoliang Wang
|
Nguyen Cam-Tu
|
Fei Tan
Findings of the Association for Computational Linguistics: ACL 2025
Large language models (LLMs) have demonstrated exceptional performance across various applications, but their conversational abilities decline sharply as model size decreases, presenting a barrier to their deployment in resource-constrained environments. Knowledge distillation (KD) with Direct Preference Optimization (DPO) has emerged as a promising approach to enhance the conversational abilities of smaller models using a larger teacher model. However, current methods primarily focus on “black-box” KD, which only uses the teacher’s responses, overlooking the rich distributional information within the teacher’s probability distribution. This paper addresses this gap by introducing daDPO (Distillation-Aware DPO), a novel framework that integrates the teacher’s distributional information into DPO distillation while preserving theoretical guarantees. Our framework offers a unified objective that enhances both preference optimization and distribution-based distillation. We provide rigorous theoretical analysis and empirical validation, showing that daDPO outperforms existing methods in restoring performance for pruned models and enhancing smaller models within the same LLM family. Notably, in in-domain evaluation, our method enables a 20% pruned Vicuna1.5-7B to achieve near-teacher performance (-7.3% preference rate), and allows Qwen2.5-1.5B to occasionally outperform its 7b teacher model (14.0% win rate).
Consultant Decoding: Yet Another Synergistic Mechanism
Chuanghao Ding
|
Jiaping Wang
|
Ziqing Yang
|
Xiaoliang Wang
|
Dahua Lin
|
Nguyen Cam-Tu
|
Fei Tan
Findings of the Association for Computational Linguistics: ACL 2025
The synergistic mechanism based on Speculative Decoding (SD) has garnered considerable attention as a simple yet effective approach for accelerating the inference of large language models (LLMs). Nonetheless, the high rejection rates require repeated LLMs calls to validate draft tokens, undermining the overall efficiency gain of SD.In this work, we revisit existing verification mechanisms and propose a novel synergetic mechanism Consultant Decoding (CD). CD achieves up to a 2.5-fold increase in inference speed compared to the target model, while maintaining comparable generation quality (~100% of the target model’s performance). Interestingly, this is achieved by combining models whose parameter sizes differ by two orders of magnitude.In addition, CD reduces the call frequency of the large target model to below 10%, particularly in more demanding tasks.CD’s performance was even found to surpass that of the large target model, which theoretically represents the upper bound for speculative decoding.
2018
Joint learning of frequency and word embeddings for multilingual readability assessment
Dieu-Thu Le
|
Cam-Tu Nguyen
|
Xiaoliang Wang
Proceedings of the 5th Workshop on Natural Language Processing Techniques for Educational Applications
This paper describes two models that employ word frequency embeddings to deal with the problem of readability assessment in multiple languages. The task is to determine the difficulty level of a given document, i.e., how hard it is for a reader to fully comprehend the text. The proposed models show how frequency information can be integrated to improve the readability assessment. The experimental results testing on both English and Chinese datasets show that the proposed models improve the results notably when comparing to those using only traditional word embeddings.
Search
Fix author
Co-authors
- Nguyen Cam-Tu 2
- Dahua Lin 2
- Fei Tan 2
- Chuanghao Ding 1
- Simin Guo 1
- show all...