Jiaping Wang
2025
Consultant Decoding: Yet Another Synergistic Mechanism
Chuanghao Ding
|
Jiaping Wang
|
Ziqing Yang
|
Xiaoliang Wang
|
Dahua Lin
|
Nguyen Cam-Tu
|
Fei Tan
Findings of the Association for Computational Linguistics: ACL 2025
The synergistic mechanism based on Speculative Decoding (SD) has garnered considerable attention as a simple yet effective approach for accelerating the inference of large language models (LLMs). Nonetheless, the high rejection rates require repeated LLMs calls to validate draft tokens, undermining the overall efficiency gain of SD.In this work, we revisit existing verification mechanisms and propose a novel synergetic mechanism Consultant Decoding (CD). CD achieves up to a 2.5-fold increase in inference speed compared to the target model, while maintaining comparable generation quality (~100% of the target model’s performance). Interestingly, this is achieved by combining models whose parameter sizes differ by two orders of magnitude.In addition, CD reduces the call frequency of the large target model to below 10%, particularly in more demanding tasks.CD’s performance was even found to surpass that of the large target model, which theoretically represents the upper bound for speculative decoding.
Search
Fix author
Co-authors
- Nguyen Cam-Tu 1
- Chuanghao Ding 1
- Dahua Lin 1
- Fei Tan 1
- Xiaoliang Wang 1
- show all...