@inproceedings{zhao-etal-2025-qspec,
title = "{QS}pec: Speculative Decoding with Complementary Quantization Schemes",
author = "Zhao, Juntao and
Lu, Wenhao and
Wang, Sheng and
Kong, Lingpeng and
Wu, Chuan",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.240/",
pages = "4779--4795",
ISBN = "979-8-89176-332-6",
abstract = "Quantization is widely adopted to accelerate inference and reduce memory consumption in large language models (LLMs). While activation-weight joint quantization enables efficient low-precision decoding, it suffers substantial performance degradation on multi-step reasoning tasks. We propose QSPEC, a novel quantization paradigm that decouples efficiency from quality by integrating two complementary schemes via speculative decoding: low-precision joint quantization for fast drafting and high-precision weight-only quantization for accurate verification. QSPEC reuses both weights and KV cache across stages, enabling near-zero-cost switching without retraining or auxiliary models. Compared to high-precision baselines, QSPEC achieves up to 1.64x speedup without quality degradation, and outperforms state-of-the-art speculative decoding methods by up to 1.55x in batched settings. Furthermore, QSPEC supports plug-and-play deployment and generalizes well across model scales, quantization methods, and workloads. These properties make QSPEC a practical and scalable solution for high-fidelity quantized LLM serving under memory-constrained scenarios."
}Markdown (Informal)
[QSpec: Speculative Decoding with Complementary Quantization Schemes](https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.240/) (Zhao et al., EMNLP 2025)
ACL