@inproceedings{zhu-etal-2025-post,
title = "Can Post-Training Quantization Benefit from an Additional {QL}o{RA} Integration?",
author = "Zhu, Xiliang and
Khasanova, Elena and
Chen, Cheng",
editor = "Chen, Weizhu and
Yang, Yi and
Kachuee, Mohammad and
Fu, Xue-Yong",
booktitle = "Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 3: Industry Track)",
month = apr,
year = "2025",
address = "Albuquerque, New Mexico",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/Ingest-2025-COMPUTEL/2025.naacl-industry.41/",
pages = "506--514",
ISBN = "979-8-89176-194-0",
abstract = "Large language models (LLMs) have transformed natural language processing but pose significant challenges for real-world deployment. These models necessitate considerable computing resources, which can be costly and frequently unavailable. Model compression techniques such as quantization are often leveraged to alleviate resource demand, but they may have a negative impact on the generation quality. In this study, we explore the integration of 4-bit Post-training Quantization (PTQ) with QLoRA to address these issues. We demonstrate through extensive experiments that this integration outperforms standard PTQ, and in some cases even 16-bit full-parameter fine-tuning on LLMs, validated across proprietary and public datasets with different quantization algorithms. The results demonstrate the efficacy of PTQ-QLoRA integration, offering a viable solution for deploying powerful LLMs in resource-constrained environments without compromising on performance."
}
Markdown (Informal)
[Can Post-Training Quantization Benefit from an Additional QLoRA Integration?](https://preview.aclanthology.org/Ingest-2025-COMPUTEL/2025.naacl-industry.41/) (Zhu et al., NAACL 2025)
ACL
- Xiliang Zhu, Elena Khasanova, and Cheng Chen. 2025. Can Post-Training Quantization Benefit from an Additional QLoRA Integration?. In Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 3: Industry Track), pages 506–514, Albuquerque, New Mexico. Association for Computational Linguistics.