@inproceedings{jeon-etal-2023-frustratingly,
title = "A Frustratingly Easy Post-Training Quantization Scheme for {LLM}s",
author = "Jeon, Yongkweon and
Lee, Chungman and
Park, Kyungphil and
Kim, Ho-young",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.emnlp-main.892/",
doi = "10.18653/v1/2023.emnlp-main.892",
pages = "14446--14461",
abstract = "Efficient inference has become crucial for hyper-scale AI models, including large language models, as their parameter count continues to increase for enhanced performance. This necessity holds true regardless of the computing environment, whether it be mobile devices or cloud servers. Quantization emerges as a solution to alleviate the computational burden during inference. By representing models with a reduced bit-width, quantization minimizes the frequency of DRAM access while fully exploiting the parallelism of operations through a dense matrix format. Consequently, quantized models achieve low end-to-end latency and optimize resource utilization by addressing both memory and computing bottlenecks. In this paper, we propose a straightforward post-training quantization scheme, called Z-Fold, that fully utilizes the feature of the Transformer structure widely employed in large language models."
}
Markdown (Informal)
[A Frustratingly Easy Post-Training Quantization Scheme for LLMs](https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.emnlp-main.892/) (Jeon et al., EMNLP 2023)
ACL