@inproceedings{choi-etal-2025-grouped,
title = "Grouped Sequency-arranged Rotation: Optimizing Rotation Transformation for Quantization for Free",
author = "Choi, Euntae and
Song, Sumin and
Lim, Woosang and
Yoo, Sungjoo",
editor = "Zhao, Jin and
Wang, Mingyang and
Liu, Zhu",
booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop)",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/landing_page/2025.acl-srw.10/",
pages = "165--172",
ISBN = "979-8-89176-254-1",
abstract = "Large Language Models (LLMs) face deployment challenges due to high computational costs, and while Post-Training Quantization (PTQ) offers a solution, existing rotation-based methods struggle at very low bit-widths like 2-bit. We introduce a novel, training-free approach to construct an improved rotation matrix, addressing the limitations of current methods. The key contributions include leveraging the Walsh-Hadamard transform with sequency ordering, which clusters similar frequency components to reduce quantization error compared to standard Hadamard matrices, significantly improving performance. Furthermore, we propose a Grouped Sequency-arranged Rotation (GSR) using block-diagonal matrices with smaller Walsh blocks, effectively isolating outlier impacts and achieving performance comparable to optimization-based methods without requiring any training. Our method demonstrates robust performance on reasoning tasks and Perplexity (PPL) score on WikiText-2. Our method also enhances results even when applied over existing learned rotation techniques."
}
Markdown (Informal)
[Grouped Sequency-arranged Rotation: Optimizing Rotation Transformation for Quantization for Free](https://preview.aclanthology.org/landing_page/2025.acl-srw.10/) (Choi et al., ACL 2025)
ACL