Beyond Fixed-Length Calibration for Post-Training Compression of LLMs

Jaehoon Oh, Dokwan Oh


Abstract
As large language models (LLMs) continue to grow in size, their practical deployment increasingly relies on a range of compression techniques, such as quantization, pruning, and low-rank approximation. Especially, post-training compression methods–which do not require re-training–have drawn considerable interest. Many recent methods leverage calibration data to capture magnitude or second-order characteristics of input activations. However, the role and significance of calibration data remain underexplored. In this study, we demonstrate that the sequence length of calibration data plays a crucial role in the effectiveness of post-training compression methods for LLMs. We then analyze input activations and find that, within the normalized hidden states, the embedding of the first token exhibits characteristics opposite to those of subsequent tokens. Building on this insight, we introduce state-aware length calibration, a technique that applies masking along the sequence axis, specifically targeting normalized hidden states. Experimental results show that our approach improves perplexity and zero-shot downstream tasks performance.
Anthology ID:
2025.findings-emnlp.1054
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2025
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
19355–19366
Language:
URL:
https://preview.aclanthology.org/ingest-luhme/2025.findings-emnlp.1054/
DOI:
10.18653/v1/2025.findings-emnlp.1054
Bibkey:
Cite (ACL):
Jaehoon Oh and Dokwan Oh. 2025. Beyond Fixed-Length Calibration for Post-Training Compression of LLMs. In Findings of the Association for Computational Linguistics: EMNLP 2025, pages 19355–19366, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Beyond Fixed-Length Calibration for Post-Training Compression of LLMs (Oh & Oh, Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-luhme/2025.findings-emnlp.1054.pdf
Checklist:
 2025.findings-emnlp.1054.checklist.pdf