A Semantic-Aware Layer-Freezing Approach to Computation-Efficient Fine-Tuning of Language Models

Jian Gu, Aldeida Aleti, Chunyang Chen, Hongyu Zhang


Abstract
Finetuning language models (LMs) is crucial for adapting the models to downstream data and tasks. However, full finetuning is usually costly. Existing work, such as parameter-efficient finetuning (PEFT), often focuses on how to finetune but neglects the issue of where to finetune. As a pioneering work on reducing the cost of backpropagation (at the layer level) by answering where to finetune, we conduct a semantic analysis of the LM inference process. We first propose using transition traces of the latent representation to compute deviations (or loss). Then, using a derived formula of scaling law, we estimate the gain of each layer in reducing deviation (or loss). Further, we narrow down the scope for finetuning, and also, study the cost-benefit balance of LM finetuning. We perform extensive experiments across well-known LMs and datasets. The results show that our approach is effective and efficient, and outperforms the existing baselines. Our approach is orthogonal to other techniques for improving finetuning efficiency, such as PEFT methods, offering practical values on LM finetuning.
Anthology ID:
2025.findings-acl.420
Volume:
Findings of the Association for Computational Linguistics: ACL 2025
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Wanxiang Che, Joyce Nabende, Ekaterina Shutova, Mohammad Taher Pilehvar
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
8019–8033
Language:
URL:
https://preview.aclanthology.org/mtsummit-25-ingestion/2025.findings-acl.420/
DOI:
10.18653/v1/2025.findings-acl.420
Bibkey:
Cite (ACL):
Jian Gu, Aldeida Aleti, Chunyang Chen, and Hongyu Zhang. 2025. A Semantic-Aware Layer-Freezing Approach to Computation-Efficient Fine-Tuning of Language Models. In Findings of the Association for Computational Linguistics: ACL 2025, pages 8019–8033, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
A Semantic-Aware Layer-Freezing Approach to Computation-Efficient Fine-Tuning of Language Models (Gu et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/mtsummit-25-ingestion/2025.findings-acl.420.pdf