Towards Infinite-Long Prefix in Transformer

Yingyu Liang, Zhenmei Shi, Zhao Song, Chiwun Yang


Abstract
Prompting and context-based fine-tuning methods, which we call Prefix Learning, have been proposed to enhance the performance of language models on various downstream tasks. They are empirically efficient and effective, matching the performance of full parameter fine-tuning, but the theoretical understandings are limited. In this paper, we aim to address this limitation by studying their ability from the perspective of prefix length. In particular, we provide a convergence guarantee for training an ultra-long prefix in a stylized setting using the Neural Tangent Kernel (NTK) framework. Based on this strong theoretical guarantee, we design and implement an algorithm that only needs to introduce and fine-tune a few extra trainable parameters instead of an infinite-long prefix in each layer of a transformer, and can approximate the prefix attention to a guaranteed polynomial-small error.Preliminary experimental results on vision, natural language, and math data show that our method achieves superior or competitive performance compared to existing methods like full parameters fine-tuning, P-Tuning V2, and LoRA. This demonstrates our method is promising for parameter-efficient fine-tuning.
Anthology ID:
2025.emnlp-main.563
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
11138–11202
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.563/
DOI:
Bibkey:
Cite (ACL):
Yingyu Liang, Zhenmei Shi, Zhao Song, and Chiwun Yang. 2025. Towards Infinite-Long Prefix in Transformer. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 11138–11202, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Towards Infinite-Long Prefix in Transformer (Liang et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.563.pdf
Checklist:
 2025.emnlp-main.563.checklist.pdf