Prompt Tuning Can Simply Adapt Large Language Models to Text Encoders

Kaiyan Zhao, Qiyu Wu, Zhongtao Miao, Yoshimasa Tsuruoka


Abstract
Recently, many works have been attempting to adapt Large Language Models (LLMs) for sentence embedding, with most of them fine-tuning LLMs towards the contrastive objective and enabling bi-directional attention for better performance, using LoRA to address the large model scale.In this work, we suggest that this adaptation can also be simply and effectively achieved using causal attention and with even fewer trainable parameters through soft prompt tuning, as an alternative to fine-tuning with LoRA and other methods with extra post-training tasks.Our method only optimizes a few learnable tokens while keeping the rest of the model frozen.Through experiments on a diverse set of evaluation tasks, we find that simply tuning only a few tokens can achieve a competitive performance with that of fine-tuning with LoRA. The percentage of trainable parameters can be reduced to less than 0.001%. Moreover, we also demonstrate that turning causal attention to bi-directional attention with or without extra post-training tasks does not provide additional benefit when soft prompt tuning is applied, suggesting that causal attention can be naturally used in decoder-only LLMs for sentence embedding adaptation.
Anthology ID:
2025.repl4nlp-1.3
Volume:
Proceedings of the 10th Workshop on Representation Learning for NLP (RepL4NLP-2025)
Month:
May
Year:
2025
Address:
Albuquerque, NM
Editors:
Vaibhav Adlakha, Alexandra Chronopoulou, Xiang Lorraine Li, Bodhisattwa Prasad Majumder, Freda Shi, Giorgos Vernikos
Venues:
RepL4NLP | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
38–50
Language:
URL:
https://preview.aclanthology.org/landing_page/2025.repl4nlp-1.3/
DOI:
Bibkey:
Cite (ACL):
Kaiyan Zhao, Qiyu Wu, Zhongtao Miao, and Yoshimasa Tsuruoka. 2025. Prompt Tuning Can Simply Adapt Large Language Models to Text Encoders. In Proceedings of the 10th Workshop on Representation Learning for NLP (RepL4NLP-2025), pages 38–50, Albuquerque, NM. Association for Computational Linguistics.
Cite (Informal):
Prompt Tuning Can Simply Adapt Large Language Models to Text Encoders (Zhao et al., RepL4NLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/landing_page/2025.repl4nlp-1.3.pdf