@inproceedings{zhang-etal-2024-selective,
title = "Selective Prefix Tuning for Pre-trained Language Models",
author = "Zhang, Hongyi and
Li, Zuchao and
Wang, Ping and
Zhao, Hai",
editor = "Ku, Lun-Wei and
Martins, Andre and
Srikumar, Vivek",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2024",
month = aug,
year = "2024",
address = "Bangkok, Thailand",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.findings-acl.164/",
doi = "10.18653/v1/2024.findings-acl.164",
pages = "2806--2813",
abstract = "The prevalent approach for optimizing pre-trained language models in downstream tasks is fine-tuning. However, it is both time-consuming and memory-inefficient. In response, a more efficient method called Prefix Tuning, which insert learnable vectors into each Transformer layers, has been proposed and proven effective. Recent investigations reveal that prefix tokens carry context-specific information, prompting the hypothesis that enhancing their specialization can improve model performance. To address this, we propose Selective Prefix Tuning (SPT), integrating a selective mechanism inspired by selective self-attention. Additionally, we introduce Selective Loss (SL) to encourage diversity in prefix tokens. Extensive experiments validate the effectiveness of SPT in sentence and token classification tasks. We contribute insight into understanding the role of prefix in model adaptation."
}
Markdown (Informal)
[Selective Prefix Tuning for Pre-trained Language Models](https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.findings-acl.164/) (Zhang et al., Findings 2024)
ACL