@inproceedings{kim-kim-2024-tuning,
title = "{IT}-Tuning : Parameter Efficient Information Token Tuning for Language Model",
author = "Kim, Jungu and
Kim, Hyeoncheol",
editor = "Zhao, Chen and
Mosbach, Marius and
Atanasova, Pepa and
Goldfarb-Tarrent, Seraphina and
Hase, Peter and
Hosseini, Arian and
Elbayad, Maha and
Pezzelle, Sandro and
Mozes, Maximilian",
booktitle = "Proceedings of the 9th Workshop on Representation Learning for NLP (RepL4NLP-2024)",
month = aug,
year = "2024",
address = "Bangkok, Thailand",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/landing_page/2024.repl4nlp-1.6/",
pages = "58--68",
abstract = "Recently, language models have demonstrated exceptional performance compared to their predecessors. In this context, attention mechanisms and pre-training significantly contribute to the enhanced performance of modern language models. Additionally, a continuously increasing number of parameters plays a crucial role in these advancements . However, an increase in the number of parameters significantly increases the GPU memory and training time required during fine-tuning of language models, this makes fine-tuning infeasible in environments with limited computing resources. Furthermore, after fine-tuning, the storage space required for deployment increases proportionally with the number of tasks, making it challenging to deploy devices with limited storage capacities. In this study, we propose IT-Tuning, a Parameter Efficient Fine-Tuning method that introduces a new concept called information tokens to address these issues."
}
Markdown (Informal)
[IT-Tuning : Parameter Efficient Information Token Tuning for Language Model](https://preview.aclanthology.org/landing_page/2024.repl4nlp-1.6/) (Kim & Kim, RepL4NLP 2024)
ACL