@inproceedings{galimzianova-sanochkin-2024-efficient,
title = "Efficient Active Learning with Adapters",
author = "Galimzianova, Daria and
Sanochkin, Leonid",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2024",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/Add-Cong-Liu-Florida-Atlantic-University-author-id/2024.findings-emnlp.840/",
doi = "10.18653/v1/2024.findings-emnlp.840",
pages = "14374--14383",
abstract = "One of the main obstacles for deploying Active Learning (AL) in practical NLP tasks is high computational cost of modern deep learning models. This issue can be partially mitigated by applying lightweight models as an acquisition model, but it can lead to the acquisition-successor mismatch (ASM) problem. Previous works show that the ASM problem can be partially alleviated by using distilled versions of a successor models as acquisition ones. However, distilled versions of pretrained models are not always available. Also, the exact pipeline of model distillation that does not lead to the ASM problem is not clear. To address these issues, we propose to use adapters as an alternative to full fine-tuning for acquisition model training. Since adapters are lightweight, this approach reduces the training cost of the model. We provide empirical evidence that it does not cause the ASM problem and can help to deploy active learning in practical NLP tasks."
}
Markdown (Informal)
[Efficient Active Learning with Adapters](https://preview.aclanthology.org/Add-Cong-Liu-Florida-Atlantic-University-author-id/2024.findings-emnlp.840/) (Galimzianova & Sanochkin, Findings 2024)
ACL
- Daria Galimzianova and Leonid Sanochkin. 2024. Efficient Active Learning with Adapters. In Findings of the Association for Computational Linguistics: EMNLP 2024, pages 14374–14383, Miami, Florida, USA. Association for Computational Linguistics.