@inproceedings{bao-etal-2021-span-fine,
title = "Span Fine-tuning for Pre-trained Language Models",
author = "Bao, Rongzhou and
Zhang, Zhuosheng and
Zhao, Hai",
editor = "Moens, Marie-Francine and
Huang, Xuanjing and
Specia, Lucia and
Yih, Scott Wen-tau",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2021",
month = nov,
year = "2021",
address = "Punta Cana, Dominican Republic",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2021.findings-emnlp.169/",
doi = "10.18653/v1/2021.findings-emnlp.169",
pages = "1970--1979",
abstract = "Pre-trained language models (PrLM) have to carefully manage input units when training on a very large text with a vocabulary consisting of millions of words. Previous works have shown that incorporating span-level information over consecutive words in pre-training could further improve the performance of PrLMs. However, given that span-level clues are introduced and fixed in pre-training, previous methods are time-consuming and lack of flexibility. To alleviate the inconvenience, this paper presents a novel span fine-tuning method for PrLMs, which facilitates the span setting to be adaptively determined by specific downstream tasks during the fine-tuning phase. In detail, any sentences processed by the PrLM will be segmented into multiple spans according to a pre-sampled dictionary. Then the segmentation information will be sent through a hierarchical CNN module together with the representation outputs of the PrLM and ultimately generate a span-enhanced representation. Experiments on GLUE benchmark show that the proposed span fine-tuning method significantly enhances the PrLM, and at the same time, offer more flexibility in an efficient way."
}
Markdown (Informal)
[Span Fine-tuning for Pre-trained Language Models](https://preview.aclanthology.org/add-emnlp-2024-awards/2021.findings-emnlp.169/) (Bao et al., Findings 2021)
ACL
- Rongzhou Bao, Zhuosheng Zhang, and Hai Zhao. 2021. Span Fine-tuning for Pre-trained Language Models. In Findings of the Association for Computational Linguistics: EMNLP 2021, pages 1970–1979, Punta Cana, Dominican Republic. Association for Computational Linguistics.