@inproceedings{tai-etal-2020-exbert,
title = "ex{BERT}: Extending Pre-trained Models with Domain-specific Vocabulary Under Constrained Training Resources",
author = "Tai, Wen and
Kung, H. T. and
Dong, Xin and
Comiter, Marcus and
Kuo, Chang-Fu",
editor = "Cohn, Trevor and
He, Yulan and
Liu, Yang",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2020",
month = nov,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2020.findings-emnlp.129/",
doi = "10.18653/v1/2020.findings-emnlp.129",
pages = "1433--1439",
abstract = "We introduce exBERT, a training method to extend BERT pre-trained models from a general domain to a new pre-trained model for a specific domain with a new additive vocabulary under constrained training resources (i.e., constrained computation and data). exBERT uses a small extension module to learn to adapt an augmenting embedding for the new domain in the context of the original BERT{'}s embedding of a general vocabulary. The exBERT training method is novel in learning the new vocabulary and the extension module while keeping the weights of the original BERT model fixed, resulting in a substantial reduction in required training resources. We pre-train exBERT with biomedical articles from ClinicalKey and PubMed Central, and study its performance on biomedical downstream benchmark tasks using the MTL-Bioinformatics-2016 datasets. We demonstrate that exBERT consistently outperforms prior approaches when using limited corpus and pre-training computation resources."
}
Markdown (Informal)
[exBERT: Extending Pre-trained Models with Domain-specific Vocabulary Under Constrained Training Resources](https://preview.aclanthology.org/fix-sig-urls/2020.findings-emnlp.129/) (Tai et al., Findings 2020)
ACL