Yuefeng Zhan
2022
Snapshot-Guided Domain Adaptation for ELECTRA
Daixuan Cheng
|
Shaohan Huang
|
Jianfeng Liu
|
Yuefeng Zhan
|
Hao Sun
|
Furu Wei
|
Denvy Deng
|
Qi Zhang
Findings of the Association for Computational Linguistics: EMNLP 2022
Discriminative pre-trained language models, such as ELECTRA, have achieved promising performances in a variety of general tasks. However, these generic pre-trained models struggle to capture domain-specific knowledge of domain-related tasks. In this work, we propose a novel domain-adaptation method for ELECTRA, which can dynamically select domain-specific tokens and guide the discriminator to emphasize them, without introducing new training parameters. We show that by re-weighting the losses of domain-specific tokens, ELECTRA can be effectively adapted to different domains. The experimental results in both computer science and biomedical domains show that the proposed method can achieve state-of-the-art results on the domain-related tasks.
Search
Co-authors
- Daixuan Cheng 1
- Shaohan Huang 1
- Jianfeng Liu 1
- Hao Sun 1
- Furu Wei 1
- show all...