@inproceedings{peinelt-etal-2021-gibert-enhancing,
title = "{G}i{BERT}: Enhancing {BERT} with Linguistic Information using a Lightweight Gated Injection Method",
author = "Peinelt, Nicole and
Rei, Marek and
Liakata, Maria",
editor = "Moens, Marie-Francine and
Huang, Xuanjing and
Specia, Lucia and
Yih, Scott Wen-tau",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2021",
month = nov,
year = "2021",
address = "Punta Cana, Dominican Republic",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2021.findings-emnlp.200/",
doi = "10.18653/v1/2021.findings-emnlp.200",
pages = "2322--2336",
abstract = "Large pre-trained language models such as BERT have been the driving force behind recent improvements across many NLP tasks. However, BERT is only trained to predict missing words {--} either through masking or next sentence prediction {--} and has no knowledge of lexical, syntactic or semantic information beyond what it picks up through unsupervised pre-training. We propose a novel method to explicitly inject linguistic information in the form of word embeddings into any layer of a pre-trained BERT. When injecting counter-fitted and dependency-based embeddings, the performance improvements on multiple semantic similarity datasets indicate that such information is beneficial and currently missing from the original model. Our qualitative analysis shows that counter-fitted embedding injection is particularly beneficial, with notable improvements on examples that require synonym resolution."
}
Markdown (Informal)
[GiBERT: Enhancing BERT with Linguistic Information using a Lightweight Gated Injection Method](https://preview.aclanthology.org/add-emnlp-2024-awards/2021.findings-emnlp.200/) (Peinelt et al., Findings 2021)
ACL