@inproceedings{srinivasagan-ostermann-2024-hybridbert,
    title = "{H}ybrid{BERT} - Making {BERT} Pretraining More Efficient Through Hybrid Mixture of Attention Mechanisms",
    author = "Srinivasagan, Gokul  and
      Ostermann, Simon",
    editor = "Cao, Yang (Trista)  and
      Papadimitriou, Isabel  and
      Ovalle, Anaelia  and
      Zampieri, Marcos  and
      Ferraro, Francis  and
      Swayamdipta, Swabha",
    booktitle = "Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 4: Student Research Workshop)",
    month = jun,
    year = "2024",
    address = "Mexico City, Mexico",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/2024.naacl-srw.30/",
    doi = "10.18653/v1/2024.naacl-srw.30",
    pages = "285--291",
    abstract = "Pretrained transformer-based language models have produced state-of-the-art performance in most natural language understanding tasks. These models undergo two stages of training: pretraining on a huge corpus of data and fine-tuning on a specific downstream task. The pretraining phase is extremely compute-intensive and requires several high-performance computing devices like GPUs and several days or even months of training, but it is crucial for the model to capture global knowledge and also has a significant impact on the fine-tuning task. This is a major roadblock for researchers without access to sophisticated computing resources. To overcome this challenge, we propose two novel hybrid architectures called HybridBERT (HBERT), which combine self-attention and additive attention mechanisms together with sub-layer normalization. We introduce a computing budget to the pretraining phase, limiting the training time and usage to a single GPU. We show that HBERT attains twice the pretraining accuracy of a vanilla-BERT baseline. We also evaluate our proposed models on two downstream tasks, where we outperform BERT-base while accelerating inference. Moreover, we study the effect of weight initialization with a limited pretraining budget. The code and models are publicly available at: www.github.com/gokulsg/HBERT/."
}Markdown (Informal)
[HybridBERT - Making BERT Pretraining More Efficient Through Hybrid Mixture of Attention Mechanisms](https://preview.aclanthology.org/ingest-emnlp/2024.naacl-srw.30/) (Srinivasagan & Ostermann, NAACL 2024)
ACL