@inproceedings{jianbang-etal-2023-adder,
title = "Adder Encoder for Pre-trained Language Model",
author = "Jianbang, Ding and
Suiyun, Zhang and
Linlin, Li",
editor = "Sun, Maosong and
Qin, Bing and
Qiu, Xipeng and
Jiang, Jing and
Han, Xianpei",
booktitle = "Proceedings of the 22nd Chinese National Conference on Computational Linguistics",
month = aug,
year = "2023",
address = "Harbin, China",
publisher = "Chinese Information Processing Society of China",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.ccl-1.76/",
pages = "898--905",
language = "eng",
abstract = "{\textquotedblleft}BERT, a pre-trained language model entirely based on attention, has proven to be highly per-formant for many natural language understanding tasks. However, pre-trained language mod-els (PLMs) are often computationally expensive and can hardly be implemented with limitedresources. To reduce energy burden, we introduce adder operations into the Transformer en-coder and propose a novel AdderBERT with powerful representation capability. Moreover, weadopt mapping-based distillation to further improve its energy efficiency with an assured perfor-mance. Empirical results demonstrate that AddderBERT6 achieves highly competitive perfor-mance against that of its teacher BERTBASE on the GLUE benchmark while obtaining a 4.9xreduction in energy consumption.{\textquotedblright}"
}
Markdown (Informal)
[Adder Encoder for Pre-trained Language Model](https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.ccl-1.76/) (Jianbang et al., CCL 2023)
ACL
- Ding Jianbang, Zhang Suiyun, and Li Linlin. 2023. Adder Encoder for Pre-trained Language Model. In Proceedings of the 22nd Chinese National Conference on Computational Linguistics, pages 898–905, Harbin, China. Chinese Information Processing Society of China.