@inproceedings{gao-callan-2021-condenser,
title = "Condenser: a Pre-training Architecture for Dense Retrieval",
author = "Gao, Luyu and
Callan, Jamie",
editor = "Moens, Marie-Francine and
Huang, Xuanjing and
Specia, Lucia and
Yih, Scott Wen-tau",
booktitle = "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2021",
address = "Online and Punta Cana, Dominican Republic",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/moar-dois/2021.emnlp-main.75/",
doi = "10.18653/v1/2021.emnlp-main.75",
pages = "981--993",
abstract = "Pre-trained Transformer language models (LM) have become go-to text representation encoders. Prior research fine-tunes deep LMs to encode text sequences such as sentences and passages into single dense vector representations for efficient text comparison and retrieval. However, dense encoders require a lot of data and sophisticated techniques to effectively train and suffer in low data situations. This paper finds a key reason is that standard LMs' internal attention structure is not ready-to-use for dense encoders, which needs to aggregate text information into the dense representation. We propose to pre-train towards dense encoder with a novel Transformer architecture, Condenser, where LM prediction CONditions on DENSE Representation. Our experiments show Condenser improves over standard LM by large margins on various text retrieval and similarity tasks."
}
Markdown (Informal)
[Condenser: a Pre-training Architecture for Dense Retrieval](https://preview.aclanthology.org/moar-dois/2021.emnlp-main.75/) (Gao & Callan, EMNLP 2021)
ACL