@inproceedings{cheng-etal-2023-task,
title = "Task-Aware Specialization for Efficient and Robust Dense Retrieval for Open-Domain Question Answering",
author = "Cheng, Hao and
Fang, Hao and
Liu, Xiaodong and
Gao, Jianfeng",
editor = "Rogers, Anna and
Boyd-Graber, Jordan and
Okazaki, Naoaki",
booktitle = "Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)",
month = jul,
year = "2023",
address = "Toronto, Canada",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.acl-short.159/",
doi = "10.18653/v1/2023.acl-short.159",
pages = "1864--1875",
abstract = "Given its effectiveness on knowledge-intensive natural language processing tasks, dense retrieval models have become increasingly popular. Specifically, the de-facto architecture for open-domain question answering uses two isomorphic encoders that are initialized from the same pretrained model but separately parameterized for questions and passages. This biencoder architecture is parameter-inefficient in that there is no parameter sharing between encoders. Further, recent studies show that such dense retrievers underperform BM25 in various settings. We thus propose a new architecture, Task-Aware Specialization for dEnse Retrieval (TASER), which enables parameter sharing by interleaving shared and specialized blocks in a single encoder. Our experiments on five question answering datasets show that TASER can achieve superior accuracy, surpassing BM25, while using about 60{\%} of the parameters as bi-encoder dense retrievers. In out-of-domain evaluations, TASER is also empirically more robust than bi-encoder dense retrievers. Our code is available at \url{https://github.com/microsoft/taser}."
}
Markdown (Informal)
[Task-Aware Specialization for Efficient and Robust Dense Retrieval for Open-Domain Question Answering](https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.acl-short.159/) (Cheng et al., ACL 2023)
ACL