Vincent Zhao
2023
RARR: Researching and Revising What Language Models Say, Using Language Models
Luyu Gao
|
Zhuyun Dai
|
Panupong Pasupat
|
Anthony Chen
|
Arun Tejasvi Chaganty
|
Yicheng Fan
|
Vincent Zhao
|
Ni Lao
|
Hongrae Lee
|
Da-Cheng Juan
|
Kelvin Guu
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Language models (LMs) now excel at many tasks such as question answering, reasoning, and dialog. However, they sometimes generate unsupported or misleading content. A user cannot easily determine whether their outputs are trustworthy or not, because most LMs do not have any built-in mechanism for attribution to external evidence. To enable attribution while still preserving all the powerful advantages of recent generation models, we propose RARR (Retrofit Attribution using Research and Revision), a system that 1) automatically finds attribution for the output of any text generation model, and 2) post-edits the output to fix unsupported content while preserving the original output as much as possible. When applied to the output of several state-of-the-art LMs on a diverse set of generation tasks, we find that RARR significantly improves attribution while otherwise preserving the original input to a much greater degree than previously explored edit models. Furthermore, the implementation of RARR requires only a handful of training examples, a large language model, and standard web search.
2022
Large Dual Encoders Are Generalizable Retrievers
Jianmo Ni
|
Chen Qu
|
Jing Lu
|
Zhuyun Dai
|
Gustavo Hernandez Abrego
|
Ji Ma
|
Vincent Zhao
|
Yi Luan
|
Keith Hall
|
Ming-Wei Chang
|
Yinfei Yang
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing
It has been shown that dual encoders trained on one domain often fail to generalize to other domains for retrieval tasks. One widespread belief is that the bottleneck layer of a dual encoder, where the final score is simply a dot-product between a query vector and a passage vector, is too limited compared to models with fine-grained interactions between the query and the passage. In this paper, we challenge this belief by scaling up the size of the dual encoder model while keeping the bottleneck layer as a single dot-product with a fixed size. With multi-stage training, scaling up the model size brings significant improvement on a variety of retrieval tasks, especially for out-of-domain generalization. We further analyze the impact of the bottleneck layer and demonstrate diminishing improvement when scaling up the embedding size. Experimental results show that our dual encoders, Generalizable T5-based dense Retrievers (GTR), outperform previous sparse and dense retrievers on the BEIR dataset significantly. Most surprisingly, our ablation study finds that GTR is very data efficient, as it only needs 10% of MS Marco supervised data to match the out-of-domain performance of using all supervised data.
Search
Co-authors
- Zhuyun Dai 2
- Jianmo Ni 1
- Chen Qu 1
- Jing Lu 1
- Gustavo Hernandez Abrego 1
- show all...