Alberto Bernacchia
2023
Hierarchical Representations in Dense Passage Retrieval for Question-Answering
Philipp Ennen
|
Federica Freddi
|
Chyi-Jiunn Lin
|
Po-Nien Kung
|
RenChu Wang
|
Chien-Yi Yang
|
Da-shan Shiu
|
Alberto Bernacchia
Proceedings of the Sixth Fact Extraction and VERification Workshop (FEVER)
An approach to improve question-answering performance is to retrieve accompanying information that contains factual evidence matching the question. These retrieved documents are then fed into a reader that generates an answer. A commonly applied retriever is dense passage retrieval. In this retriever, the output of a transformer neural network is used to query a knowledge database for matching documents. Inspired by the observation that different layers of a transformer network provide rich representations with different levels of abstraction, we hypothesize that useful queries can be generated not only at the output layer, but at every layer of a transformer network, and that the hidden representations of different layers may combine to improve the fetched documents for reader performance. Our novel approach integrates retrieval into each layer of a transformer network, exploiting the hierarchical representations of the input question. We show that our technique outperforms prior work on downstream tasks such as question answering, demonstrating the effectiveness of our approach.
2021
Cross-Lingual Transfer with MAML on Trees
Jezabel Garcia
|
Federica Freddi
|
Jamie McGowan
|
Tim Nieradzik
|
Feng-Ting Liao
|
Ye Tian
|
Da-shan Shiu
|
Alberto Bernacchia
Proceedings of the Second Workshop on Domain Adaptation for NLP
In meta-learning, the knowledge learned from previous tasks is transferred to new ones, but this transfer only works if tasks are related. Sharing information between unrelated tasks might hurt performance, and it is unclear how to transfer knowledge across tasks that have a hierarchical structure. Our research extends a meta-learning model, MAML, by exploiting hierarchical task relationships. Our algorithm, TreeMAML, adapts the model to each task with a few gradient steps, but the adaptation follows the hierarchical tree structure: in each step, gradients are pooled across tasks clusters and subsequent steps follow down the tree. We also implement a clustering algorithm that generates the tasks tree without previous knowledge of the task structure, allowing us to make use of implicit relationships between the tasks. We show that TreeMAML successfully trains natural language processing models for cross-lingual Natural Language Inference by taking advantage of the language phylogenetic tree. This result is useful since most languages in the world are under-resourced and the improvement on cross-lingual transfer allows the internationalization of NLP models.
Search
Co-authors
- Federica Freddi 2
- Da-shan Shiu 2
- Jezabel Garcia 1
- Jamie McGowan 1
- Tim Nieradzik 1
- show all...