Variance-reduced First-order Meta-learning for Natural Language Processing Tasks

Lingxiao Wang, Kevin Huang, Tengyu Ma, Quanquan Gu, Jing Huang


Abstract
First-order meta-learning algorithms have been widely used in practice to learn initial model parameters that can be quickly adapted to new tasks due to their efficiency and effectiveness. However, existing studies find that meta-learner can overfit to some specific adaptation when we have heterogeneous tasks, leading to significantly degraded performance. In Natural Language Processing (NLP) applications, datasets are often diverse and each task has its unique characteristics. Therefore, to address the overfitting issue when applying first-order meta-learning to NLP applications, we propose to reduce the variance of the gradient estimator used in task adaptation. To this end, we develop a variance-reduced first-order meta-learning algorithm. The core of our algorithm is to introduce a novel variance reduction term to the gradient estimation when performing the task adaptation. Experiments on two NLP applications: few-shot text classification and multi-domain dialog state tracking demonstrate the superior performance of our proposed method.
Anthology ID:
2021.naacl-main.206
Volume:
Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
Month:
June
Year:
2021
Address:
Online
Editors:
Kristina Toutanova, Anna Rumshisky, Luke Zettlemoyer, Dilek Hakkani-Tur, Iz Beltagy, Steven Bethard, Ryan Cotterell, Tanmoy Chakraborty, Yichao Zhou
Venue:
NAACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
2609–2615
Language:
URL:
https://aclanthology.org/2021.naacl-main.206
DOI:
10.18653/v1/2021.naacl-main.206
Bibkey:
Cite (ACL):
Lingxiao Wang, Kevin Huang, Tengyu Ma, Quanquan Gu, and Jing Huang. 2021. Variance-reduced First-order Meta-learning for Natural Language Processing Tasks. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 2609–2615, Online. Association for Computational Linguistics.
Cite (Informal):
Variance-reduced First-order Meta-learning for Natural Language Processing Tasks (Wang et al., NAACL 2021)
Copy Citation:
PDF:
https://preview.aclanthology.org/nschneid-patch-5/2021.naacl-main.206.pdf
Video:
 https://preview.aclanthology.org/nschneid-patch-5/2021.naacl-main.206.mp4
Data
FewRel