Few-Shot Regularization to Tackle Catastrophic Forgetting in Multilingual Machine Translation

Salvador Carrión-Ponz , Francisco Casacuberta


Abstract
Increasing the number of tasks supported by a machine learning model without forgetting previously learned tasks is the goal of any lifelong learning system. In this work, we study how to mitigate the effects of the catastrophic forgetting problem to sequentially train a multilingual neural machine translation model using minimal past information. First, we describe the catastrophic forgetting phenomenon as a function of the number of tasks learned (language pairs) and the ratios of past data used during the learning of the new task. Next, we explore the importance of applying oversampling strategies for scenarios where only minimal amounts of past data are available. Finally, we derive a new loss function that minimizes the forgetting of previously learned tasks by actively re-weighting past samples and penalizing weights that deviate too much from the original model. Our work suggests that by using minimal amounts of past data and a simple regularization function, we can significantly mitigate the effects of the catastrophic forgetting phenomenon without increasing the computational costs.
Anthology ID:
2022.amta-research.14
Volume:
Proceedings of the 15th biennial conference of the Association for Machine Translation in the Americas (Volume 1: Research Track)
Month:
September
Year:
2022
Address:
Orlando, USA
Venue:
AMTA
SIG:
Publisher:
Association for Machine Translation in the Americas
Note:
Pages:
188–199
Language:
URL:
https://aclanthology.org/2022.amta-research.14
DOI:
Bibkey:
Cite (ACL):
Salvador Carrión-Ponz and Francisco Casacuberta. 2022. Few-Shot Regularization to Tackle Catastrophic Forgetting in Multilingual Machine Translation. In Proceedings of the 15th biennial conference of the Association for Machine Translation in the Americas (Volume 1: Research Track), pages 188–199, Orlando, USA. Association for Machine Translation in the Americas.
Cite (Informal):
Few-Shot Regularization to Tackle Catastrophic Forgetting in Multilingual Machine Translation ( & Casacuberta, AMTA 2022)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingestion-script-update/2022.amta-research.14.pdf