Continual-learning for Modelling Low-Resource Languages from Large Language Models

Santosh Srinath K, Mudit Somani, Varun Reddy Padala, Prajna Upadhyay, Abhijit Das


Abstract
Modelling a language model for a multi-lingual scenario includes several potential challenges, among which catastrophic forgetting is the major challenge. For example, small language models (SLM) built for low-resource languages by adapting large language models (LLMs) pose the challenge of catastrophic forgetting. This work proposes to employ a continual learning strategy using parts-of-speech (POS)-based code-switching along with a replay adapter strategy to mitigate the identified gap of catastrophic forgetting while training SLM from LLM. Experiments conducted on vision language tasks such as visual question answering and language modelling task exhibits the success of the proposed architecture.
Anthology ID:
2026.eacl-long.293
Volume:
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Month:
March
Year:
2026
Address:
Rabat, Morocco
Editors:
Vera Demberg, Kentaro Inui, Lluís Marquez
Venue:
EACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
6258–6275
Language:
URL:
https://preview.aclanthology.org/ingest-eacl/2026.eacl-long.293/
DOI:
Bibkey:
Cite (ACL):
Santosh Srinath K, Mudit Somani, Varun Reddy Padala, Prajna Upadhyay, and Abhijit Das. 2026. Continual-learning for Modelling Low-Resource Languages from Large Language Models. In Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers), pages 6258–6275, Rabat, Morocco. Association for Computational Linguistics.
Cite (Informal):
Continual-learning for Modelling Low-Resource Languages from Large Language Models (K et al., EACL 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-eacl/2026.eacl-long.293.pdf