SEE: Continual Fine-tuning with Sequential Ensemble of Experts

Zhilin Wang, Yafu Li, Xiaoye Qu, Yu Cheng


Abstract
Continual fine-tuning of large language models (LLMs) suffers from catastrophic forgetting. Rehearsal-based methods mitigate this problem by retaining a small set of old data. Nevertheless, they still suffer inevitable performance loss. Although training separate experts for each task can help prevent forgetting, effectively assembling them remains a challenge. Some approaches use routers to assign tasks to experts, but in continual learning, they often require retraining for optimal performance. To address these challenges, we introduce the Sequential Ensemble of Experts (SEE) framework. SEE removes the need for an additional router, allowing each expert to independently decide whether a query should be handled. The framework employs distributed routing, and during continual fine-tuning, SEE only requires the training of new experts for incoming tasks, rather than retraining the entire system. Experiments reveal that the SEE outperforms prior approaches, including multi-task learning, in continual fine-tuning. It also demonstrates remarkable generalization ability, as the expert can effectively identify out-of-distribution queries, which can then be directed to a more generalized model for resolution. This work highlights the promising potential of integrating routing and response mechanisms within each expert, paving the way for the future of distributed model ensembling.
Anthology ID:
2025.findings-acl.387
Volume:
Findings of the Association for Computational Linguistics: ACL 2025
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Wanxiang Che, Joyce Nabende, Ekaterina Shutova, Mohammad Taher Pilehvar
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
7418–7432
Language:
URL:
https://preview.aclanthology.org/display_plenaries/2025.findings-acl.387/
DOI:
Bibkey:
Cite (ACL):
Zhilin Wang, Yafu Li, Xiaoye Qu, and Yu Cheng. 2025. SEE: Continual Fine-tuning with Sequential Ensemble of Experts. In Findings of the Association for Computational Linguistics: ACL 2025, pages 7418–7432, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
SEE: Continual Fine-tuning with Sequential Ensemble of Experts (Wang et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/display_plenaries/2025.findings-acl.387.pdf