Mixtures of In-Context Learners
Giwon Hong, Emile Van Krieken, Edoardo Ponti, Nikolay Malkin, Pasquale Minervini
Abstract
In-context learning (ICL) adapts LLMs by providing demonstrations without fine-tuning the model parameters; however, it is very sensitive to the choice of in-context demonstrations, and processing many demonstrations can be computationally demanding. We propose Mixtures of In-Context Learners (MoICL), a novel approach that uses subsets of demonstrations to train a set of experts via ICL and learns a weighting function to merge their output distributions via gradient-based optimisation. In our experiments, we show performance improvements on 5 out of 7 classification datasets compared to a set of strong baselines (e.g., up to +13% compared to ICL and LENS). Moreover, we improve the Pareto frontier of ICL by reducing the inference time needed to achieve the same performance with fewer demonstrations. Finally, MoICL is more robust to out-of-domain (up to +11%), imbalanced (up to +49%) and perturbed demonstrations (up to +38%).- Anthology ID:
- 2025.acl-long.1277
- Volume:
- Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
- Month:
- July
- Year:
- 2025
- Address:
- Vienna, Austria
- Editors:
- Wanxiang Che, Joyce Nabende, Ekaterina Shutova, Mohammad Taher Pilehvar
- Venue:
- ACL
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 26332–26351
- Language:
- URL:
- https://preview.aclanthology.org/ingestion-acl-25/2025.acl-long.1277/
- DOI:
- Cite (ACL):
- Giwon Hong, Emile Van Krieken, Edoardo Ponti, Nikolay Malkin, and Pasquale Minervini. 2025. Mixtures of In-Context Learners. In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 26332–26351, Vienna, Austria. Association for Computational Linguistics.
- Cite (Informal):
- Mixtures of In-Context Learners (Hong et al., ACL 2025)
- PDF:
- https://preview.aclanthology.org/ingestion-acl-25/2025.acl-long.1277.pdf