@inproceedings{lin-etal-2025-xampler,
title = "{XAMPLER}: Learning to Retrieve Cross-Lingual In-Context Examples",
author = "Lin, Peiqin and
Martins, Andre and
Schuetze, Hinrich",
editor = "Chiruzzo, Luis and
Ritter, Alan and
Wang, Lu",
booktitle = "Findings of the Association for Computational Linguistics: NAACL 2025",
month = apr,
year = "2025",
address = "Albuquerque, New Mexico",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2025.findings-naacl.221/",
pages = "3968--3977",
ISBN = "979-8-89176-195-7",
abstract = "Recent studies indicate that leveraging off-the-shelf or fine-tuned retrievers, capable of retrieving relevant in-context examples tailored to the input query, enhances few-shot in-context learning of English. However, adapting these methods to other languages, especially low-resource ones, poses challenges due to the scarcity of cross-lingual retrievers and annotated data. Thus, we introduce XAMPLER: Cross-Lingual Example Retrieval, a method tailored to tackle the challenge of cross-lingual in-context learning using only annotated English data. XAMPLER first trains a retriever based on Glot500, a multilingual small language model, using positive and negative English examples constructed from the predictions of a multilingual large language model, i.e., MaLA500. Leveraging the cross-lingual capacity of the retriever, it can directly retrieve English examples as few-shot examples for in-context learning of target languages. Experiments on two multilingual text classification benchmarks, namely SIB200 with 176 languages and MasakhaNEWS with 16 languages, demonstrate that XAMPLER substantially improves the in-context learning performance across languages."
}
Markdown (Informal)
[XAMPLER: Learning to Retrieve Cross-Lingual In-Context Examples](https://preview.aclanthology.org/fix-sig-urls/2025.findings-naacl.221/) (Lin et al., Findings 2025)
ACL