@inproceedings{schoch-ji-2025-good,
title = "The Good, the Bad, and the Debatable: A Survey on the Impacts of Data for In-Context Learning",
author = "Schoch, Stephanie and
Ji, Yangfeng",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1514/",
pages = "29786--29800",
ISBN = "979-8-89176-332-6",
abstract = "In-context learning is an emergent learning paradigm that enables an LLM to learn an unseen task by seeing a number of demonstrations in the context window. The quality of the demonstrations is of paramount importance as 1) context window size limitations restrict the number of demonstrations that can be presented to the model, and 2) the model must identify the task and potentially learn new, unseen input-output mappings from the limited demonstration set. An increasing body of work has also shown the sensitivity of predictions to perturbations on the demonstration set. Given this importance, this work presents a survey on the current literature pertaining to the relationship between data and in-context learning. We present our survey in three parts: the ``good'' {--} qualities that are desirable when selecting demonstrations, the ``bad'' {--} qualities of demonstrations that can negatively impact the model, as well as issues that can arise in presenting demonstrations, and the ``debatable'' {--} qualities of demonstrations with mixed results or factors modulating data impacts."
}Markdown (Informal)
[The Good, the Bad, and the Debatable: A Survey on the Impacts of Data for In-Context Learning](https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1514/) (Schoch & Ji, EMNLP 2025)
ACL