@inproceedings{honda-oka-2025-exploring,
    title = "Exploring Explanations Improves the Robustness of In-Context Learning",
    author = "Honda, Ukyo  and
      Oka, Tatsushi",
    editor = "Che, Wanxiang  and
      Nabende, Joyce  and
      Shutova, Ekaterina  and
      Pilehvar, Mohammad Taher",
    booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
    month = jul,
    year = "2025",
    address = "Vienna, Austria",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/2025.acl-long.1155/",
    doi = "10.18653/v1/2025.acl-long.1155",
    pages = "23693--23714",
    ISBN = "979-8-89176-251-0",
    abstract = "In-context learning (ICL) has emerged as a successful paradigm for leveraging large language models (LLMs). However, it often struggles to generalize beyond the distribution of the provided demonstrations. A recent advancement in enhancing robustness is ICL with explanations (X-ICL), which improves prediction reliability by guiding LLMs to understand and articulate the reasoning behind correct labels. Building on this approach, we introduce an advanced framework that extends X-ICL by systematically exploring explanations for all possible labels (X$^2$-ICL), thereby enabling more comprehensive and robust decision-making. Experimental results on multiple natural language understanding datasets validate the effectiveness of X$^2$-ICL, demonstrating significantly improved robustness to out-of-distribution data compared to the existing ICL approaches."
}Markdown (Informal)
[Exploring Explanations Improves the Robustness of In-Context Learning](https://preview.aclanthology.org/ingest-emnlp/2025.acl-long.1155/) (Honda & Oka, ACL 2025)
ACL