@inproceedings{waldis-etal-2024-dive,
title = "Dive into the Chasm: Probing the Gap between In- and Cross-Topic Generalization",
author = "Waldis, Andreas and
Hou, Yufang and
Gurevych, Iryna",
editor = "Graham, Yvette and
Purver, Matthew",
booktitle = "Findings of the Association for Computational Linguistics: EACL 2024",
month = mar,
year = "2024",
address = "St. Julian{'}s, Malta",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.findings-eacl.146/",
pages = "2197--2214",
abstract = "Pre-trained language models (PLMs) perform well in In-Topic setups, where training and testing data come from the same topics. However, they face challenges in Cross-Topic scenarios where testing data is derived from distinct topics. This paper analyzes various PLMs with three probing-based experiments to better understand the reasons behind such generalization gaps. For the first time, we demonstrate that the extent of these generalization gaps and the sensitivity to token-level interventions vary significantly across PLMs. By evaluating large language models (LLMs), we show the usefulness of our analysis for these recent models. Overall, we observe diverse pre-training objectives and architectural regularization contribute to more robust PLMs and mitigate generalization gaps. Our research contributes to a deeper understanding and comparison of language models across different generalization scenarios."
}
Markdown (Informal)
[Dive into the Chasm: Probing the Gap between In- and Cross-Topic Generalization](https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.findings-eacl.146/) (Waldis et al., Findings 2024)
ACL