@inproceedings{delbari-pilehvar-2025-beyond,
title = "Beyond Accuracy: Revisiting Out-of-Distribution Generalization in {NLI} Models",
author = "Delbari, Zahra and
Pilehvar, Mohammad Taher",
editor = "Boleda, Gemma and
Roth, Michael",
booktitle = "Proceedings of the 29th Conference on Computational Natural Language Learning",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/acl25-workshop-ingestion/2025.conll-1.36/",
pages = "557--570",
ISBN = "979-8-89176-271-8",
abstract = "This study investigates how well discriminative transformers generalize in Natural Language Inference (NLI) tasks. We specifically focus on a well-studied bias in this task: the tendency of models to rely on superficial features and dataset biases rather than a true understanding of language. We argue that the performance differences observed between training and analysis datasets do not necessarily indicate a lack of knowledge within the model. Instead, the gap often points to a misalignment between the decision boundaries of the classifier head and the representations learned by the encoder for the analysis samples. By investigating the representation space of NLI models across different analysis datasets, we demonstrate that even when the accuracy is nearly random in some settings, still samples from opposing classes remain almost perfectly linearly separable in the encoder{'}s representation space. This suggests that, although the classifier head may fail on analysis data, the encoder still generalizes and encodes representations that allow for effective discrimination between NLI classes."
}
Markdown (Informal)
[Beyond Accuracy: Revisiting Out-of-Distribution Generalization in NLI Models](https://preview.aclanthology.org/acl25-workshop-ingestion/2025.conll-1.36/) (Delbari & Pilehvar, CoNLL 2025)
ACL