How Does Data Corruption Affect Natural Language Understanding Models? A Study on GLUE datasets
Aarne Talman, Marianna Apidianaki, Stergios Chatzikyriakidis, Jörg Tiedemann
Abstract
A central question in natural language understanding (NLU) research is whether high performance demonstrates the models’ strong reasoning capabilities. We present an extensive series of controlled experiments where pre-trained language models are exposed to data that have undergone specific corruption transformations. These involve removing instances of specific word classes and often lead to non-sensical sentences. Our results show that performance remains high on most GLUE tasks when the models are fine-tuned or tested on corrupted data, suggesting that they leverage other cues for prediction even in non-sensical contexts. Our proposed data transformations can be used to assess the extent to which a specific dataset constitutes a proper testbed for evaluating models’ language understanding capabilities.- Anthology ID:
- 2022.starsem-1.20
- Volume:
- Proceedings of the 11th Joint Conference on Lexical and Computational Semantics
- Month:
- July
- Year:
- 2022
- Address:
- Seattle, Washington
- Editors:
- Vivi Nastase, Ellie Pavlick, Mohammad Taher Pilehvar, Jose Camacho-Collados, Alessandro Raganato
- Venue:
- *SEM
- SIG:
- SIGLEX
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 226–233
- Language:
- URL:
- https://aclanthology.org/2022.starsem-1.20
- DOI:
- 10.18653/v1/2022.starsem-1.20
- Cite (ACL):
- Aarne Talman, Marianna Apidianaki, Stergios Chatzikyriakidis, and Jörg Tiedemann. 2022. How Does Data Corruption Affect Natural Language Understanding Models? A Study on GLUE datasets. In Proceedings of the 11th Joint Conference on Lexical and Computational Semantics, pages 226–233, Seattle, Washington. Association for Computational Linguistics.
- Cite (Informal):
- How Does Data Corruption Affect Natural Language Understanding Models? A Study on GLUE datasets (Talman et al., *SEM 2022)
- PDF:
- https://preview.aclanthology.org/ingest-acl-2023-videos/2022.starsem-1.20.pdf
- Code
- helsinki-nlp/nlu-dataset-diagnostics
- Data
- CoLA, GLUE, MRPC, MultiNLI, QNLI, SST, SST-2