Abstract
The task of natural language inference (NLI), to decide if a hypothesis entails or contradicts a premise, received considerable attention in recent years. All competitive systems build on top of contextualized representations and make use of transformer architectures for learning an NLI model. When somebody is faced with a particular NLI task, they need to select the best model that is available. This is a time-consuming and resource-intense endeavour. To solve this practical problem, we propose a simple method for predicting the performance without actually fine-tuning the model. We do this by testing how well the pre-trained models perform on the aNLI task when just comparing sentence embeddings with cosine similarity to what kind of performance is achieved when training a classifier on top of these embeddings. We show that the accuracy of the cosine similarity approach correlates strongly with the accuracy of the classification approach with a Pearson correlation coefficient of 0.65. Since the similarity is orders of magnitude faster to compute on a given dataset (less than a minute vs. hours), our method can lead to significant time savings in the process of model selection.- Anthology ID:
- 2022.naacl-main.441
- Volume:
- Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
- Month:
- July
- Year:
- 2022
- Address:
- Seattle, United States
- Editors:
- Marine Carpuat, Marie-Catherine de Marneffe, Ivan Vladimir Meza Ruiz
- Venue:
- NAACL
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 6031–6037
- Language:
- URL:
- https://aclanthology.org/2022.naacl-main.441
- DOI:
- 10.18653/v1/2022.naacl-main.441
- Cite (ACL):
- Emīls Kadiķis, Vaibhav Srivastav, and Roman Klinger. 2022. Embarrassingly Simple Performance Prediction for Abductive Natural Language Inference. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 6031–6037, Seattle, United States. Association for Computational Linguistics.
- Cite (Informal):
- Embarrassingly Simple Performance Prediction for Abductive Natural Language Inference (Kadiķis et al., NAACL 2022)
- PDF:
- https://preview.aclanthology.org/ingest-acl-2023-videos/2022.naacl-main.441.pdf
- Code
- Vaibhavs10/anli-performance-prediction
- Data
- ART Dataset