Confidence Calibration in Large Language Model-Based Entity Matching
Iris Kamsteeg, Juan Cardenas-Cartagena, Floris van Beers, Tsegaye Misikir Tashu, Matias Valdenegro-Toro
Abstract
This research aims to explore the intersection of Large Language Models and confidence calibration in Entity Matching. To this end, we perform an empirical study to compare baseline RoBERTa confidences for an Entity Matching task against confidences that are calibrated using Temperature Scaling, Monte Carlo Dropout and Ensembles. We use the Abt-Buy, DBLP-ACM, iTunes-Amazon and Company datasets. The findings indicate that the proposed modified RoBERTa model exhibits a slight overconfidence, with Expected Calibration Error scores ranging from 0.0043 to 0.0552 across datasets. We find that this overconfidence can be mitigated using Temperature Scaling, reducing Expected Calibration Error scores by up to 23.83%.- Anthology ID:
- 2025.uncertainlp-main.12
- Volume:
- Proceedings of the 2nd Workshop on Uncertainty-Aware NLP (UncertaiNLP 2025)
- Month:
- November
- Year:
- 2025
- Address:
- Suzhou, China
- Editor:
- Noidea Noidea
- Venues:
- UncertaiNLP | WS
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 120–137
- Language:
- URL:
- https://preview.aclanthology.org/ingest-emnlp/2025.uncertainlp-main.12/
- DOI:
- Cite (ACL):
- Iris Kamsteeg, Juan Cardenas-Cartagena, Floris van Beers, Tsegaye Misikir Tashu, and Matias Valdenegro-Toro. 2025. Confidence Calibration in Large Language Model-Based Entity Matching. In Proceedings of the 2nd Workshop on Uncertainty-Aware NLP (UncertaiNLP 2025), pages 120–137, Suzhou, China. Association for Computational Linguistics.
- Cite (Informal):
- Confidence Calibration in Large Language Model-Based Entity Matching (Kamsteeg et al., UncertaiNLP 2025)
- PDF:
- https://preview.aclanthology.org/ingest-emnlp/2025.uncertainlp-main.12.pdf