Beyond Averages: Learning with Annotator Disagreement in STS

Alejandro Benito-Santos, Adrian Ghajari


Abstract
This work investigates capturing and modeling disagreement in Semantic Textual Similarity (STS), where sentence pairs are assigned ordinal similarity labels (0–5). Conventional STS systems average multiple annotator scores and focus on a single numeric estimate, overlooking label dispersion. By leveraging the disaggregated SemEval-2015 dataset (Soft-STS-15), this paper proposes and compares two disagreement-aware strategies that treat STS as an ordinal distribution prediction problem: a lightweight truncated Gaussian head for standard regression models, and a cross-encoder trained with a distance-aware objective, refined with temperature scaling. Results show improved performance in distance-based metrics, with the calibrated soft-label model proving best overall and notably more accurate on the most ambiguous pairs. This demonstrates that modeling disagreement benefits both calibration and ranking accuracy, highlighting the value of retaining and modeling full annotation distributions rather than collapsing them to a single mean label.
Anthology ID:
2025.emnlp-main.1800
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
35539–35545
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1800/
DOI:
Bibkey:
Cite (ACL):
Alejandro Benito-Santos and Adrian Ghajari. 2025. Beyond Averages: Learning with Annotator Disagreement in STS. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 35539–35545, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Beyond Averages: Learning with Annotator Disagreement in STS (Benito-Santos & Ghajari, EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1800.pdf
Checklist:
 2025.emnlp-main.1800.checklist.pdf