On the Interplay between Human Label Variation and Model Fairness

Kemal Kurniawan, Meladel Mistica, Timothy Baldwin, Jey Han Lau


Abstract
The impact of human label variation (HLV) on model fairness is an unexplored topic. This paper examines the interplay by comparing training on majority-vote labels with a range of HLV methods. Our experiments show that without explicit debiasing, HLV training methods have a positive impact on fairness under certain configurations.
Anthology ID:
2026.findings-eacl.50
Volume:
Findings of the Association for Computational Linguistics: EACL 2026
Month:
March
Year:
2026
Address:
Rabat, Morocco
Editors:
Vera Demberg, Kentaro Inui, Lluís Marquez
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
967–976
Language:
URL:
https://preview.aclanthology.org/ingest-eacl/2026.findings-eacl.50/
DOI:
Bibkey:
Cite (ACL):
Kemal Kurniawan, Meladel Mistica, Timothy Baldwin, and Jey Han Lau. 2026. On the Interplay between Human Label Variation and Model Fairness. In Findings of the Association for Computational Linguistics: EACL 2026, pages 967–976, Rabat, Morocco. Association for Computational Linguistics.
Cite (Informal):
On the Interplay between Human Label Variation and Model Fairness (Kurniawan et al., Findings 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-eacl/2026.findings-eacl.50.pdf
Checklist:
 2026.findings-eacl.50.checklist.pdf