Enhancing RLHF with Human Gaze Modeling

Karim Galliamov, Ivan Titov, Ilya Pershin


Abstract
Reinforcement Learning from Human Feedback (RLHF) aligns language models with human preferences but faces efficiency challenges. We explore two approaches leveraging human gaze prediction to enhance RLHF: (1) gaze-aware reward models and (2) gaze-based distribution of sparse rewards at token level. Our experiments show gaze-informed RLHF achieves faster convergence while maintaining or slightly improving performance, reducing computational requirements during policy optimization. Human visual attention patterns provide valuable signals for policy training, suggesting a promising direction for improving RLHF efficiency through human-like attention mechanisms.
Anthology ID:
2025.emnlp-main.1559
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
30625–30631
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1559/
DOI:
Bibkey:
Cite (ACL):
Karim Galliamov, Ivan Titov, and Ilya Pershin. 2025. Enhancing RLHF with Human Gaze Modeling. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 30625–30631, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Enhancing RLHF with Human Gaze Modeling (Galliamov et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1559.pdf
Checklist:
 2025.emnlp-main.1559.checklist.pdf