Disentangling Emotion Understanding and Generation in Large Language Models

Sadegh Jafari, Els Lefever, Veronique Hoste


Abstract
Large language models (LLMs) have demonstrated strong performance on emotion understanding tasks, yet their ability to faithfully generate emotionally aligned text remains less well understood.We propose a semantic evaluation framework that jointly assesses emotion understanding, emotion generation, and internal consistency, using a VAE-based emotion cost matrix that captures graded semantic similarity between emotion categories.Our framework introduces four complementary metrics that disentangle baseline understanding, human-perceived emotion in generated text, generation quality, and model consistency.Experimental results show that while understanding and consistency scores are highly correlated, emotion generation exhibits substantially weaker correlations with these metrics.These findings motivate the development of specialized evaluation protocols that independently measure emotional understanding and generation, enabling more reliable assessments of LLM emotional intelligence.
Anthology ID:
2026.wassa-1.14
Volume:
The Proceedings for the 15th Workshop on Computational Approaches to Subjectivity, Sentiment Social Media Analysis (WASSA 2026)
Month:
March
Year:
2026
Address:
Rabat, Morocco
Editors:
Jeremy Barnes, Valentin Barriere, Orphée De Clercq, Roman Klinger, Célia Nouri, Debora Nozza, Pranaydeep Singh
Venues:
WASSA | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
161–171
Language:
URL:
https://preview.aclanthology.org/ingest-eacl/2026.wassa-1.14/
DOI:
Bibkey:
Cite (ACL):
Sadegh Jafari, Els Lefever, and Veronique Hoste. 2026. Disentangling Emotion Understanding and Generation in Large Language Models. In The Proceedings for the 15th Workshop on Computational Approaches to Subjectivity, Sentiment Social Media Analysis (WASSA 2026), pages 161–171, Rabat, Morocco. Association for Computational Linguistics.
Cite (Informal):
Disentangling Emotion Understanding and Generation in Large Language Models (Jafari et al., WASSA 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-eacl/2026.wassa-1.14.pdf