Which Model Mimics Human Mental Lexicon Better? A Comparative Study of Word Embedding and Generative Models

Huacheng Song, Zhaoxin Feng, Emmanuele Chersoni, Chu-Ren Huang


Abstract
Word associations are commonly applied in psycholinguistics to investigate the nature and structure of the human mental lexicon, and at the same time an important data source for measuring the alignment of language models with human semantic representations.Taking this view, we compare the capacities of different language models to model collective human association norms via five word association tasks (WATs), with predictions about associations driven by either word vector similarities for traditional embedding models or prompting large language models (LLMs).Our results demonstrate that neither approach could produce human-like performances in all five WATs. Hence, none of them can successfully model the human mental lexicon yet. Our detailed analysis shows that static word-type embeddings and prompted LLMs have overall better alignment with human norms compared to word-token embeddings from pretrained models like BERT. Further analysis suggests that the performance discrepancies may be due to different model architectures, especially in terms of approximating human-like associative reasoning through either semantic similarity or relatedness evaluation. Our codes and data are publicly available at: https://github.com/florethsong/word_association.
Anthology ID:
2025.iwcs-1.20
Volume:
Proceedings of the 16th International Conference on Computational Semantics
Month:
September
Year:
2025
Address:
Düsseldorf, Germany
Editors:
Kilian Evang, Laura Kallmeyer, Sylvain Pogodalla
Venues:
IWCS | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
218–240
Language:
URL:
https://preview.aclanthology.org/iwcs-25-ingestion/2025.iwcs-1.20/
DOI:
Bibkey:
Cite (ACL):
Huacheng Song, Zhaoxin Feng, Emmanuele Chersoni, and Chu-Ren Huang. 2025. Which Model Mimics Human Mental Lexicon Better? A Comparative Study of Word Embedding and Generative Models. In Proceedings of the 16th International Conference on Computational Semantics, pages 218–240, Düsseldorf, Germany. Association for Computational Linguistics.
Cite (Informal):
Which Model Mimics Human Mental Lexicon Better? A Comparative Study of Word Embedding and Generative Models (Song et al., IWCS 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/iwcs-25-ingestion/2025.iwcs-1.20.pdf