Which Model Mimics Human Mental Lexicon Better? A Comparative Study of Word Embedding and Generative Models
Huacheng Song, Zhaoxin Feng, Emmanuele Chersoni, Chu-Ren Huang
Abstract
Word associations are commonly applied in psycholinguistics to investigate the nature and structure of the human mental lexicon, and at the same time an important data source for measuring the alignment of language models with human semantic representations.Taking this view, we compare the capacities of different language models to model collective human association norms via five word association tasks (WATs), with predictions about associations driven by either word vector similarities for traditional embedding models or prompting large language models (LLMs).Our results demonstrate that neither approach could produce human-like performances in all five WATs. Hence, none of them can successfully model the human mental lexicon yet. Our detailed analysis shows that static word-type embeddings and prompted LLMs have overall better alignment with human norms compared to word-token embeddings from pretrained models like BERT. Further analysis suggests that the performance discrepancies may be due to different model architectures, especially in terms of approximating human-like associative reasoning through either semantic similarity or relatedness evaluation. Our codes and data are publicly available at: https://github.com/florethsong/word_association.- Anthology ID:
- 2025.iwcs-1.20
- Volume:
- Proceedings of the 16th International Conference on Computational Semantics
- Month:
- September
- Year:
- 2025
- Address:
- Düsseldorf, Germany
- Editors:
- Kilian Evang, Laura Kallmeyer, Sylvain Pogodalla
- Venues:
- IWCS | WS
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 218–240
- Language:
- URL:
- https://preview.aclanthology.org/iwcs-25-ingestion/2025.iwcs-1.20/
- DOI:
- Cite (ACL):
- Huacheng Song, Zhaoxin Feng, Emmanuele Chersoni, and Chu-Ren Huang. 2025. Which Model Mimics Human Mental Lexicon Better? A Comparative Study of Word Embedding and Generative Models. In Proceedings of the 16th International Conference on Computational Semantics, pages 218–240, Düsseldorf, Germany. Association for Computational Linguistics.
- Cite (Informal):
- Which Model Mimics Human Mental Lexicon Better? A Comparative Study of Word Embedding and Generative Models (Song et al., IWCS 2025)
- PDF:
- https://preview.aclanthology.org/iwcs-25-ingestion/2025.iwcs-1.20.pdf