Serial Recall Effects in Neural Language Modeling
Hassan Hajipoor, Hadi Amiri, Maseud Rahgozar, Farhad Oroumchian
Abstract
Serial recall experiments study the ability of humans to recall words in the order in which they occurred. The following serial recall effects are generally investigated in studies with humans: word length and frequency, primacy and recency, semantic confusion, repetition, and transposition effects. In this research, we investigate neural language models in the context of these serial recall effects. Our work provides a framework to better understand and analyze neural language models and opens a new window to develop accurate language models.- Anthology ID:
- N19-1073
- Volume:
- Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)
- Month:
- June
- Year:
- 2019
- Address:
- Minneapolis, Minnesota
- Editors:
- Jill Burstein, Christy Doran, Thamar Solorio
- Venue:
- NAACL
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 688–694
- Language:
- URL:
- https://aclanthology.org/N19-1073
- DOI:
- 10.18653/v1/N19-1073
- Cite (ACL):
- Hassan Hajipoor, Hadi Amiri, Maseud Rahgozar, and Farhad Oroumchian. 2019. Serial Recall Effects in Neural Language Modeling. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 688–694, Minneapolis, Minnesota. Association for Computational Linguistics.
- Cite (Informal):
- Serial Recall Effects in Neural Language Modeling (Hajipoor et al., NAACL 2019)
- PDF:
- https://preview.aclanthology.org/autopr/N19-1073.pdf