Optimizing Hidden Markov Language Models: An Empirical Study of Reparameterization and Initialization Techniques

Ivan Lee, Taylor Berg-Kirkpatrick


Abstract
Hidden Markov models (HMMs) are valuable for their ability to provide exact and tractable inference. However, learning an HMM in an unsupervised manner involves a non-convex optimization problem that is plagued by poor local optima. Recent work on scaling-up HMMs to perform competitively as language models has indicated that this challenge only increases with larger hidden state sizes. Several techniques to address this problem have been proposed, but have not be evaluated comprehensively. This study provides a comprehensive empirical analysis of two recent strategies that use neural networks to enhance HMM optimization: neural reparameterization and neural initialization. We find that (1) these techniques work effectively for scaled HMM language modeling, (2) linear reparameterizations can be as effective as non-linear ones, and (3) the strategies are complementary.
Anthology ID:
2025.findings-naacl.429
Volume:
Findings of the Association for Computational Linguistics: NAACL 2025
Month:
April
Year:
2025
Address:
Albuquerque, New Mexico
Editors:
Luis Chiruzzo, Alan Ritter, Lu Wang
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
7712–7723
Language:
URL:
https://preview.aclanthology.org/landing_page/2025.findings-naacl.429/
DOI:
Bibkey:
Cite (ACL):
Ivan Lee and Taylor Berg-Kirkpatrick. 2025. Optimizing Hidden Markov Language Models: An Empirical Study of Reparameterization and Initialization Techniques. In Findings of the Association for Computational Linguistics: NAACL 2025, pages 7712–7723, Albuquerque, New Mexico. Association for Computational Linguistics.
Cite (Informal):
Optimizing Hidden Markov Language Models: An Empirical Study of Reparameterization and Initialization Techniques (Lee & Berg-Kirkpatrick, Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/landing_page/2025.findings-naacl.429.pdf