Ngo Van Dong
2026
Beyond Coherence: Improving Temporal Consistency and Interpretability in Dynamic Topic Models
Thanh Vinh Nguyen | Ngo Van Dong | Minh Chu Xuan | Tung Nguyen | Linh Ngo Van | Dinh Viet Sang | Trung Le
Findings of the Association for Computational Linguistics: EACL 2026
Thanh Vinh Nguyen | Ngo Van Dong | Minh Chu Xuan | Tung Nguyen | Linh Ngo Van | Dinh Viet Sang | Trung Le
Findings of the Association for Computational Linguistics: EACL 2026
Dynamic topic models aim to reveal how themes emerge, evolve, and dissolve in time-stamped corpora, but existing approaches still face three major challenges: (i) encoders capture bag-of-words statistics but fail to align with the rich semantic priors of large pre-trained language models, (ii) temporal linkages are often modeled as rigid one-to-one chains, limiting the ability to track non-linear evolution such as topic splits or merges, and (iii) interpretability remains shallow, relying on noisy top-word lists that obscure thematic clarity. We propose L-DNTM (LLM-Augmented for Dynamic Neural Topic Model), a variational framework designed to capture more faithful temporal trajectories. Our model integrates three key components: multi-objective distillation to inject PLM-derived semantic knowledge into the encoder, entropy-regularized optimal transport to align entire topic constellations across time for smooth yet flexible evolution, and LLM-guided refinement to sharpen topic–word distributions for improved interpretability. Extensive experiments on diverse corpora show that L-DNTM yields more coherent, temporally consistent, and interpretable topic dynamics, and further enhances downstream classification and clustering tasks.