The Emergence of Chunking Structures with Hierarchical RNN

Zijun Wu, Anup Anand Deshmukh, Yongkang Wu, Jimmy Lin, Lili Mou


Abstract
In Natural Language Processing (NLP), predicting linguistic structures, such as parsing and chunking, has mostly relied on manual annotations of syntactic structures. This article introduces an unsupervised approach to chunking, a syntactic task that involves grouping words in a non-hierarchical manner. We present a Hierarchical Recurrent Neural Network (HRNN) designed to model word-to-chunk and chunk-to-sentence compositions. Our approach involves a two-stage training process: pretraining with an unsupervised parser and finetuning on downstream NLP tasks. Experiments on multiple datasets reveal a notable improvement of unsupervised chunking performance in both pretraining and finetuning stages. Interestingly, we observe that the emergence of the chunking structure is transient during the neural model’s downstream-task training. This study contributes to the advancement of unsupervised syntactic structure discovery and opens avenues for further research in linguistic theory.1
Anthology ID:
2025.cl-3.4
Volume:
Computational Linguistics, Volume 51, Issue 3 - September 2025
Month:
September
Year:
2025
Address:
Cambridge, MA
Venue:
CL
SIG:
Publisher:
MIT Press
Note:
Pages:
815–841
Language:
URL:
https://preview.aclanthology.org/ingest-eacl/2025.cl-3.4/
DOI:
10.1162/coli_a_00545
Bibkey:
Cite (ACL):
Zijun Wu, Anup Anand Deshmukh, Yongkang Wu, Jimmy Lin, and Lili Mou. 2025. The Emergence of Chunking Structures with Hierarchical RNN. Computational Linguistics, 51(3):815–841.
Cite (Informal):
The Emergence of Chunking Structures with Hierarchical RNN (Wu et al., CL 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-eacl/2025.cl-3.4.pdf