Does Higher Order LSTM Have Better Accuracy for Segmenting and Labeling Sequence Data?

Yi Zhang, Xu Sun, Shuming Ma, Yang Yang, Xuancheng Ren


Abstract
Existing neural models usually predict the tag of the current token independent of the neighboring tags. The popular LSTM-CRF model considers the tag dependencies between every two consecutive tags. However, it is hard for existing neural models to take longer distance dependencies between tags into consideration. The scalability is mainly limited by the complex model structures and the cost of dynamic programming during training. In our work, we first design a new model called “high order LSTM” to predict multiple tags for the current token which contains not only the current tag but also the previous several tags. We call the number of tags in one prediction as “order”. Then we propose a new method called Multi-Order BiLSTM (MO-BiLSTM) which combines low order and high order LSTMs together. MO-BiLSTM keeps the scalability to high order models with a pruning technique. We evaluate MO-BiLSTM on all-phrase chunking and NER datasets. Experiment results show that MO-BiLSTM achieves the state-of-the-art result in chunking and highly competitive results in two NER datasets.
Anthology ID:
C18-1061
Volume:
Proceedings of the 27th International Conference on Computational Linguistics
Month:
August
Year:
2018
Address:
Santa Fe, New Mexico, USA
Editors:
Emily M. Bender, Leon Derczynski, Pierre Isabelle
Venue:
COLING
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
723–733
Language:
URL:
https://aclanthology.org/C18-1061
DOI:
Bibkey:
Cite (ACL):
Yi Zhang, Xu Sun, Shuming Ma, Yang Yang, and Xuancheng Ren. 2018. Does Higher Order LSTM Have Better Accuracy for Segmenting and Labeling Sequence Data?. In Proceedings of the 27th International Conference on Computational Linguistics, pages 723–733, Santa Fe, New Mexico, USA. Association for Computational Linguistics.
Cite (Informal):
Does Higher Order LSTM Have Better Accuracy for Segmenting and Labeling Sequence Data? (Zhang et al., COLING 2018)
Copy Citation:
PDF:
https://preview.aclanthology.org/naacl-24-ws-corrections/C18-1061.pdf
Code
 lancopku/Multi-Order-LSTM