@inproceedings{zhou-etal-2022-bridging,
title = "Bridging Pre-trained Language Models and Hand-crafted Features for Unsupervised {POS} Tagging",
author = "Zhou, Houquan and
Li, Yang and
Li, Zhenghua and
Zhang, Min",
editor = "Muresan, Smaranda and
Nakov, Preslav and
Villavicencio, Aline",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2022",
month = may,
year = "2022",
address = "Dublin, Ireland",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest_wac_2008/2022.findings-acl.259/",
doi = "10.18653/v1/2022.findings-acl.259",
pages = "3276--3290",
abstract = "In recent years, large-scale pre-trained language models (PLMs) have made extraordinary progress in most NLP tasks. But, in the unsupervised POS tagging task, works utilizing PLMs are few and fail to achieve state-of-the-art (SOTA) performance. The recent SOTA performance is yielded by a Guassian HMM variant proposed by He et al. (2018). However, as a generative model, HMM makes very strong independence assumptions, making it very challenging to incorporate contexualized word representations from PLMs. In this work, we for the first time propose a neural conditional random field autoencoder (CRF-AE) model for unsupervised POS tagging. The discriminative encoder of CRF-AE can straightforwardly incorporate ELMo word representations. Moreover, inspired by feature-rich HMM, we reintroduce hand-crafted features into the decoder of CRF-AE. Finally, experiments clearly show that our model outperforms previous state-of-the-art models by a large margin on Penn Treebank and multilingual Universal Dependencies treebank v2.0."
}
Markdown (Informal)
[Bridging Pre-trained Language Models and Hand-crafted Features for Unsupervised POS Tagging](https://preview.aclanthology.org/ingest_wac_2008/2022.findings-acl.259/) (Zhou et al., Findings 2022)
ACL