@inproceedings{he-etal-2022-weighted,
title = "Weighted self Distillation for {C}hinese word segmentation",
author = "He, Rian and
Cai, Shubin and
Ming, Zhong and
Zhang, Jialei",
editor = "Muresan, Smaranda and
Nakov, Preslav and
Villavicencio, Aline",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2022",
month = may,
year = "2022",
address = "Dublin, Ireland",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2022.findings-acl.139/",
doi = "10.18653/v1/2022.findings-acl.139",
pages = "1757--1770",
abstract = "Recent researches show that multi-criteria resources and n-gram features are beneficial to Chinese Word Segmentation (CWS). However, these methods rely heavily on such additional information mentioned above and focus less on the model itself. We thus propose a novel neural framework, named Weighted self Distillation for Chinese word segmentation (WeiDC). The framework, which only requires unigram features, adopts self-distillation technology with four hand-crafted weight modules and two teacher models configurations. Experiment results show that WeiDC can make use of character features to learn contextual knowledge and successfully achieve state-of-the-art or competitive performance in terms of strictly closed test settings on SIGHAN Bakeoff benchmark datasets. Moreover, further experiments and analyses also demonstrate the robustness of WeiDC. Source codes of this paper are available on Github."
}
Markdown (Informal)
[Weighted self Distillation for Chinese word segmentation](https://preview.aclanthology.org/fix-sig-urls/2022.findings-acl.139/) (He et al., Findings 2022)
ACL