@inproceedings{maveli-cohen-2022-co,
title = "{C}o-training an {U}nsupervised {C}onstituency {P}arser with {W}eak {S}upervision",
author = "Maveli, Nickil and
Cohen, Shay",
editor = "Muresan, Smaranda and
Nakov, Preslav and
Villavicencio, Aline",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2022",
month = may,
year = "2022",
address = "Dublin, Ireland",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2022.findings-acl.101/",
doi = "10.18653/v1/2022.findings-acl.101",
pages = "1274--1291",
abstract = "We introduce a method for unsupervised parsing that relies on bootstrapping classifiers to identify if a node dominates a specific span in a sentence. There are two types of classifiers, an inside classifier that acts on a span, and an outside classifier that acts on everything outside of a given span. Through self-training and co-training with the two classifiers, we show that the interplay between them helps improve the accuracy of both, and as a result, effectively parse. A seed bootstrapping technique prepares the data to train these classifiers. Our analyses further validate that such an approach in conjunction with weak supervision using prior branching knowledge of a known language (left/right-branching) and minimal heuristics injects strong inductive bias into the parser, achieving 63.1 F$_1$ on the English (PTB) test set. In addition, we show the effectiveness of our architecture by evaluating on treebanks for Chinese (CTB) and Japanese (KTB) and achieve new state-of-the-art results."
}
Markdown (Informal)
[Co-training an Unsupervised Constituency Parser with Weak Supervision](https://preview.aclanthology.org/fix-sig-urls/2022.findings-acl.101/) (Maveli & Cohen, Findings 2022)
ACL