@inproceedings{stymne-etal-2018-parser,
title = "Parser Training with Heterogeneous Treebanks",
author = "Stymne, Sara and
de Lhoneux, Miryam and
Smith, Aaron and
Nivre, Joakim",
editor = "Gurevych, Iryna and
Miyao, Yusuke",
booktitle = "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)",
month = jul,
year = "2018",
address = "Melbourne, Australia",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/P18-2098/",
doi = "10.18653/v1/P18-2098",
pages = "619--625",
abstract = "How to make the most of multiple heterogeneous treebanks when training a monolingual dependency parser is an open question. We start by investigating previously suggested, but little evaluated, strategies for exploiting multiple treebanks based on concatenating training sets, with or without fine-tuning. We go on to propose a new method based on treebank embeddings. We perform experiments for several languages and show that in many cases fine-tuning and treebank embeddings lead to substantial improvements over single treebanks or concatenation, with average gains of 2.0{--}3.5 LAS points. We argue that treebank embeddings should be preferred due to their conceptual simplicity, flexibility and extensibility."
}
Markdown (Informal)
[Parser Training with Heterogeneous Treebanks](https://preview.aclanthology.org/fix-sig-urls/P18-2098/) (Stymne et al., ACL 2018)
ACL
- Sara Stymne, Miryam de Lhoneux, Aaron Smith, and Joakim Nivre. 2018. Parser Training with Heterogeneous Treebanks. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 619–625, Melbourne, Australia. Association for Computational Linguistics.