Abstract
We present a semi-supervised way of training a character-based encoder-decoder recurrent neural network for morphological reinflection—the task of generating one inflected wordform from another. This is achieved by using unlabeled tokens or random strings as training data for an autoencoding task, adapting a network for morphological reinflection, and performing multi-task training. We thus use limited labeled data more effectively, obtaining up to 9.92% improvement over state-of-the-art baselines for 8 different languages.- Anthology ID:
- W17-4111
- Volume:
- Proceedings of the First Workshop on Subword and Character Level Models in NLP
- Month:
- September
- Year:
- 2017
- Address:
- Copenhagen, Denmark
- Editors:
- Manaal Faruqui, Hinrich Schuetze, Isabel Trancoso, Yadollah Yaghoobzadeh
- Venue:
- SCLeM
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 76–81
- Language:
- URL:
- https://aclanthology.org/W17-4111
- DOI:
- 10.18653/v1/W17-4111
- Cite (ACL):
- Katharina Kann and Hinrich Schütze. 2017. Unlabeled Data for Morphological Generation With Character-Based Sequence-to-Sequence Models. In Proceedings of the First Workshop on Subword and Character Level Models in NLP, pages 76–81, Copenhagen, Denmark. Association for Computational Linguistics.
- Cite (Informal):
- Unlabeled Data for Morphological Generation With Character-Based Sequence-to-Sequence Models (Kann & Schütze, SCLeM 2017)
- PDF:
- https://preview.aclanthology.org/emnlp22-frontmatter/W17-4111.pdf