Abstract
Character-level models have been used extensively in recent years in NLP tasks as both supplements and replacements for closed-vocabulary token-level word representations. In one popular architecture, character-level LSTMs are used to feed token representations into a sequence tagger predicting token-level annotations such as part-of-speech (POS) tags. In this work, we examine the behavior of POS taggers across languages from the perspective of individual hidden units within the character LSTM. We aggregate the behavior of these units into language-level metrics which quantify the challenges that taggers face on languages with different morphological properties, and identify links between synthesis and affixation preference and emergent behavior of the hidden tagger layer. In a comparative experiment, we show how modifying the balance between forward and backward hidden units affects model arrangement and performance in these types of languages.- Anthology ID:
- W19-4811
- Volume:
- Proceedings of the 2019 ACL Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP
- Month:
- August
- Year:
- 2019
- Address:
- Florence, Italy
- Editors:
- Tal Linzen, Grzegorz Chrupała, Yonatan Belinkov, Dieuwke Hupkes
- Venue:
- BlackboxNLP
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 95–102
- Language:
- URL:
- https://aclanthology.org/W19-4811
- DOI:
- 10.18653/v1/W19-4811
- Cite (ACL):
- Yuval Pinter, Marc Marone, and Jacob Eisenstein. 2019. Character Eyes: Seeing Language through Character-Level Taggers. In Proceedings of the 2019 ACL Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, pages 95–102, Florence, Italy. Association for Computational Linguistics.
- Cite (Informal):
- Character Eyes: Seeing Language through Character-Level Taggers (Pinter et al., BlackboxNLP 2019)
- PDF:
- https://preview.aclanthology.org/nschneid-patch-2/W19-4811.pdf
- Code
- ruyimarone/character-eyes