Abstract
Choosing the most suitable classifier in a linguistic context is a well-known problem in the production of Mandarin and many other languages. The present paper proposes a solution based on BERT, compares this solution to previous neural and rule-based models, and argues that the BERT model performs particularly well on those difficult cases where the classifier adds information to the text.- Anthology ID:
- 2021.inlg-1.17
- Volume:
- Proceedings of the 14th International Conference on Natural Language Generation
- Month:
- August
- Year:
- 2021
- Address:
- Aberdeen, Scotland, UK
- Editors:
- Anya Belz, Angela Fan, Ehud Reiter, Yaji Sripada
- Venue:
- INLG
- SIG:
- SIGGEN
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 172–176
- Language:
- URL:
- https://aclanthology.org/2021.inlg-1.17
- DOI:
- 10.18653/v1/2021.inlg-1.17
- Cite (ACL):
- Jani Järnfors, Guanyi Chen, Kees van Deemter, and Rint Sybesma. 2021. Using BERT for choosing classifiers in Mandarin. In Proceedings of the 14th International Conference on Natural Language Generation, pages 172–176, Aberdeen, Scotland, UK. Association for Computational Linguistics.
- Cite (Informal):
- Using BERT for choosing classifiers in Mandarin (Järnfors et al., INLG 2021)
- PDF:
- https://preview.aclanthology.org/ml4al-ingestion/2021.inlg-1.17.pdf
- Data
- Chinese Classifier