@inproceedings{sekizawa-yanaka-2023-analyzing,
title = "Analyzing Syntactic Generalization Capacity of Pre-trained Language Models on {J}apanese Honorific Conversion",
author = "Sekizawa, Ryo and
Yanaka, Hitomi",
editor = "Palmer, Alexis and
Camacho-collados, Jose",
booktitle = "Proceedings of the 12th Joint Conference on Lexical and Computational Semantics (*SEM 2023)",
month = jul,
year = "2023",
address = "Toronto, Canada",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2023.starsem-1.5/",
doi = "10.18653/v1/2023.starsem-1.5",
pages = "40--47",
abstract = "Using Japanese honorifics is challenging because it requires not only knowledge of the grammatical rules but also contextual information, such as social relationships. It remains unclear whether pre-trained large language models (LLMs) can flexibly handle Japanese honorifics like humans. To analyze this, we introduce an honorific conversion task that considers social relationships among people mentioned in a conversation. We construct a Japanese honorifics dataset from problem templates of various sentence structures to investigate the syntactic generalization capacity of GPT-3, one of the leading LLMs, on this task under two settings: fine-tuning and prompt learning. Our results showed that the fine-tuned GPT-3 performed better in a context-aware honorific conversion task than the prompt-based one. The fine-tuned model demonstrated overall syntactic generalizability towards compound honorific sentences, except when tested with the data involving direct speech."
}
Markdown (Informal)
[Analyzing Syntactic Generalization Capacity of Pre-trained Language Models on Japanese Honorific Conversion](https://preview.aclanthology.org/fix-sig-urls/2023.starsem-1.5/) (Sekizawa & Yanaka, *SEM 2023)
ACL