Abstract
EvaHan2024 focuses on sentence punctuation in ancient Chinese. Xunzi large language base model, which is specifically trained for ancient Chinese processing, is advised in the campaign. In general, we adopted the in-context learning (ICL) paradigm for this task and designed a post-processing scheme to ensure the standardability of final results. When constructing ICL prompts, we did feature extraction by LLM QA and selected demonstrations based on non-parametric metrics. We used Xunzi in two stages and neither did further training, so the model was generic and other fundamental abilities remained unaffected. Moreover, newly acquired training data can be directly utilized after identical feature extraction, showcasing the scalability of our system. As for the result, we achieved an F1-score of 67.7% on a complex test dataset consisting of multiple types of documents and 77.98% on Zuozhuan data.- Anthology ID:
- 2024.lt4hala-1.33
- Volume:
- Proceedings of the Third Workshop on Language Technologies for Historical and Ancient Languages (LT4HALA) @ LREC-COLING-2024
- Month:
- May
- Year:
- 2024
- Address:
- Torino, Italia
- Editors:
- Rachele Sprugnoli, Marco Passarotti
- Venues:
- LT4HALA | WS
- SIG:
- Publisher:
- ELRA and ICCL
- Note:
- Pages:
- 261–265
- Language:
- URL:
- https://aclanthology.org/2024.lt4hala-1.33
- DOI:
- Cite (ACL):
- Jie Huang. 2024. Ancient Chinese Punctuation via In-Context Learning. In Proceedings of the Third Workshop on Language Technologies for Historical and Ancient Languages (LT4HALA) @ LREC-COLING-2024, pages 261–265, Torino, Italia. ELRA and ICCL.
- Cite (Informal):
- Ancient Chinese Punctuation via In-Context Learning (Huang, LT4HALA-WS 2024)
- PDF:
- https://preview.aclanthology.org/nschneid-patch-3/2024.lt4hala-1.33.pdf