Stuart Michael McManus
2023
The Ups and Downs of Training RoBERTa-based models on Smaller Datasets for Translation Tasks from Classical Chinese into Modern Standard Mandarin and Modern English
Stuart Michael McManus
|
Roslin Liu
|
Yuji Li
|
Leo Tam
|
Stephanie Qiu
|
Letian Yu
Proceedings of ALT2023: Ancient Language Translation Workshop
The paper presents an investigation into the effectiveness of pre-trained language models, Siku-RoBERTa and RoBERTa, for Classical Chinese to Modern Standard Mandarin and Classical Chinese to English translation tasks. The English translation model resulted in unsatisfactory performance due to the small dataset, while the Modern Standard Mandarin model gave reasonable results.