Sevi Sariisik
2022
GoURMET – Machine Translation for Low-Resourced Languages
Peggy van der Kreeft
|
Alexandra Birch
|
Sevi Sariisik
|
Felipe Sánchez-Martínez
|
Wilker Aziz
Proceedings of the 23rd Annual Conference of the European Association for Machine Translation
The GoURMET project, funded by the European Commission’s H2020 program (under grant agreement 825299), develops models for machine translation, in particular for low-resourced languages. Data, models and software releases as well as the GoURMET Translate Tool are made available as open source.
2021
Surprise Language Challenge: Developing a Neural Machine Translation System between Pashto and English in Two Months
Alexandra Birch
|
Barry Haddow
|
Antonio Valerio Miceli Barone
|
Jindrich Helcl
|
Jonas Waldendorf
|
Felipe Sánchez Martínez
|
Mikel Forcada
|
Víctor Sánchez Cartagena
|
Juan Antonio Pérez-Ortiz
|
Miquel Esplà-Gomis
|
Wilker Aziz
|
Lina Murady
|
Sevi Sariisik
|
Peggy van der Kreeft
|
Kay Macquarrie
Proceedings of Machine Translation Summit XVIII: Research Track
In the media industry and the focus of global reporting can shift overnight. There is a compelling need to be able to develop new machine translation systems in a short period of time and in order to more efficiently cover quickly developing stories. As part of the EU project GoURMET and which focusses on low-resource machine translation and our media partners selected a surprise language for which a machine translation system had to be built and evaluated in two months(February and March 2021). The language selected was Pashto and an Indo-Iranian language spoken in Afghanistan and Pakistan and India. In this period we completed the full pipeline of development of a neural machine translation system: data crawling and cleaning and aligning and creating test sets and developing and testing models and and delivering them to the user partners. In this paperwe describe rapid data creation and experiments with transfer learning and pretraining for this low-resource language pair. We find that starting from an existing large model pre-trained on 50languages leads to far better BLEU scores than pretraining on one high-resource language pair with a smaller model. We also present human evaluation of our systems and which indicates that the resulting systems perform better than a freely available commercial system when translating from English into Pashto direction and and similarly when translating from Pashto into English.