2022
abs
Feeding NMT a Healthy Diet – The Impact of Quality, Quantity, or the Right Type of Nutrients
Abdallah Nasir
|
Sara Alisis
|
Ruba W Jaikat
|
Rebecca Jonsson
|
Sara Qardan
|
Eyas Shawahneh
|
Nour Al-Khdour
Proceedings of the 15th Biennial Conference of the Association for Machine Translation in the Americas (Volume 2: Users and Providers Track and Government Track)
In the era of gigantic language models, and in our case, Neural Machine Translation (NMT) models, where merely size seems to matter, we’ve been asking ourselves, is it healthy to just feed our NMT model with more and more data? In this presentation, we want to show our findings on the impact of NMT performance of different data “nutrients” we were feeding our models. We have explored the impact of quantity, quality and the type of data we feed to our English-Arabic NMT models. The presentation will show the impact of adding millions of parallel sentences into our training data as opposed to a much smaller data set with much higher quality, and the results from additional experiments with different data nutrients. We will highlight our learnings, challenges and share insights from our Linguistics Quality Assurance team, on what are the advantages and disadvantages of each type of data source and define the criteria of high-quality data with respect to a healthy NMT diet.
abs
Unlocking the value of bilingual translated documents with Deep Learning Segmentation and Alignment for Arabic
Nour Al-Khdour
|
Rebecca Jonsson
|
Ruba W Jaikat
|
Abdallah Nasir
|
Sara Alisis
|
Sara Qardan
|
Eyas Shawahneh
Proceedings of the 15th Biennial Conference of the Association for Machine Translation in the Americas (Volume 2: Users and Providers Track and Government Track)
To unlock the value of high-quality bilingual translated documents we need parallel data. With sentence-aligned translation pairs, we can fuel our neural machine translation, customize MT or create translation memories for our clients. To automate this process, automatic segmentation and alignment are required. Despite Arabic being the fifth biggest language in the world, language technology for Arabic is many times way behind other languages. We will show how we struggled to find a proper sentence segmentation for Arabic and instead explored different frameworks, from statistical to deep learning, to end up fine-tuning our own Arabic DL segmentation model. We will highlight our learnings and challenges with segmenting and aligning Arabic and English bilingual data. Finally, we will show the impact on our proprietary NMT engine as we started to unlock the value and could leverage data that had been translated offline, outside CAT tools, as well as comparable corpora, to feed our NMT.
2021
abs
A Data-Centric Approach to Real-World Custom NMT for Arabic
Rebecca Jonsson
|
Ruba Jaikat
|
Abdallah Nasir
|
Nour Al-Khdour
|
Sara Alisis
Proceedings of Machine Translation Summit XVIII: Users and Providers Track
In this presentation, we will present our approach to taking Custom NMT to the next level by building tailor-made NMT to fit the needs of businesses seeking to scale in the Arabic-speaking world. In close collaboration with customers in the MENA region and with a deep understanding of their data, we work on building a variety of NMT models that accommodate to the unique challenges of the Arabic language. This session will provide insights into the challenges of acquiring, analyzing, and processing customer data in various sectors, as well as insights into how to best make use of this data to build high-quality Custom NMT models in English-Arabic. Feedback from usage of these models in production will be provided. Furthermore, we will show how to use our translation management system to make the most of the custom NMT, by leveraging the models, fine-tuning and continuing to improve them over time.
2019
pdf
bib
AnonyMate: A Toolkit for Anonymizing Unstructured Chat Data
Allison Adams
|
Eric Aili
|
Daniel Aioanei
|
Rebecca Jonsson
|
Lina Mickelsson
|
Dagmar Mikmekova
|
Fred Roberts
|
Javier Fernandez Valencia
|
Roger Wechsler
Proceedings of the Workshop on NLP and Pseudonymisation