Yuliia Teslia


2023

pdf
MUNI-NLP Submission for Czech-Ukrainian Translation Task at WMT23
Pavel Rychly | Yuliia Teslia
Proceedings of the Eighth Conference on Machine Translation

The system is trained on officialy provided data only. We have heavily filtered all the data to remove machine translated text, Russian text and other noise. We use the DeepNorm modification of the transformer architecture in the TorchScale library with 18 encoder layers and 6 decoder layers. The initial systems for backtranslation uses HFT tokenizer, the final system uses custom tokenizer derived from HFT.
Search
Co-authors
Venues