The taraXÜ corpus of human-annotated machine translations
Eleftherios Avramidis, Aljoscha Burchardt, Sabine Hunsicker, Maja Popović, Cindy Tscherwinka, David Vilar, Hans Uszkoreit
Abstract
Human translators are the key to evaluating machine translation (MT) quality and also to addressing the so far unanswered question when and how to use MT in professional translation workflows. This paper describes the corpus developed as a result of a detailed large scale human evaluation consisting of three tightly connected tasks: ranking, error classification and post-editing.- Anthology ID:
- L14-1347
- Volume:
- Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC'14)
- Month:
- May
- Year:
- 2014
- Address:
- Reykjavik, Iceland
- Editors:
- Nicoletta Calzolari, Khalid Choukri, Thierry Declerck, Hrafn Loftsson, Bente Maegaard, Joseph Mariani, Asuncion Moreno, Jan Odijk, Stelios Piperidis
- Venue:
- LREC
- SIG:
- Publisher:
- European Language Resources Association (ELRA)
- Note:
- Pages:
- 2679–2682
- Language:
- URL:
- http://www.lrec-conf.org/proceedings/lrec2014/pdf/401_Paper.pdf
- DOI:
- Cite (ACL):
- Eleftherios Avramidis, Aljoscha Burchardt, Sabine Hunsicker, Maja Popović, Cindy Tscherwinka, David Vilar, and Hans Uszkoreit. 2014. The taraXÜ corpus of human-annotated machine translations. In Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC'14), pages 2679–2682, Reykjavik, Iceland. European Language Resources Association (ELRA).
- Cite (Informal):
- The taraXÜ corpus of human-annotated machine translations (Avramidis et al., LREC 2014)
- PDF:
- http://www.lrec-conf.org/proceedings/lrec2014/pdf/401_Paper.pdf