Abstract
Effectively assessing Natural Language Processing output tasks is a challenge for research in the area. In the case of Machine Translation (MT), automatic metrics are usually preferred over human evaluation, given time and budget constraints.However, traditional automatic metrics (such as BLEU) are not reliable for absolute quality assessment of documents, often producing similar scores for documents translated by the same MT system.For scenarios where absolute labels are necessary for building models, such as document-level Quality Estimation, these metrics can not be fully trusted. In this paper, we introduce a corpus of reading comprehension tests based on machine translated documents, where we evaluate documents based on answers to questions by fluent speakers of the target language. We describe the process of creating such a resource, the experiment design and agreement between the test takers. Finally, we discuss ways to convert the reading comprehension test into document-level quality scores.- Anthology ID:
- L16-1579
- Volume:
- Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC'16)
- Month:
- May
- Year:
- 2016
- Address:
- Portorož, Slovenia
- Venue:
- LREC
- SIG:
- Publisher:
- European Language Resources Association (ELRA)
- Note:
- Pages:
- 3652–3658
- Language:
- URL:
- https://aclanthology.org/L16-1579
- DOI:
- Cite (ACL):
- Carolina Scarton and Lucia Specia. 2016. A Reading Comprehension Corpus for Machine Translation Evaluation. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC'16), pages 3652–3658, Portorož, Slovenia. European Language Resources Association (ELRA).
- Cite (Informal):
- A Reading Comprehension Corpus for Machine Translation Evaluation (Scarton & Specia, LREC 2016)
- PDF:
- https://preview.aclanthology.org/paclic-22-ingestion/L16-1579.pdf
- Code
- carolscarton/CREG-MT-eval