Abstract
Work on comparing a set of linguistic test scores for MT output to a set of the same tests’ scores for naturally-occurring target language text (Jones and Rusk 2000) broke new ground in automating MT Evaluation. However, the tests used were selected on an ad hoc basis. In this paper, we report on work to extend our understanding, through refinement and validation, of suitable linguistic tests in the context of our novel approach to MTE. This approach was introduced in Miller and Vanni (2001a) and employs standard, rather than randomly-chosen, tests of MT output quality selected from the ISLE framework as well as a scoring system for predicting the type of information processing task performable with the output. Since the intent is to automate the scoring system, this work can also be viewed as the preliminary steps of algorithm design.- Anthology ID:
- 2001.mtsummit-eval.9
- Volume:
- Workshop on MT Evaluation
- Month:
- September 18-22
- Year:
- 2001
- Address:
- Santiago de Compostela, Spain
- Editors:
- Eduard Hovy, Margaret King, Sandra Manzi, Florence Reeder
- Venue:
- MTSummit
- SIG:
- Publisher:
- Note:
- Pages:
- Language:
- URL:
- https://aclanthology.org/2001.mtsummit-eval.9
- DOI:
- Cite (ACL):
- Michelle Vanni and Keith J. Miller. 2001. Scaling the ISLE framework: validating tests of machine translation quality for multi-dimensional measurement. In Workshop on MT Evaluation, Santiago de Compostela, Spain.
- Cite (Informal):
- Scaling the ISLE framework: validating tests of machine translation quality for multi-dimensional measurement (Vanni & Miller, MTSummit 2001)
- PDF:
- https://preview.aclanthology.org/nschneid-patch-4/2001.mtsummit-eval.9.pdf