Iva Bojic
2023
Hierarchical Evaluation Framework: Best Practices for Human Evaluation
Iva Bojic
|
Jessica Chen
|
Si Yuan Chang
|
Qi Chwen Ong
|
Shafiq Joty
|
Josip Car
Proceedings of the 3rd Workshop on Human Evaluation of NLP Systems
Human evaluation plays a crucial role in Natural Language Processing (NLP) as it assesses the quality and relevance of developed systems, thereby facilitating their enhancement. However, the absence of widely accepted human evaluation metrics in NLP hampers fair comparisons among different systems and the establishment of universal assessment standards. Through an extensive analysis of existing literature on human evaluation metrics, we identified several gaps in NLP evaluation methodologies. These gaps served as motivation for developing our own hierarchical evaluation framework. The proposed framework offers notable advantages, particularly in providing a more comprehensive representation of the NLP system’s performance. We applied this framework to evaluate the developed Machine Reading Comprehension system, which was utilized within a human-AI symbiosis model. The results highlighted the associations between the quality of inputs and outputs, underscoring the necessity to evaluate both components rather than solely focusing on outputs. In future work, we will investigate the potential time-saving benefits of our proposed framework for evaluators assessing NLP systems.
A Data-centric Framework for Improving Domain-specific Machine Reading Comprehension Datasets
Iva Bojic
|
Josef Halim
|
Verena Suharman
|
Sreeja Tar
|
Qi Chwen Ong
|
Duy Phung
|
Mathieu Ravaut
|
Shafiq Joty
|
Josip Car
Proceedings of the Fourth Workshop on Insights from Negative Results in NLP
Low-quality data can cause downstream problems in high-stakes applications. Data-centric approach emphasizes on improving dataset quality to enhance model performance. High-quality datasets are needed for general-purpose Large Language Models (LLMs) training, as well as for domain-specific models, which are usually small in size as it is costly to engage a large number of domain experts for their creation. Thus, it is vital to ensure high-quality domain-specific training data. In this paper, we propose a framework for enhancing the data quality of original datasets. (Code and dataset are available at https://github.com/IvaBojic/framework). We applied the proposed framework to four biomedical datasets and showed relative improvement of up to 33%/40% for fine-tuning of retrieval/reader models on the BioASQ dataset when using back translation to enhance the original dataset quality.
Search
Co-authors
- Qi Chwen Ong 2
- Shafiq Joty 2
- Josip Car 2
- Jessica Chen 1
- Si Yuan Chang 1
- show all...