Yugyeong Ji


2025

pdf bib
Transparent Reference-free Automated Evaluation of Open-Ended User Survey Responses
Subin An | Yugyeong Ji | Junyoung Kim | Heejin Kook | Yang Lu | Josh Seltzer
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track

Open-ended survey responses provide valuable insights in marketing research, but low-quality responses not only burden researchers with manual filtering but also risk leading to misleading conclusions, underscoring the need for effective evaluation. Existing automatic evaluation methods target LLM-generated text and inadequately assess human-written responses with their distinct characteristics. To address such characteristics, we propose a two-stage evaluation framework specifically designed for human survey responses. First, gibberish filtering removes nonsensical responses. Then, three dimensions—effort, relevance, and complete- ness—are evaluated using LLM capabilities, grounded in empirical analysis of real-world survey data. Validation on English and Korean datasets shows that our framework not only outperforms existing metrics but also demonstrates high practical applicability for real-world applications such as response quality prediction and response rejection, showing strong correlations with expert assessment.