Increasing the Generalizability of Similarity-Based Essay Scoring Through Cross-Prompt Training

Marie Bexte, Yuning Ding, Andrea Horbach


Abstract
In this paper, we address generic essay scoring, i.e., the use of training data from one writing task to score data from a different task. We approach this by generalizing a similarity-based essay scoring method (Xie et al., 2022) to learning from texts that are written in response to a mixture of different prompts. In our experiments, we compare within-prompt and cross-prompt performance on two large datasets (ASAP and PERSUADE). We combine different amounts of prompts in the training data and show that our generalized method substantially improves cross-prompt performance, especially when an increasing number of prompts is used to form the training data. In the most extreme case, this leads to more than double the performance, increasing QWK from .26 to .55.
Anthology ID:
2025.bea-1.17
Volume:
Proceedings of the 20th Workshop on Innovative Use of NLP for Building Educational Applications (BEA 2025)
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Ekaterina Kochmar, Bashar Alhafni, Marie Bexte, Jill Burstein, Andrea Horbach, Ronja Laarmann-Quante, Anaïs Tack, Victoria Yaneva, Zheng Yuan
Venues:
BEA | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
225–236
Language:
URL:
https://preview.aclanthology.org/acl25-workshop-ingestion/2025.bea-1.17/
DOI:
Bibkey:
Cite (ACL):
Marie Bexte, Yuning Ding, and Andrea Horbach. 2025. Increasing the Generalizability of Similarity-Based Essay Scoring Through Cross-Prompt Training. In Proceedings of the 20th Workshop on Innovative Use of NLP for Building Educational Applications (BEA 2025), pages 225–236, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Increasing the Generalizability of Similarity-Based Essay Scoring Through Cross-Prompt Training (Bexte et al., BEA 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/acl25-workshop-ingestion/2025.bea-1.17.pdf