@inproceedings{parulekar-jyothi-2025-laser,
title = "{LASER}: An {LLM}-based {ASR} Scoring and Evaluation Rubric",
author = "Parulekar, Amruta and
Jyothi, Preethi",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1257/",
pages = "24773--24782",
ISBN = "979-8-89176-332-6",
abstract = "Standard ASR evaluation metrics like Word Error Rate (WER) tend to unfairly penalize morphological and syntactic nuances that do not significantly alter sentence semantics. We introduce an LLM-based scoring rubric LASER that leverages state-of-the-art LLMs' in-context learning abilities to learn from prompts with detailed examples. Hindi LASER scores using Gemini 2.5 Pro achieved a very high correlation score of 94{\%} with human annotations. Hindi examples in the prompt were also effective in analyzing errors in other Indian languages such as Marathi, Kannada and Malayalam. We also demonstrate how a smaller LLM like Llama 3 can be finetuned on word-pair examples derived from reference and ASR predictions to predict what kind of penalty should be applied with close to 89{\%} accuracy."
}Markdown (Informal)
[LASER: An LLM-based ASR Scoring and Evaluation Rubric](https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1257/) (Parulekar & Jyothi, EMNLP 2025)
ACL
- Amruta Parulekar and Preethi Jyothi. 2025. LASER: An LLM-based ASR Scoring and Evaluation Rubric. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 24773–24782, Suzhou, China. Association for Computational Linguistics.