CA*: Addressing Evaluation Pitfalls in Computation-Aware Latency for Simultaneous Speech Translation

Xi Xu, Wenda Xu, Siqi Ouyang, Lei Li


Abstract
Simultaneous speech translation (SimulST) systems must balance translation quality with response time, making latency measurement crucial for evaluating their real-world performance. However, there has been a longstanding belief that current metrics yield unrealistically high latency measurements in unsegmented streaming settings. In this paper, we investigate this phenomenon, revealing its root cause in a fundamental misconception underlying existing latency evaluation approaches. We demonstrate that this issue affects not only streaming but also segment-level latency evaluation across different metrics. Furthermore, we propose a modification to correctly measure computation-aware latency for SimulST systems, addressing the limitations present in existing metrics.
Anthology ID:
2025.findings-naacl.393
Volume:
Findings of the Association for Computational Linguistics: NAACL 2025
Month:
April
Year:
2025
Address:
Albuquerque, New Mexico
Editors:
Luis Chiruzzo, Alan Ritter, Lu Wang
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
7062–7067
Language:
URL:
https://preview.aclanthology.org/landing_page/2025.findings-naacl.393/
DOI:
Bibkey:
Cite (ACL):
Xi Xu, Wenda Xu, Siqi Ouyang, and Lei Li. 2025. CA*: Addressing Evaluation Pitfalls in Computation-Aware Latency for Simultaneous Speech Translation. In Findings of the Association for Computational Linguistics: NAACL 2025, pages 7062–7067, Albuquerque, New Mexico. Association for Computational Linguistics.
Cite (Informal):
CA*: Addressing Evaluation Pitfalls in Computation-Aware Latency for Simultaneous Speech Translation (Xu et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/landing_page/2025.findings-naacl.393.pdf