From Input Perception to Predictive Insight: Modeling Model Blind Spots Before They Become Errors

Maggie Mi, Aline Villavicencio, Nafise Sadat Moosavi


Abstract
Language models often struggle with idiomatic, figurative, or context-sensitive inputs, not because they produce flawed outputs, but because they misinterpret the input from the outset. We propose an input-only method for anticipating such failures using token-level likelihood features inspired by surprisal and the Uniform Information Density hypothesis. These features capture localized uncertainty in input comprehension and outperform standard baselines across five linguistically challenging datasets. We show that span-localized features improve error detection for larger models, while smaller models benefit from global patterns. Our method requires no access to outputs or hidden activations, offering a lightweight and generalizable approach to pre-generation error prediction.
Anthology ID:
2025.emnlp-main.1740
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
34316–34329
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1740/
DOI:
Bibkey:
Cite (ACL):
Maggie Mi, Aline Villavicencio, and Nafise Sadat Moosavi. 2025. From Input Perception to Predictive Insight: Modeling Model Blind Spots Before They Become Errors. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 34316–34329, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
From Input Perception to Predictive Insight: Modeling Model Blind Spots Before They Become Errors (Mi et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1740.pdf
Checklist:
 2025.emnlp-main.1740.checklist.pdf