Revealing the Numeracy Gap: An Empirical Investigation of Text Embedding Models

Ningyuan Deng, Hanyu Duan, Yixuan Tang, Yi Yang


Abstract
Text embedding models are widely used in natural language processing applications. However, their capability is often benchmarked on tasks that do not require understanding nuanced numerical information in text. As a result, it remains unclear whether current embedding models can precisely encode numerical content, such as numbers, into embeddings. This question is critical because embedding models are increasingly applied in domains where numbers matter, such as finance and healthcare. For example, ”Company X’s market share grew by 2%” should be interpreted very differently from ”Company X’s market share grew by 20%” , even though both indicate growth in market share. This study aims to examine whether text embedding models can capture such nuances. Using synthetic data in a financial context, we evaluate 13 widely used text embedding models and find that they generally struggle to capture numerical details accurately. Our further analyses provide deeper insights into embedding numeracy, informing future research to strengthen embedding model-based NLP systems with improved capacity for handling numerical content.
Anthology ID:
2026.findings-eacl.289
Volume:
Findings of the Association for Computational Linguistics: EACL 2026
Month:
March
Year:
2026
Address:
Rabat, Morocco
Editors:
Vera Demberg, Kentaro Inui, Lluís Marquez
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
5448–5461
Language:
URL:
https://preview.aclanthology.org/ingest-eacl/2026.findings-eacl.289/
DOI:
Bibkey:
Cite (ACL):
Ningyuan Deng, Hanyu Duan, Yixuan Tang, and Yi Yang. 2026. Revealing the Numeracy Gap: An Empirical Investigation of Text Embedding Models. In Findings of the Association for Computational Linguistics: EACL 2026, pages 5448–5461, Rabat, Morocco. Association for Computational Linguistics.
Cite (Informal):
Revealing the Numeracy Gap: An Empirical Investigation of Text Embedding Models (Deng et al., Findings 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-eacl/2026.findings-eacl.289.pdf
Checklist:
 2026.findings-eacl.289.checklist.pdf