TabComp: A Dataset for Visual Table Reading Comprehension

Somraj Gautam, Abhishek Bhandari, Gaurav Harit


Abstract
Reaching a human-level understanding of real-world documents necessitates effective machine reading comprehension, yet recent developments in this area often struggle with table images. In response, we introduce the Visual Table Reading Comprehension (TabComp) dataset, which includes table images, questions, and generative answers designed to evaluate OCR-free models. Unlike general Visual Question Answering (VQA) datasets, TabComp uniquely focuses on table images, fostering the development of systems which obviate the use of optical character recognition (OCR) technology, which often struggles with complex table layouts. Our findings reveal that current OCR-free models perform poorly on TabComp, highlighting the need for robust, specialized models for accurate table reading comprehension. We propose TabComp as a benchmark for evaluating OCR-free models in table reading comprehension and encourage the research community to collaborate on developing more effective solutions. The code and data are available at - https://github.com/dialabiitj/TabComp/
Anthology ID:
2025.findings-naacl.320
Volume:
Findings of the Association for Computational Linguistics: NAACL 2025
Month:
April
Year:
2025
Address:
Albuquerque, New Mexico
Editors:
Luis Chiruzzo, Alan Ritter, Lu Wang
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
5773–5780
Language:
URL:
https://preview.aclanthology.org/corrections-2025-06/2025.findings-naacl.320/
DOI:
10.18653/v1/2025.findings-naacl.320
Bibkey:
Cite (ACL):
Somraj Gautam, Abhishek Bhandari, and Gaurav Harit. 2025. TabComp: A Dataset for Visual Table Reading Comprehension. In Findings of the Association for Computational Linguistics: NAACL 2025, pages 5773–5780, Albuquerque, New Mexico. Association for Computational Linguistics.
Cite (Informal):
TabComp: A Dataset for Visual Table Reading Comprehension (Gautam et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/corrections-2025-06/2025.findings-naacl.320.pdf