Rethinking Reading Order: Toward Generalizable Document Understanding with LLM-based Relation Modeling

Weishi Wang, Hengchang Hu, Daniel Dahlmeier


Abstract
Document understanding requires modeling both structural and semantic relationships between the layout elements within the document, with human-perceived reading order (RO) playing a crucial yet often neglected role compared to heuristic OCR sequences used by most existing models. Previous approaches depend on costly, inconsistent human annotations, limiting scalability and generalization. To bridge the gap, we propose a cost-effective paradigm that leverages large language models (LLMs) to infer global RO and inter-element layout relations without human supervision. By explicitly incorporating RO as structural guidance, our method captures hierarchical, document-level dependencies beyond local adjacency. Experiments on Semantic Entity Recognition, Entity Linking, and Document Question Answering show consistent improvements over baseline methods. Notably, LLM-inferred RO, even when differing from ground-truth adjacency, provides richer global structural priors and yields superior downstream performance. These results and findings demonstrate the scalability and significance of RO-aware modeling, advancing both LLMs and lightweight layout-aware models for robust document understanding. Code, data, and more details will be made publicly available after corporate review, in accordance with SAP’s corporate open-source policy.
Anthology ID:
2026.eacl-long.192
Volume:
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Month:
March
Year:
2026
Address:
Rabat, Morocco
Editors:
Vera Demberg, Kentaro Inui, Lluís Marquez
Venue:
EACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
4110–4130
Language:
URL:
https://preview.aclanthology.org/ingest-eacl/2026.eacl-long.192/
DOI:
Bibkey:
Cite (ACL):
Weishi Wang, Hengchang Hu, and Daniel Dahlmeier. 2026. Rethinking Reading Order: Toward Generalizable Document Understanding with LLM-based Relation Modeling. In Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers), pages 4110–4130, Rabat, Morocco. Association for Computational Linguistics.
Cite (Informal):
Rethinking Reading Order: Toward Generalizable Document Understanding with LLM-based Relation Modeling (Wang et al., EACL 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-eacl/2026.eacl-long.192.pdf