@inproceedings{colakoglu-etal-2025-problem,
title = "Problem Solved? Information Extraction Design Space for Layout-Rich Documents using {LLM}s",
author = {Colakoglu, Gaye and
Solmaz, G{\"u}rkan and
F{\"u}rst, Jonathan},
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.973/",
doi = "10.18653/v1/2025.findings-emnlp.973",
pages = "17908--17927",
ISBN = "979-8-89176-335-7",
abstract = "This paper defines and explores the design space for information extraction (IE) from layout-rich documents using large language models (LLMs). The three core challenges of layout-aware IE with LLMs are 1) data structuring, 2) model engagement, and 3) output refinement. Our study investigates the sub-problems and methods within these core challenges, such as input representation, chunking, prompting, selection of LLMs, and multimodal models. It examines the effect of different design choices through LayIE-LLM, a new, open-source, layout-aware IE test suite, benchmarking against traditional, fine-tuned IE models. The results on two IE datasets show that LLMs require adjustment of the IE pipeline to achieve competitive performance: the optimized configuration found with LayIE-LLM achieves 13.3{--}37.5 F1 points more than a general-practice baseline configuration using the same LLM. To find a well-working configuration, we develop a one-factor-at-a-time (OFAT) method that achieves near-optimal results. Our method is only 0.8{--}1.8 points lower than the best full factorial exploration with a fraction ({\textasciitilde}2.8{\%}) of the required computation. Overall, we demonstrate that, if well-configured, general-purpose LLMs match the performance of specialized models, providing a cost-effective, finetuning-free alternative. Our test-suite is available at https://github.com/gayecolakoglu/LayIE-LLM"
}Markdown (Informal)
[Problem Solved? Information Extraction Design Space for Layout-Rich Documents using LLMs](https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.973/) (Colakoglu et al., Findings 2025)
ACL