Zhenbin Chen
2026
MMUIE: Massive Multi-Domain Universal Information Extraction for Long Documents
Shuyi Zhang | Zhenbin Chen | Shuting Li | Kewei Tu | Li Jing | Zixia Jia | Zilong Zheng
Findings of the Association for Computational Linguistics: EACL 2026
Shuyi Zhang | Zhenbin Chen | Shuting Li | Kewei Tu | Li Jing | Zixia Jia | Zilong Zheng
Findings of the Association for Computational Linguistics: EACL 2026
We present **MMUIE**, a large-scale universal dataset for multi-domain, document-level information extraction (IE) from long texts.Existing IE systems predominantly operate at the sentence level or within narrow domains due to annotation constraints.MMUIE addresses this gap by introducing an automated annotation pipeline that integrates traditional knowledge bases with large language models to extract fine-grained entities, aliases, and relation triples across 34 domains.The dataset comprises a weakly-supervised training set and a manually verified test set, featuring 723 entity types and 456 relation types.Empirical evaluations reveal that existing sentence-level IE models and even advanced LLMs underperform on this task, highlighting the need for better domain-aware document-level models.To this end, we develop DocUIE, a universal IE model fine-tuned on MMUIE, which achieves strong generalization and transferability across domains. MMUIE lays the foundation for robust, scalable, and universal information extraction from long-form text in diverse real-world scenarios. All code, data, and models are available in https://github.com/Shuyi-zsy/Massive-Multi-Domain-UIE.