General Collaborative Framework between Large Language Model and Experts for Universal Information Extraction

K Bao, Ning Wang


Abstract
Recently, unified information extraction has garnered widespread attention from the NLP community, which aims to use a unified paradigm to perform various information extraction tasks. However, prevalent unified IE approaches inevitably encounter challenges such as noise interference, abstract label semantics, and diverse span granularity. In this paper, we first present three problematic assumptions regarding the capabilities of unified information extraction model. Furthermore, we propose the General Collaborative Information Extraction (GCIE) framework to address these challenges in universal information extraction tasks. Specifically, GCIE consists of a general Recognizer as well as multiple task-specific Experts for recognizing predefined types and extracting spans respectively. The Recognizer is a large language model, while the Experts comprise a series of smaller language models. Together, they collaborate in a two-stage pipeline to perform unified information extraction. Extensive empirical experiments on 6 IE tasks and several datasets, validate the effectiveness and generality of our approach.
Anthology ID:
2024.findings-emnlp.4
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2024
Month:
November
Year:
2024
Address:
Miami, Florida, USA
Editors:
Yaser Al-Onaizan, Mohit Bansal, Yun-Nung Chen
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
52–77
Language:
URL:
https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.findings-emnlp.4/
DOI:
10.18653/v1/2024.findings-emnlp.4
Bibkey:
Cite (ACL):
K Bao and Ning Wang. 2024. General Collaborative Framework between Large Language Model and Experts for Universal Information Extraction. In Findings of the Association for Computational Linguistics: EMNLP 2024, pages 52–77, Miami, Florida, USA. Association for Computational Linguistics.
Cite (Informal):
General Collaborative Framework between Large Language Model and Experts for Universal Information Extraction (Bao & Wang, Findings 2024)
Copy Citation:
PDF:
https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.findings-emnlp.4.pdf