@inproceedings{pang-etal-2023-guideline,
title = "Guideline Learning for In-Context Information Extraction",
author = "Pang, Chaoxu and
Cao, Yixuan and
Ding, Qiang and
Luo, Ping",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2023.emnlp-main.950/",
doi = "10.18653/v1/2023.emnlp-main.950",
pages = "15372--15389",
abstract = "Large language models (LLMs) can perform a new task by merely conditioning on task instructions and a few input-output examples, without optimizing any parameters. This is called In-Context Learning (ICL). In-context Information Extraction (IE) has recently garnered attention in the research community. However, the performance of In-context IE generally lags behind the state-of-the-art supervised expert models. We highlight a key reason for this shortfall: underspecified task description. The limited-length context struggles to thoroughly express the intricate IE task instructions and various edge cases, leading to misalignment in task comprehension with humans. In this paper, we propose a Guideline Learning (GL) framework for In-context IE which reflectively learns and follows guidelines. During the learning phrase, GL automatically synthesizes a set of guidelines based on a few error cases, and during inference, GL retrieves helpful guidelines for better ICL. Moreover, we propose a self-consistency-based active learning method to enhance the efficiency of GL. Experiments on event extraction and relation extraction show that GL can significantly improve the performance of in-context IE."
}
Markdown (Informal)
[Guideline Learning for In-Context Information Extraction](https://preview.aclanthology.org/fix-sig-urls/2023.emnlp-main.950/) (Pang et al., EMNLP 2023)
ACL