@inproceedings{yi-etal-2024-integrating,
title = "Integrating Structural Semantic Knowledge for Enhanced Information Extraction Pre-training",
author = "Yi, Xiaoyang and
Bao, Yuru and
Zhang, Jian and
Qin, Yifang and
Lin, Faxin",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2024.emnlp-main.129/",
doi = "10.18653/v1/2024.emnlp-main.129",
pages = "2156--2171",
abstract = "Information Extraction (IE), aiming to extract structured information from unstructured natural language texts, can significantly benefit from pre-trained language models. However, existing pre-training methods solely focus on exploiting the textual knowledge, relying extensively on annotated large-scale datasets, which is labor-intensive and thus limits the scalability and versatility of the resulting models. To address these issues, we propose SKIE, a novel pre-training framework tailored for IE that integrates structural semantic knowledge via contrastive learning, effectively alleviating the annotation burden. Specifically, SKIE utilizes Abstract Meaning Representation (AMR) as a low-cost supervision source to boost model performance without human intervention. By enhancing the topology of AMR graphs, SKIE derives high-quality cohesive subgraphs as additional training samples, providing diverse multi-level structural semantic knowledge. Furthermore, SKIE refines the graph encoder to better capture cohesive information and edge relation information, thereby improving the pre-training efficacy. Extensive experimental results demonstrate that SKIE outperforms state-of-the-art baselines across multiple IE tasks and showcases exceptional performance in few-shot and zero-shot settings."
}
Markdown (Informal)
[Integrating Structural Semantic Knowledge for Enhanced Information Extraction Pre-training](https://preview.aclanthology.org/add-emnlp-2024-awards/2024.emnlp-main.129/) (Yi et al., EMNLP 2024)
ACL