@inproceedings{lan-etal-2024-multi,
title = "Multi-label Sequential Sentence Classification via Large Language Model",
author = "Lan, Mengfei and
Zheng, Lecheng and
Ming, Shufan and
Kilicoglu, Halil",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2024",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.findings-emnlp.944/",
doi = "10.18653/v1/2024.findings-emnlp.944",
pages = "16086--16104",
abstract = "Sequential sentence classification (SSC) in scientific publications is crucial for supporting downstream tasks such as fine-grained information retrieval and extractive summarization. However, current SSC methods are constrained by model size, sequence length, and single-label setting. To address these limitations, this paper proposes LLM-SSC, a large language model (LLM)-based framework for both single- and multi-label SSC tasks. Unlike previous approaches that employ small- or medium-sized language models, the proposed framework utilizes LLMs to generate SSC labels through designed prompts, which enhance task understanding by incorporating demonstrations and a query to describe the prediction target. We also present a multi-label contrastive learning loss with auto-weighting scheme, enabling the multi-label classification task. To support our multi-label SSC analysis, we introduce and release a new dataset, biorc800, which mainly contains unstructured abstracts in the biomedical domain with manual annotations. Experiments demonstrate LLM-SSC`s strong performance in SSC under both in-context learning and task-specific tuning settings. We release biorc800 and our code at: https://github.com/ScienceNLP-Lab/LLM-SSC."
}
Markdown (Informal)
[Multi-label Sequential Sentence Classification via Large Language Model](https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.findings-emnlp.944/) (Lan et al., Findings 2024)
ACL