Do we still need Human Annotators? Prompting Large Language Models for Aspect Sentiment Quad Prediction

Nils Hellwig, Jakob Fehle, Udo Kruschwitz, Christian Wolff


Abstract
Aspect sentiment quad prediction (ASQP) facilitates a detailed understanding of opinions expressed in a text by identifying the opinion term, aspect term, aspect category and sentiment polarity for each opinion. However, annotating a full set of training examples to fine-tune models for ASQP is a resource-intensive process. In this study, we explore the capabilities of large language models (LLMs) for zero- and few-shot learning on the ASQP task across five diverse datasets. We report F1 scores almost up to par with those obtained with state-of-the-art fine-tuned models and exceeding previously reported zero- and few-shot performance. In the 20-shot setting on the Rest16 restaurant domain dataset, LLMs achieved an F1 score of 51.54, compared to 60.39 by the best-performing fine-tuned method MVP. Additionally, we report the performance of LLMs in target aspect sentiment detection (TASD), where the F1 scores were close to fine-tuned models, achieving 68.93 on Rest16 in the 30-shot setting, compared to 72.76 with MVP. While human annotators remain essential for achieving optimal performance, LLMs can reduce the need for extensive manual annotation in ASQP tasks.
Anthology ID:
2025.xllm-1.15
Volume:
Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025)
Month:
August
Year:
2025
Address:
Vienna, Austria
Editors:
Hao Fei, Kewei Tu, Yuhui Zhang, Xiang Hu, Wenjuan Han, Zixia Jia, Zilong Zheng, Yixin Cao, Meishan Zhang, Wei Lu, N. Siddharth, Lilja Øvrelid, Nianwen Xue, Yue Zhang
Venues:
XLLM | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
153–172
Language:
URL:
https://preview.aclanthology.org/landing_page/2025.xllm-1.15/
DOI:
Bibkey:
Cite (ACL):
Nils Hellwig, Jakob Fehle, Udo Kruschwitz, and Christian Wolff. 2025. Do we still need Human Annotators? Prompting Large Language Models for Aspect Sentiment Quad Prediction. In Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025), pages 153–172, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Do we still need Human Annotators? Prompting Large Language Models for Aspect Sentiment Quad Prediction (Hellwig et al., XLLM 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/landing_page/2025.xllm-1.15.pdf