Prompt Refinement or Fine-tuning? Best Practices for using LLMs in Computational Social Science Tasks

Anders Giovanni Møller, Luca Maria Aiello


Abstract
Large Language Models are expressive tools that enable complex tasks of text understanding within Computational Social Science. Their versatility, while beneficial, poses a barrier for establishing standardized best practices within the field. To bring clarity on the values of different strategies, we present an overview of the performance of modern LLM-based classification methods on a benchmark of 23 social knowledge tasks. Our results point to three best practices: prioritize models with larger vocabulary and pre-training corpora; avoid simple zero-shot in favor of AI-enhanced prompting; fine-tune on task-specific data, and consider more complex forms instruction-tuning on multiple datasets only when only training data is more abundant.
Anthology ID:
2025.sicon-1.2
Volume:
Proceedings of the Third Workshop on Social Influence in Conversations (SICon 2025)
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
James Hale, Brian Deuksin Kwon, Ritam Dutt
Venues:
SICon | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
27–49
Language:
URL:
https://preview.aclanthology.org/landing_page/2025.sicon-1.2/
DOI:
Bibkey:
Cite (ACL):
Anders Giovanni Møller and Luca Maria Aiello. 2025. Prompt Refinement or Fine-tuning? Best Practices for using LLMs in Computational Social Science Tasks. In Proceedings of the Third Workshop on Social Influence in Conversations (SICon 2025), pages 27–49, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Prompt Refinement or Fine-tuning? Best Practices for using LLMs in Computational Social Science Tasks (Giovanni Møller & Maria Aiello, SICon 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/landing_page/2025.sicon-1.2.pdf