@inproceedings{kermani-etal-2025-systematic,
title = "A Systematic Evaluation of {LLM} Strategies for Mental Health Text Analysis: Fine-tuning vs. Prompt Engineering vs. {RAG}",
author = "Kermani, Arshia and
Perez-Rosas, Veronica and
Metsis, Vangelis",
editor = "Zirikly, Ayah and
Yates, Andrew and
Desmet, Bart and
Ireland, Molly and
Bedrick, Steven and
MacAvaney, Sean and
Bar, Kfir and
Ophir, Yaakov",
booktitle = "Proceedings of the 10th Workshop on Computational Linguistics and Clinical Psychology (CLPsych 2025)",
month = may,
year = "2025",
address = "Albuquerque, New Mexico",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2025.clpsych-1.14/",
pages = "172--180",
ISBN = "979-8-89176-226-8",
abstract = "This study presents a systematic comparison of three approaches for the analysis of mental health text using large language models (LLMs): prompt engineering, retrieval augmented generation (RAG), and fine-tuning. Using LLaMA 3, we evaluate these approaches on emotion classification and mental health condition detection tasks across two datasets. Fine-tuning achieves the highest accuracy (91{\%} for emotion classification, 80{\%} for mental health conditions) but requires substantial computational resources and large training sets, while prompt engineering and RAG offer more flexible deployment with moderate performance (40-68{\%} accuracy). Our findings provide practical insights for implementing LLM-based solutions in mental health applications, highlighting the trade-offs between accuracy, computational requirements, and deployment flexibility."
}
Markdown (Informal)
[A Systematic Evaluation of LLM Strategies for Mental Health Text Analysis: Fine-tuning vs. Prompt Engineering vs. RAG](https://preview.aclanthology.org/fix-sig-urls/2025.clpsych-1.14/) (Kermani et al., CLPsych 2025)
ACL