Abstract
Hallucinations in large language models (LLMs), where they generate fluent but factually incorrect outputs, pose challenges for applications requiring strict truthfulness. This work proposes a multi-faceted approach to detect such hallucinations across various language tasks. We leverage automatic data annotation using a proprietary LLM, fine-tuning of the Mistral-7B-instruct-v0.2 model on annotated and benchmark data, role-based and rationale-based prompting strategies, and an ensemble method combining different model outputs through majority voting. This comprehensive framework aims to improve the robustness and reliability of hallucination detection for LLM generations.- Anthology ID:
- 2024.semeval-1.208
- Volume:
- Proceedings of the 18th International Workshop on Semantic Evaluation (SemEval-2024)
- Month:
- June
- Year:
- 2024
- Address:
- Mexico City, Mexico
- Editors:
- Atul Kr. Ojha, A. Seza Doğruöz, Harish Tayyar Madabushi, Giovanni Da San Martino, Sara Rosenthal, Aiala Rosá
- Venue:
- SemEval
- SIG:
- SIGLEX
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 1449–1454
- Language:
- URL:
- https://preview.aclanthology.org/add_missing_videos/2024.semeval-1.208/
- DOI:
- 10.18653/v1/2024.semeval-1.208
- Cite (ACL):
- Souvik Das and Rohini Srihari. 2024. Compos Mentis at SemEval2024 Task6: A Multi-Faceted Role-based Large Language Model Ensemble to Detect Hallucination. In Proceedings of the 18th International Workshop on Semantic Evaluation (SemEval-2024), pages 1449–1454, Mexico City, Mexico. Association for Computational Linguistics.
- Cite (Informal):
- Compos Mentis at SemEval2024 Task6: A Multi-Faceted Role-based Large Language Model Ensemble to Detect Hallucination (Das & Srihari, SemEval 2024)
- PDF:
- https://preview.aclanthology.org/add_missing_videos/2024.semeval-1.208.pdf