A Comparative Analysis of Ethical and Safety Gaps in LLMs using Relative Danger Coefficient

Yehor Tereshchenko, Mika Hämäläinen


Abstract
Artificial Intelligence (AI) and Large Language Models (LLMs) have rapidly evolved in recent years, showcasing remarkable capabilities in natural language understanding and generation. However, these advancements also raise critical ethical questions regarding safety, potential misuse, discrimination and overall societal impact. This article provides a comparative analysis of the ethical performance of various AI models, including the brand new DeepSeek-V3(R1 with reasoning and without), various GPT variants (4o, 3.5 Turbo, 4 Turbo, o1/o3 mini) and Gemini (1.5 flash, 2.0 flash and 2.0 flash exp) and highlights the need for robust human oversight, especially in situations with high stakes. Furthermore, we present a new metric for calculating harm in LLMs called Relative Danger Coefficient (RDC).
Anthology ID:
2025.nlp4dh-1.40
Volume:
Proceedings of the 5th International Conference on Natural Language Processing for Digital Humanities
Month:
May
Year:
2025
Address:
Albuquerque, USA
Editors:
Mika Hämäläinen, Emily Öhman, Yuri Bizzoni, So Miyagawa, Khalid Alnajjar
Venues:
NLP4DH | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
464–477
Language:
URL:
https://preview.aclanthology.org/landing_page/2025.nlp4dh-1.40/
DOI:
Bibkey:
Cite (ACL):
Yehor Tereshchenko and Mika Hämäläinen. 2025. A Comparative Analysis of Ethical and Safety Gaps in LLMs using Relative Danger Coefficient. In Proceedings of the 5th International Conference on Natural Language Processing for Digital Humanities, pages 464–477, Albuquerque, USA. Association for Computational Linguistics.
Cite (Informal):
A Comparative Analysis of Ethical and Safety Gaps in LLMs using Relative Danger Coefficient (Tereshchenko & Hämäläinen, NLP4DH 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/landing_page/2025.nlp4dh-1.40.pdf