AI Tools Can Generate Misculture Visuals! Detecting Prompts Generating Misculture Visuals For Prevention

Venkatesh Velugubantla, Raj Sonani, Msvpj Sathvik


Abstract
Advanced AI models that generate realistic images from text prompts offer new creative possibilities but also risk producing culturally insensitive or offensive content. To address this issue, we introduce a novel dataset designed to classify text prompts that could lead to the generation of harmful images misrepresenting different cultures and communities. By training machine learning models on this dataset, we aim to automatically identify and filter out harmful prompts before image generation, balancing cultural sensitivity with creative freedom. Benchmarking with state-ofthe-art language models, our baseline models achieved an accuracy of 73.34%.
Anthology ID:
2025.nlp4pi-1.25
Volume:
Proceedings of the Fourth Workshop on NLP for Positive Impact (NLP4PI)
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Katherine Atwell, Laura Biester, Angana Borah, Daryna Dementieva, Oana Ignat, Neema Kotonya, Ziyi Liu, Ruyuan Wan, Steven Wilson, Jieyu Zhao
Venues:
NLP4PI | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
285–293
Language:
URL:
https://preview.aclanthology.org/display_plenaries/2025.nlp4pi-1.25/
DOI:
Bibkey:
Cite (ACL):
Venkatesh Velugubantla, Raj Sonani, and Msvpj Sathvik. 2025. AI Tools Can Generate Misculture Visuals! Detecting Prompts Generating Misculture Visuals For Prevention. In Proceedings of the Fourth Workshop on NLP for Positive Impact (NLP4PI), pages 285–293, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
AI Tools Can Generate Misculture Visuals! Detecting Prompts Generating Misculture Visuals For Prevention (Velugubantla et al., NLP4PI 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/display_plenaries/2025.nlp4pi-1.25.pdf