@inproceedings{sahu-etal-2025-mining,
title = "Mining Contextualized Visual Associations from Images for Creativity Understanding",
author = "Sahu, Ananya and
Ananthram, Amith and
McKeown, Kathleen",
editor = "Flek, Lucie and
Narayan, Shashi and
Phương, L{\^e} Hồng and
Pei, Jiahuan",
booktitle = "Proceedings of the 18th International Natural Language Generation Conference",
month = oct,
year = "2025",
address = "Hanoi, Vietnam",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/author-page-you-zhang-rochester/2025.inlg-main.11/",
pages = "165--181",
abstract = "Understanding another person{'}s creative output requires a shared language of association. However, when training vision-language models such as CLIP, we rely on web-scraped datasets containing short, predominantly literal, alt-text. In this work, we introduce a method for mining contextualized associations for salient visual elements in an image that can scale to any unlabeled dataset. Given an image, we can use these mined associations to generate high quality creative captions at increasing degrees of abstraction. With our method, we produce a new dataset of visual associations and 1.7m creative captions for the images in MSCOCO. Human evaluation confirms that these captions remain visually grounded while exhibiting recognizably increasing abstraction. Moreover, fine-tuning a visual encoder on this dataset yields meaningful improvements in zero-shot image-text retrieval in two creative domains: poetry and metaphor visualization. We release our dataset, our generation code and our models for use by the broader community."
}Markdown (Informal)
[Mining Contextualized Visual Associations from Images for Creativity Understanding](https://preview.aclanthology.org/author-page-you-zhang-rochester/2025.inlg-main.11/) (Sahu et al., INLG 2025)
ACL