Sjoerd Van Steenkiste
2025
DreamSync: Aligning Text-to-Image Generation with Image Understanding Feedback
Jiao Sun
|
Deqing Fu
|
Yushi Hu
|
Su Wang
|
Royi Rassin
|
Da-Cheng Juan
|
Dana Alon
|
Charles Herrmann
|
Sjoerd Van Steenkiste
|
Ranjay Krishna
|
Cyrus Rashtchian
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Despite their widespread success, Text-to-Image models (T2I) still struggle to produce images that are both aesthetically pleasing and faithful to the user’s input text. We introduce DreamSync, a simple yet effective training algorithm that improves T2I models to be faithful to the text input. DreamSync utilizes large vision-language models (VLMs) to effectively identify the fine-grained discrepancies between generated images and the text inputs and enable T2I models to self-improve without labeled data. First, it prompts the model to generate several candidate images for a given input text. Then, it uses two VLMs to select the best generation: a Visual Question Answering model that measures the alignment of generated images to the text, and another that measures the generation’s aesthetic quality. After selection, we use LoRA to iteratively finetune the T2I model to guide its generation towards the selected best generations. DreamSync does not need any additional human annotation, model architecture changes, or reinforcement learning. Despite its simplicity, DreamSync improves both the semantic alignment and aesthetic appeal of two diffusion-based T2I models, evidenced by multiple benchmarks (+1.7% on TIFA, +2.9% on DSG1K, +3.4% on VILA aesthetic) and human evaluation shows that DreamSync improves text rendering compared to SDXL by 18.5% on DSG1K benchmark.
2024
Benchmarking Vision Language Models for Cultural Understanding
Shravan Nayak
|
Kanishk Jain
|
Rabiul Awal
|
Siva Reddy
|
Sjoerd Van Steenkiste
|
Lisa Anne Hendricks
|
Karolina Stanczak
|
Aishwarya Agrawal
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
Foundation models and vision-language pre-training have notably advanced Vision Language Models (VLMs), enabling multimodal processing of visual and linguistic data. However, their performance has been typically assessed on general scene understanding - recognizing objects, attributes, and actions - rather than cultural comprehension. This study introduces CulturalVQA, a visual question-answering benchmark aimed at assessing VLM’s geo-diverse cultural understanding. We curate a diverse collection of 2,378 image-question pairs with 1-5 answers per question representing cultures from 11 countries across 5 continents. The questions probe understanding of various facets of culture such as clothing, food, drinks, rituals, and traditions. Benchmarking VLMs on CulturalVQA, including GPT-4V and Gemini, reveals disparity in their level of cultural understanding across regions, with strong cultural understanding capabilities for North America while significantly weaker capabilities for Africa. We observe disparity in their performance across cultural facets too, with clothing, rituals, and traditions seeing higher performances than food and drink. These disparities help us identify areas where VLMs lack cultural understanding and demonstrate the potential of CulturalVQA as a comprehensive evaluation set for gauging VLM progress in understanding diverse cultures.
Search
Fix data
Co-authors
- Aishwarya Agrawal 1
- Dana Alon 1
- Rabiul Awal 1
- Deqing Fu 1
- Lisa Anne Hendricks 1
- show all...