CIVET: Systematic Evaluation of Understanding in VLMs

Massimo Rizzoli, Simone Alghisi, Olha Khomyn, Gabriel Roccabruna, Seyed Mahed Mousavi, Giuseppe Riccardi


Abstract
While Vision-Language Models (VLMs) have achieved competitive performance in various tasks, their comprehension of the underlying structure and semantics of a scene remains understudied. To investigate the understanding of VLMs, we study their capability regarding object properties and relations in a controlled and interpretable manner. To this scope, we introduce CIVET, a novel and extensible framework for systematiC evaluatIon Via controllEd sTimuli. CIVET addresses the lack of standardized systematic evaluation for assessing VLMs’ understanding, enabling researchers to test hypotheses with statistical rigor. With CIVET, we evaluate five state-of-the-art VLMs on exhaustive sets of stimuli, free from annotation noise, dataset-specific biases, and uncontrolled scene complexity. Our findings reveal that 1) current VLMs can accurately recognize only a limited set of basic object properties; 2) their performance heavily depends on the position of the object in the scene; 3) they struggle to understand basic relations among objects. Furthermore, a comparative evaluation with human annotators reveals that VLMs still fall short of achieving human-level accuracy.
Anthology ID:
2025.findings-emnlp.239
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2025
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
4462–4480
Language:
URL:
https://preview.aclanthology.org/author-page-jing-yang-campinas/2025.findings-emnlp.239/
DOI:
10.18653/v1/2025.findings-emnlp.239
Bibkey:
Cite (ACL):
Massimo Rizzoli, Simone Alghisi, Olha Khomyn, Gabriel Roccabruna, Seyed Mahed Mousavi, and Giuseppe Riccardi. 2025. CIVET: Systematic Evaluation of Understanding in VLMs. In Findings of the Association for Computational Linguistics: EMNLP 2025, pages 4462–4480, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
CIVET: Systematic Evaluation of Understanding in VLMs (Rizzoli et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/author-page-jing-yang-campinas/2025.findings-emnlp.239.pdf
Checklist:
 2025.findings-emnlp.239.checklist.pdf