@article{liu-etal-2023-visual,
title = "Visual Spatial Reasoning",
author = "Liu, Fangyu and
Emerson, Guy and
Collier, Nigel",
journal = "Transactions of the Association for Computational Linguistics",
volume = "11",
year = "2023",
address = "Cambridge, MA",
publisher = "MIT Press",
url = "https://preview.aclanthology.org/fix-sig-urls/2023.tacl-1.37/",
doi = "10.1162/tacl_a_00566",
pages = "635--651",
abstract = "Spatial relations are a basic part of human cognition. However, they are expressed in natural language in a variety of ways, and previous work has suggested that current vision-and-language models (VLMs) struggle to capture relational information. In this paper, we present Visual Spatial Reasoning (VSR), a dataset containing more than 10k natural text-image pairs with 66 types of spatial relations in English (e.g., under, in front of, facing). While using a seemingly simple annotation format, we show how the dataset includes challenging linguistic phenomena, such as varying reference frames. We demonstrate a large gap between human and model performance: The human ceiling is above 95{\%}, while state-of-the-art models only achieve around 70{\%}. We observe that VLMs' by-relation performances have little correlation with the number of training examples and the tested models are in general incapable of recognising relations concerning the orientations of objects.1"
}
Markdown (Informal)
[Visual Spatial Reasoning](https://preview.aclanthology.org/fix-sig-urls/2023.tacl-1.37/) (Liu et al., TACL 2023)
ACL
- Fangyu Liu, Guy Emerson, and Nigel Collier. 2023. Visual Spatial Reasoning. Transactions of the Association for Computational Linguistics, 11:635–651.