Preethi Vaidyanathan


2018

pdf bib
SNAG: Spoken Narratives and Gaze Dataset
Preethi Vaidyanathan | Emily T. Prud’hommeaux | Jeff B. Pelz | Cecilia O. Alm
Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)

Humans rely on multiple sensory modalities when examining and reasoning over images. In this paper, we describe a new multimodal dataset that consists of gaze measurements and spoken descriptions collected in parallel during an image inspection task. The task was performed by multiple participants on 100 general-domain images showing everyday objects and activities. We demonstrate the usefulness of the dataset by applying an existing visual-linguistic data fusion framework in order to label important image regions with appropriate linguistic labels.

2015

pdf bib
Alignment of Eye Movements and Spoken Language for Semantic Image Understanding
Preethi Vaidyanathan | Emily Prud’hommeaux | Cecilia O. Alm | Jeff B. Pelz | Anne R. Haake
Proceedings of the 11th International Conference on Computational Semantics

pdf bib
Computational Integration of Human Vision and Natural Language through Bitext Alignment
Preethi Vaidyanathan | Emily Prud’hommeaux | Cecilia O. Alm | Jeff B. Pelz | Anne R. Haake
Proceedings of the Fourth Workshop on Vision and Language