Preethi Vaidyanathan
2018
SNAG: Spoken Narratives and Gaze Dataset
Preethi Vaidyanathan
|
Emily T. Prud’hommeaux
|
Jeff B. Pelz
|
Cecilia O. Alm
Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)
Humans rely on multiple sensory modalities when examining and reasoning over images. In this paper, we describe a new multimodal dataset that consists of gaze measurements and spoken descriptions collected in parallel during an image inspection task. The task was performed by multiple participants on 100 general-domain images showing everyday objects and activities. We demonstrate the usefulness of the dataset by applying an existing visual-linguistic data fusion framework in order to label important image regions with appropriate linguistic labels.
2015
Alignment of Eye Movements and Spoken Language for Semantic Image Understanding
Preethi Vaidyanathan
|
Emily Prud’hommeaux
|
Cecilia O. Alm
|
Jeff B. Pelz
|
Anne R. Haake
Proceedings of the 11th International Conference on Computational Semantics
Computational Integration of Human Vision and Natural Language through Bitext Alignment
Preethi Vaidyanathan
|
Emily Prud’hommeaux
|
Cecilia O. Alm
|
Jeff B. Pelz
|
Anne R. Haake
Proceedings of the Fourth Workshop on Vision and Language
Search