Representations of language in a model of visually grounded speech signal

Grzegorz Chrupała, Lieke Gelderloos, Afra Alishahi


Abstract
We present a visually grounded model of speech perception which projects spoken utterances and images to a joint semantic space. We use a multi-layer recurrent highway network to model the temporal nature of spoken speech, and show that it learns to extract both form and meaning-based linguistic knowledge from the input signal. We carry out an in-depth analysis of the representations used by different components of the trained model and show that encoding of semantic aspects tends to become richer as we go up the hierarchy of layers, whereas encoding of form-related aspects of the language input tends to initially increase and then plateau or decrease.
Anthology ID:
P17-1057
Volume:
Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Month:
July
Year:
2017
Address:
Vancouver, Canada
Editors:
Regina Barzilay, Min-Yen Kan
Venue:
ACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
613–622
Language:
URL:
https://aclanthology.org/P17-1057
DOI:
10.18653/v1/P17-1057
Bibkey:
Cite (ACL):
Grzegorz Chrupała, Lieke Gelderloos, and Afra Alishahi. 2017. Representations of language in a model of visually grounded speech signal. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 613–622, Vancouver, Canada. Association for Computational Linguistics.
Cite (Informal):
Representations of language in a model of visually grounded speech signal (Chrupała et al., ACL 2017)
Copy Citation:
PDF:
https://preview.aclanthology.org/nschneid-patch-3/P17-1057.pdf
Presentation:
 P17-1057.Presentation.pdf
Video:
 https://preview.aclanthology.org/nschneid-patch-3/P17-1057.mp4
Code
 gchrupala/visually-grounded-speech +  additional community code
Data
MS COCOSICK