@inproceedings{jhamtani-berg-kirkpatrick-2018-learning,
title = "Learning to Describe Differences Between Pairs of Similar Images",
author = "Jhamtani, Harsh and
Berg-Kirkpatrick, Taylor",
editor = "Riloff, Ellen and
Chiang, David and
Hockenmaier, Julia and
Tsujii, Jun{'}ichi",
booktitle = "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing",
month = oct # "-" # nov,
year = "2018",
address = "Brussels, Belgium",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/D18-1436/",
doi = "10.18653/v1/D18-1436",
pages = "4024--4034",
abstract = "In this paper, we introduce the task of automatically generating text to describe the differences between two similar images. We collect a new dataset by crowd-sourcing difference descriptions for pairs of image frames extracted from video-surveillance footage. Annotators were asked to succinctly describe all the differences in a short paragraph. As a result, our novel dataset provides an opportunity to explore models that align language and vision, and capture visual salience. The dataset may also be a useful benchmark for coherent multi-sentence generation. We perform a first-pass visual analysis that exposes clusters of differing pixels as a proxy for object-level differences. We propose a model that captures visual salience by using a latent variable to align clusters of differing pixels with output sentences. We find that, for both single-sentence generation and as well as multi-sentence generation, the proposed model outperforms the models that use attention alone."
}
Markdown (Informal)
[Learning to Describe Differences Between Pairs of Similar Images](https://preview.aclanthology.org/fix-sig-urls/D18-1436/) (Jhamtani & Berg-Kirkpatrick, EMNLP 2018)
ACL