@inproceedings{mohammadshahi-etal-2019-aligning-multilingual,
    title = "Aligning Multilingual Word Embeddings for Cross-Modal Retrieval Task",
    author = "Mohammadshahi, Alireza  and
      Lebret, R{\'e}mi  and
      Aberer, Karl",
    editor = "Thorne, James  and
      Vlachos, Andreas  and
      Cocarascu, Oana  and
      Christodoulopoulos, Christos  and
      Mittal, Arpit",
    booktitle = "Proceedings of the Second Workshop on Fact Extraction and VERification (FEVER)",
    month = nov,
    year = "2019",
    address = "Hong Kong, China",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/iwcs-25-ingestion/D19-6605/",
    doi = "10.18653/v1/D19-6605",
    pages = "27--33",
    abstract = "In this paper, we propose a new approach to learn multimodal multilingual embeddings for matching images and their relevant captions in two languages. We combine two existing objective functions to make images and captions close in a joint embedding space while adapting the alignment of word embeddings between existing languages in our model. We show that our approach enables better generalization, achieving state-of-the-art performance in text-to-image and image-to-text retrieval task, and caption-caption similarity task. Two multimodal multilingual datasets are used for evaluation: Multi30k with German and English captions and Microsoft-COCO with English and Japanese captions."
}Markdown (Informal)
[Aligning Multilingual Word Embeddings for Cross-Modal Retrieval Task](https://preview.aclanthology.org/iwcs-25-ingestion/D19-6605/) (Mohammadshahi et al., 2019)
ACL