@inproceedings{li-ke-2025-cross,
title = "Cross-Modal Augmentation for Low-Resource Language Understanding and Generation",
author = "Li, Zichao and
Ke, Zong",
editor = "Kriz, Reno and
Murray, Kenton",
booktitle = "Proceedings of the 1st Workshop on Multimodal Augmented Generation via Multimodal Retrieval (MAGMaR 2025)",
month = aug,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/landing_page/2025.magmar-1.9/",
pages = "90--99",
ISBN = "979-8-89176-280-0",
abstract = "This paper introduces a multimodal retrieval-augmented generation (RAG) system designed to enhance language understanding and generation for low-resource languages. By integrating textual, visual, and geospatial data, the system leverages cross-lingual adaptation and multimodal augmentation to bridge the gap between high-resource and low-resource languages. Evaluated on the MM-COVID and LORELEI datasets, the system demonstrates superior performance in retrieval (precision: 85{\%}, recall: 82{\%}) and generation (BLEU: 28.4) tasks compared to baselines. Case studies in public health communication and disaster response highlight its practical utility. The results underscore the potential of multimodal AI to democratize access to technology and address global challenges in low-resource settings."
}
Markdown (Informal)
[Cross-Modal Augmentation for Low-Resource Language Understanding and Generation](https://preview.aclanthology.org/landing_page/2025.magmar-1.9/) (Li & Ke, MAGMaR 2025)
ACL