@inproceedings{dallasen-etal-2024-retrieval,
title = "Retrieval-enriched zero-shot image classification in low-resource domains",
author = "Dall{'}Asen, Nicola and
Wang, Yiming and
Fini, Enrico and
Ricci, Elisa",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2024.emnlp-main.1186/",
doi = "10.18653/v1/2024.emnlp-main.1186",
pages = "21287--21302",
abstract = "Low-resource domains, characterized by scarce data and annotations, present significant challenges for language and visual understanding tasks, with the latter much under-explored in the literature. Recent advancements in Vision-Language Models (VLM) have shown promising results in high-resource domains but fall short in low-resource concepts that are under-represented (e.g. only a handful of images per category) in the pre-training set. We tackle the challenging task of zero-shot low-resource image classification from a novel perspective. By leveraging a retrieval-based strategy, we achieve this in a training-free fashion. Specifically, our method, named CoRE (Combination of Retrieval Enrichment), enriches the representation of both query images and class prototypes by retrieving relevant textual information from large web-crawled databases. This retrieval-based enrichment significantly boosts classification performance by incorporating the broader contextual information relevant to the specific class. We validate our method on a newly established benchmark covering diverse low-resource domains, including medical imaging, rare plants, and circuits. Our experiments demonstrate that CoRE outperforms existing state-of-the-art methods that rely on synthetic data generation and model fine-tuning."
}
Markdown (Informal)
[Retrieval-enriched zero-shot image classification in low-resource domains](https://preview.aclanthology.org/add-emnlp-2024-awards/2024.emnlp-main.1186/) (Dall’Asen et al., EMNLP 2024)
ACL