Abstract
The widely used Fact-based Visual Question Answering (FVQA) dataset contains visually-grounded questions that require information retrieval using common sense knowledge graphs to answer. It has been observed that the original dataset is highly imbalanced and concentrated on a small portion of its associated knowledge graph. We introduce FVQA 2.0 which contains adversarial variants of test questions to address this imbalance. We show that systems trained with the original FVQA train sets can be vulnerable to adversarial samples and we demonstrate an augmentation scheme to reduce this vulnerability without human annotations.- Anthology ID:
- 2023.findings-eacl.11
- Volume:
- Findings of the Association for Computational Linguistics: EACL 2023
- Month:
- May
- Year:
- 2023
- Address:
- Dubrovnik, Croatia
- Editors:
- Andreas Vlachos, Isabelle Augenstein
- Venue:
- Findings
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 149–157
- Language:
- URL:
- https://aclanthology.org/2023.findings-eacl.11
- DOI:
- 10.18653/v1/2023.findings-eacl.11
- Cite (ACL):
- Weizhe Lin, Zhilin Wang, and Bill Byrne. 2023. FVQA 2.0: Introducing Adversarial Samples into Fact-based Visual Question Answering. In Findings of the Association for Computational Linguistics: EACL 2023, pages 149–157, Dubrovnik, Croatia. Association for Computational Linguistics.
- Cite (Informal):
- FVQA 2.0: Introducing Adversarial Samples into Fact-based Visual Question Answering (Lin et al., Findings 2023)
- PDF:
- https://preview.aclanthology.org/ingest-2024-clasp/2023.findings-eacl.11.pdf