Instruction-tuned QwenChart for Chart Question Answering

Viviana Ventura, Lukas Amadeus Kleybolte, Alessandra Zarcone


Abstract
Charts, where information is delivered holistically by visual and textual features, represent a challenge when it comes to downstream tasks such as chart question answering, where both kinds of information contribute to the task. The standard approach is to decouple the task in two steps, first extracting information from the charts, or representing it as a table, text or code, and then a second reasoning step to output the answers. Today, the advancements in visual encoding of Visual Large Language Models (VLLM) have shown their capabilities to solve such complex tasks without using in-between representations of the charts or massive in-domain training. Our new instruction fine-tuned and chain-of-thought model QwenChart showed that even in a complex new benchmark such as SciVQA general models can achieve great performances with low-cost training, matching the capabilities that LLMs have showed in unimodal downstream tasks. An out-of-domain evaluation showed satisfactory results, albeit with an expected drop in performance.
Anthology ID:
2025.sdp-1.22
Volume:
Proceedings of the Fifth Workshop on Scholarly Document Processing (SDP 2025)
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Tirthankar Ghosal, Philipp Mayr, Amanpreet Singh, Aakanksha Naik, Georg Rehm, Dayne Freitag, Dan Li, Sonja Schimmler, Anita De Waard
Venues:
sdp | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
240–251
Language:
URL:
https://preview.aclanthology.org/tal-24-ingestion/2025.sdp-1.22/
DOI:
10.18653/v1/2025.sdp-1.22
Bibkey:
Cite (ACL):
Viviana Ventura, Lukas Amadeus Kleybolte, and Alessandra Zarcone. 2025. Instruction-tuned QwenChart for Chart Question Answering. In Proceedings of the Fifth Workshop on Scholarly Document Processing (SDP 2025), pages 240–251, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Instruction-tuned QwenChart for Chart Question Answering (Ventura et al., sdp 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/tal-24-ingestion/2025.sdp-1.22.pdf