@inproceedings{osei-brefo-liang-2025-oseibrefo,
title = "{O}sei{B}refo-Liang at {S}em{E}val-2025 Task 8 : A Multi-Agent {LLM} code generation approach for answering Tabular Questions",
author = "Osei - Brefo, Emmanuel and
Liang, Huizhi(elly)",
editor = "Rosenthal, Sara and
Ros{\'a}, Aiala and
Ghosh, Debanjan and
Zampieri, Marcos",
booktitle = "Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025)",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/corrections-2025-08/2025.semeval-1.49/",
pages = "343--349",
ISBN = "979-8-89176-273-2",
abstract = "This paper presents a novel multi-agent framework for automated code generation and execution in tabular question answering. Developed for the SemEval-2025 Task 8, our system utilises a structured, multi-agent approach where distinct agents handle dataset extraction, schema identification, prompt engineering, code generation, execution, and prediction. Unlike traditional methods such as semantic parsing-based SQL generation and transformer-based table models such as TAPAS, our approach leverages a large language model-driven code synthesis pipeline using the DeepSeek API. Our system follows a zero-shot inference approach, which generates Python functions that operate directly on structured data. Through the dynamic extraction of dataset schema and intergration into structured prompts, the model comprehension of tabular structures is enhanced, which leads to more precise and interpretable results. Experimental results demonstrate that our system outperforms existing tabular questioning and answering models, achieving an accuracy of 84.67{\%} on DataBench and 86.02{\%} on DataBench-lite, which significantly surpassed the performances of TAPAS (2.68{\%}) and stable-code-3b-GGUF (27{\%}). The source code used in this paper is available at t https://github.com/oseibrefo/semEval25task8"
}
Markdown (Informal)
[OseiBrefo-Liang at SemEval-2025 Task 8 : A Multi-Agent LLM code generation approach for answering Tabular Questions](https://preview.aclanthology.org/corrections-2025-08/2025.semeval-1.49/) (Osei - Brefo & Liang, SemEval 2025)
ACL