Beyond Seen Data: Improving KBQA Generalization Through Schema-Guided Logical Form Generation

Shengxiang Gao, Jey Han Lau, Jianzhong Qi


Abstract
Knowledge base question answering (KBQA) aims to answer user questions in natural language using rich human knowledge stored in large KBs. As current KBQA methods struggle with unseen knowledge base elements and their novel compositions at test time, we introduce SG-KBQA — a novel model that injects schema contexts into entity retrieval and logical form generation to tackle this issue. It exploits information about the semantics and structure of the knowledge base provided by schema contexts to enhance generalizability. We show that achieves strong generalizability, outperforming state-of-the-art models on two commonly used benchmark datasets across a variety of test settings. Our source code is available at https://github.com/gaosx2000/SG_KBQA.
Anthology ID:
2025.emnlp-main.442
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
8764–8783
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.442/
DOI:
Bibkey:
Cite (ACL):
Shengxiang Gao, Jey Han Lau, and Jianzhong Qi. 2025. Beyond Seen Data: Improving KBQA Generalization Through Schema-Guided Logical Form Generation. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 8764–8783, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Beyond Seen Data: Improving KBQA Generalization Through Schema-Guided Logical Form Generation (Gao et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.442.pdf
Checklist:
 2025.emnlp-main.442.checklist.pdf