@inproceedings{zhiyu-etal-2024-enhancing,
title = "Enhancing Free-Form Table Question Answering Models by Distilling Relevant-Cell-Based Rationales",
author = "Yang, Zhiyu and
Wang, Shuo and
Yan, Yukun and
Liu, Pengyuan and
Yu, Dong",
editor = "Sun, Maosong and
Liang, Jiye and
Han, Xianpei and
Liu, Zhiyuan and
He, Yulan",
booktitle = "Proceedings of the 23rd Chinese National Conference on Computational Linguistics (Volume 1: Main Conference)",
month = jul,
year = "2024",
address = "Taiyuan, China",
publisher = "Chinese Information Processing Society of China",
url = "https://preview.aclanthology.org/fix-sig-urls/2024.ccl-1.75/",
pages = "973--985",
language = "eng",
abstract = "``Free-form table question answering is a challenging task since tables contain structured contentscompared to plain texts, which requires high-level reasoning abilities to effectively identify cellsthat are relevant to the question and produce a correct and faithful answer based on their relations.Large language models (LLMs) have exhibited remarkable reasoning capabilities in numerousNLP applications. However, in some specific tasks, specially-trained small models can still out-perform LLMs. Furthermore, small models require extremely less computation costs comparedto LLMs. To leverage the strengths of both types of models, we propose a Relevant-Cell-basedKnowledge Distillation with inference-time Teacher Guidance (RCKD-TG) method. This ap-proach aims to combine small free-form table question answering models' abilities to learn fromhuman annotations and large language models' abilities to effectively reason from table contents,via applying Relevant-Cell-based rationales distilled from LLMs to small models' training andinference stages. Our experiments demonstrate the superiority of our method over vanilla smallmodels in correctness, faithfulness, adequacy and fluency, also over general LLMs in adheringto the style of human annotations. We achieve state-of-the-art performance on FeTaQA, a rep-resentative free-form table question answering benchmark. Our result of a 41.3 BLEU scoredemonstrates the feasibility of effectively using small models' task-specific abilities and LLMs{'}reasoning capabilities at the same time. Additionally, our method exhibits high computation ef-ficiency and data efficiency. Compared to strong baselines, we achieve better performance withsignificantly less training data.''"
}
Markdown (Informal)
[Enhancing Free-Form Table Question Answering Models by Distilling Relevant-Cell-Based Rationales](https://preview.aclanthology.org/fix-sig-urls/2024.ccl-1.75/) (Yang et al., CCL 2024)
ACL