TableKV: KV Cache Compression for In-Context Table Processing

Giulio Corallo, Elia Faure-Rolland, Miriam Lamari, Paolo Papotti


Abstract
Processing large tables provided in-context to LLMs is challenging due to token limits and information overload. While Retrieval-Augmented Generation can select relevant subsets externally, this work explores Key-Value (KV) cache compression as an alternative, applied directly to the linearized table during inference. We show that the LLM’s internal attention scores over the table context guides the retention of essential KV pairs, effectively compressing the processing context while preserving crucial relational information needed for complex queries. Experiments on Spider, WikitableQA, and QTSumm datasets validate the compression approach for in-context table processing, offering a promising path for improved table representation learning in LLMs.
Anthology ID:
2025.trl-1.13
Volume:
Proceedings of the 4th Table Representation Learning Workshop
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Shuaichen Chang, Madelon Hulsebos, Qian Liu, Wenhu Chen, Huan Sun
Venues:
TRL | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
166–171
Language:
URL:
https://preview.aclanthology.org/acl25-workshop-ingestion/2025.trl-1.13/
DOI:
Bibkey:
Cite (ACL):
Giulio Corallo, Elia Faure-Rolland, Miriam Lamari, and Paolo Papotti. 2025. TableKV: KV Cache Compression for In-Context Table Processing. In Proceedings of the 4th Table Representation Learning Workshop, pages 166–171, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
TableKV: KV Cache Compression for In-Context Table Processing (Corallo et al., TRL 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/acl25-workshop-ingestion/2025.trl-1.13.pdf