@inproceedings{veerendranath-etal-2024-calc,
title = "Calc-{CMU} at {S}em{E}val-2024 Task 7: Pre-Calc - Learning to Use the Calculator Improves Numeracy in Language Models",
author = "Veerendranath, Vishruth and
Shah, Vishwa and
Ghate, Kshitish",
editor = {Ojha, Atul Kr. and
Do{\u{g}}ru{\"o}z, A. Seza and
Tayyar Madabushi, Harish and
Da San Martino, Giovanni and
Rosenthal, Sara and
Ros{\'a}, Aiala},
booktitle = "Proceedings of the 18th International Workshop on Semantic Evaluation (SemEval-2024)",
month = jun,
year = "2024",
address = "Mexico City, Mexico",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2024.semeval-1.211/",
doi = "10.18653/v1/2024.semeval-1.211",
pages = "1468--1475",
abstract = "Quantitative and numerical comprehension in language is an important task in many fields like education and finance, but still remains a challenging task for language models. While tool and calculator usage has shown to be helpful to improve mathematical reasoning in large pretrained decoder-only language models, this remains unexplored for smaller language models with encoders. In this paper, we propose Pre-Calc, a simple pre-finetuning objective of learning to use the calculator for both encoder-only and encoder-decoder architectures, formulated as a discriminative and generative task respectively. We pre-train BERT and RoBERTa for discriminative calculator use and Flan-T5 for generative calculator use on the MAWPS, SVAMP, and AsDiv-A datasets, which improves performance on downstream tasks that require numerical understanding. Our code and data are available at https://github.com/calc-cmu/pre-calc."
}
Markdown (Informal)
[Calc-CMU at SemEval-2024 Task 7: Pre-Calc - Learning to Use the Calculator Improves Numeracy in Language Models](https://preview.aclanthology.org/add-emnlp-2024-awards/2024.semeval-1.211/) (Veerendranath et al., SemEval 2024)
ACL