Tony Mak
2024
LLMs cannot find reasoning errors, but can correct them given the error location
Gladys Tyen
|
Hassan Mansoor
|
Victor Carbune
|
Peter Chen
|
Tony Mak
Findings of the Association for Computational Linguistics ACL 2024
While self-correction has shown promise in improving LLM outputs in terms of style and quality (e.g. Chen et al., 2023b; Madaan et al.,2023), recent attempts to self-correct logical or reasoning errors often cause correct answers to become incorrect, resulting in worse performances overall (Huang et al., 2023). In this paper, we show that poor self-correction performance stems from LLMs’ inability tofind logical mistakes, rather than their ability to correct a known mistake. Firstly, we benchmark several state-of-the-art LLMs ontheir mistake-finding ability and demonstrate that they generally struggle with the task, even in highly objective, unambiguous cases. Secondly, we test the correction abilities of LLMs – separately from mistake finding – using a backtracking setup that feeds ground truth mistake location information to the model. We show that this boosts downstream task performance across our 5 reasoning tasks, indicating that LLMs’ correction abilities are robust. Finally, we show that it is possible to obtain mistake location information without ground truth labels or in-domain training data. We train a small classifier with out-of-domain data, which exhibits stronger mistake-finding performance than prompting a large model. We release our dataset of LLM-generated logical mistakes, BIG-Bench Mistake, to enable further research into locating LLM reasoning mistakes.
2021
Predicting Text Readability from Scrolling Interactions
Sian Gooding
|
Yevgeni Berzak
|
Tony Mak
|
Matt Sharifi
Proceedings of the 25th Conference on Computational Natural Language Learning
Judging the readability of text has many important applications, for instance when performing text simplification or when sourcing reading material for language learners. In this paper, we present a 518 participant study which investigates how scrolling behaviour relates to the readability of English texts. We make our dataset publicly available and show that (1) there are statistically significant differences in the way readers interact with text depending on the text level, (2) such measures can be used to predict the readability of text, and (3) the background of a reader impacts their reading interactions and the factors contributing to text difficulty.
Search
Co-authors
- Sian Gooding 1
- Yevgeni Berzak 1
- Matt Sharifi 1
- Gladys Tyen 1
- Hassan Mansoor 1
- show all...