Abstract
The recent progress in large language models (LLMs), especially the invention of chain-of-thought prompting, has made it possible to automatically answer questions by stepwise reasoning. However, when faced with more complicated problems that require non-linear thinking, even the strongest LLMs make mistakes. To address this, we explore whether LLMs are able to recognize errors in their own step-by-step reasoning, without resorting to external resources. To this end, we propose SelfCheck, a general-purpose zero-shot verification schema for recognizing such errors. We then use the results of these checks to improve question-answering performance by conducting weighted voting on multiple solutions to the question. We test SelfCheck on math- and logic-based datasets and find that it successfully recognizes errors and, in turn, increases final answer accuracies. © 2024 12th International Conference on Learning Representations, ICLR 2024. All rights reserved.
Original language | English |
---|---|
Title of host publication | 12th International Conference on Learning Representations (ICLR 2024) |
Publisher | International Conference on Learning Representations, ICLR |
Number of pages | 16 |
ISBN (Print) | 9781713898658 |
Publication status | Published - Oct 2024 |
Externally published | Yes |
Event | 12th International Conference on Learning Representations (ICLR 2024) - Messe Wien Exhibition and Congress Center, Vienna, Austria Duration: 7 May 2024 → 11 May 2024 https://iclr.cc/Conferences/2024 https://openreview.net/group?id=ICLR.cc/2024/Conference |
Publication series
Name | International Conference on Learning Representations, ICLR |
---|
Conference
Conference | 12th International Conference on Learning Representations (ICLR 2024) |
---|---|
Country/Territory | Austria |
City | Vienna |
Period | 7/05/24 → 11/05/24 |
Internet address |