AIModels.fyi

AIModels.fyi

Share this post

AIModels.fyi
AIModels.fyi
The bottleneck in LLMs is finding reasoning errors, not fixing them

The bottleneck in LLMs is finding reasoning errors, not fixing them

LLMs can't find reasoning errors, but can correct them when you tell them where to look

aimodels-fyi's avatar
aimodels-fyi
Jun 10, 2024
∙ Paid
4

Share this post

AIModels.fyi
AIModels.fyi
The bottleneck in LLMs is finding reasoning errors, not fixing them
Share
The BIG-bench-mistake UI from the repo (link inside!)

LLMs have taken the field of natural language processing by storm. With the right prompting, LLMs can solve all sorts of tasks in a zero- or few-shot way, demonstrating impressive capabilities. However, a key weakness of current LLMs seems to be self-correction - the ability to find and fix errors in their own outputs.

A new paper by researchers at Google and the University of Cambridge digs into this issue of LLM self-correction. The authors divide the self-correction process into two distinct components:

  1. Mistake finding, which refers to identifying errors in an LLM's output

  2. Output correction, which involves actually fixing those mistakes once they've been pinpointed.

Their analysis focuses specifically on reasoning tasks, where LLMs generate a multi-step chain-of-thought (CoT) style trace showing their step-by-step reasoning process. So what did they find?

AIModels.fyi is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.

Keep reading with a 7-day free trial

Subscribe to AIModels.fyi to keep reading this post and get 7 days of free access to the full post archives.

Already a paid subscriber? Sign in
© 2025 AIModels.fyi
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share