Back to Reading List
[Reasoning]·PAP-VLDZUF·2023·March 17, 2026

Let's Verify Step by Step

2023

Hunter Lightman, Vineet Kosaraju, Yura Burda et al.

4 min readReasoningAlignmentTraining

Core Insight

Process supervision beats outcome supervision in AI reasoning accuracy—think 78.2% vs 72.4% success in math tasks.

By the Numbers

78.2%

success rate of PRMs on MATH tasks

72.4%

success rate of ORMs on MATH tasks

800,000

human feedback labels used

In Plain English

The paper developed process reward models (PRMs) to assess each reasoning step, improving performance to 78.2% on MATH tasks. This surpasses outcome models' 72.4% success, illustrating that evaluating intermediate steps boosts AI accuracy.

Knowledge Prerequisites

git blame for knowledge

To fully understand Let's Verify Step by Step, trace this dependency chain first. Papers in our library are linked — click to read them.

DIRECT PREREQIN LIBRARY
Attention Is All You Need

Understanding the attention mechanism is fundamental to grasping the step-by-step verification approach used in modern language models.

attention mechanismtransformer architectureself-attention
DIRECT PREREQIN LIBRARY
Chain-of-Thought Prompting Elicits Reasoning in Large Language Models

This paper introduces the chain-of-thought prompting that underpins the structured, step-by-step reasoning process.

chain-of-thought promptingreasoning in modelsprompt engineering
DIRECT PREREQIN LIBRARY
OpenAI o1: Learning to Reason with LLMs

Building on foundational reasoning, this paper elaborates on enhancing reasoning capabilities in language models through structured learning.

learning to reasonlanguage model reasoningstructured learning
DIRECT PREREQIN LIBRARY
Self-Consistency Improves Chain of Thought Reasoning in Language Models

This demonstrates the self-consistency mechanism, which is crucial for improving the reliability of step-by-step reasoning processes.

self-consistencyreasoning improvementmodel reliability

YOU ARE HERE

Let's Verify Step by Step

The Idea Graph

The Idea Graph
10 nodes · 11 edges
Click a node to explore · Drag to pan · Scroll to zoom
622 words · 4 min read6 sections · 10 concepts

Table of Contents

01

The Problem: Limitations of Outcome Supervision

138 words

In traditional AI models, has been the norm. This approach evaluates only the final result of a task, which can often lead to oversight of errors made during the intermediate steps of reasoning. This is particularly problematic in tasks requiring complex, multi-step thinking, such as mathematics or logical reasoning, where mistakes early in the process can cascade into incorrect final outcomes.

(ORMs) are the standard implementation of this approach and have been shown to achieve a 72.4% success rate on certain reasoning tasks. However, by focusing solely on the end result, ORMs lack the granularity needed to pinpoint where the process went wrong.

The limitations of underscore the need for a new approach that can provide more detailed feedback throughout the reasoning process, which is where process supervision comes into play.

02

Key Insight: Process Supervision

96 words

The core insight of this research is the introduction of , which evaluates each step of the reasoning process rather than just the final outcome. This method allows for more granular feedback and accountability, helping to identify errors at the point they occur rather than only at the end of a task.

contrasts with traditional by providing a framework that can improve AI accuracy in tasks that require multi-step reasoning. This insight is central to the development of Process Reward Models (PRMs), which aim to improve upon the limitations of ORMs.

03

Methodology: Developing Process Reward Models

96 words

(PRMs) are an innovative approach that assess the correctness of each individual step in the reasoning process. This methodology involves using a novel dataset of 800,000 human feedback labels to score each step, providing a much-needed level of granularity.

The development of PRMs is a direct response to the limitations observed in . By focusing on the process rather than just the outcome, PRMs offer a new dimension of analysis for AI reasoning tasks. This method allows AI systems to better understand where and why errors occur, ultimately improving task performance.

04

Methodology: Utilizing Human Feedback

90 words

A significant component of the is the use of a novel . This dataset contains 800,000 labels that are used to evaluate the correctness of each reasoning step. By incorporating human judgment into the process, PRMs can more accurately assess intermediate steps.

The is crucial for training the PRMs, ensuring that the models have a robust understanding of the reasoning process. This dataset provides the necessary foundation for PRMs to function effectively, highlighting the importance of human-AI collaboration in improving AI performance.

05

Results: Improved Reasoning Accuracy

92 words

The implementation of Process Reward Models has yielded significant improvements in . In mathematical reasoning tasks, PRMs achieved a 78.2% success rate, compared to the 72.4% success rate of Outcome Reward Models. This substantial performance gap highlights the effectiveness of evaluating intermediate reasoning steps.

The results demonstrate that provides a more reliable path for training AI to correctly perform complex multi-step reasoning. By focusing on each step rather than just the outcome, PRMs offer a more detailed and accurate understanding of the reasoning process, leading to better overall performance.

06

Impact: Enhancing AI Reliability

110 words

The improvements in reasoning accuracy enabled by Process Reward Models have significant implications for various AI applications. By making AI systems more reliable, especially in tasks that require complex reasoning, PRMs can transform fields such as and .

For , the enhanced reasoning abilities of AI can lead to more accurate and reliable tutoring systems, providing better support for learners. In the realm of , improved reasoning can make chatbots more effective, delivering more accurate and contextually relevant responses to users.

Overall, the introduction of and PRMs stands to significantly enhance the trustworthiness and effectiveness of AI systems across a range of applications.

Experience It

Live Experiment

Process Reward Models

See Process Supervision in Action

Compare AI reasoning with and without evaluating each step. See how process supervision enhances accuracy in solving complex problems.

Notice how the Process Reward Models lead to more accurate and reliable solutions by verifying each step, as demonstrated by the 78.2% success rate in the paper.

Try an example — see the difference instantly

⌘↵ to run

How grounded is this content?

Metrics are computed from available source text only — abstract, summary, and impact fields ingested into this system. Full paper PDF is not ingested; numerical claims that originate from within the paper body will not appear in these scores.

Source Richness100%

8 of 8 content fields populated. More fields = better-grounded generation.

Source Depth~225 words

Total source text analyzed by the model. Includes extended deep-dive summary — high confidence.

Number Grounding3 / 3

Key statistics whose numeric values appear verbatim in ingested source text. Unverified stats may originate from the full paper body.

Quote Traceability3 / 3

Key passages whose significant vocabulary (≥4-char words) overlap ≥35% with source text. Measures lexical traceability, not semantic accuracy.

Methodology: Number grounding uses regex digit extraction against source text. Quote traceability uses token set intersection on content words stripped of stop-words. Neither metric validates semantic correctness or factual accuracy against the original paper. For full verification, cross-reference with the original paper via the arXiv link above.