## GenPRM: Scaling Test-Time Compute of Process Reward Models via Generative Reasoning
Jian Zhao 1,3 * , Runze Liu 1,2 *โ , Kaiyan Zhang 1 , Zhimu Zhou 3 , Junqi Gao 4 , Dong Li 4 , Jiafei Lyu 1 , Zhouyi Qian 4 , Biqing Qi 2 โก , Xiu Li 1 โก and Bowen Zhou 1,2 โก
1 Tsinghua University, 2 Shanghai AI Laboratory, 3 BUPT, 4 Harbin Institute of Technology
Recent advancements in Large Language Models (LLMs) have shown that it is promising to utilize Process Reward Models (PRMs) as verifiers to enhance the performance of LLMs. However, current PRMs face three key challenges: (1) limited process supervision and generalization capabilities, (2) dependence on scalar value prediction without leveraging the generative abilities of LLMs, and (3) inability to scale the test-time compute of PRMs. In this work, we introduce GenPRM, a generative process reward model that performs explicit Chain-of-Thought (CoT) reasoning with code verification before providing judgment for each reasoning step. To obtain high-quality process supervision labels and rationale data, we propose Relative Progress Estimation (RPE) and a rationale synthesis framework that incorporates code verification. Experimental results on ProcessBench and several mathematical reasoning tasks show that GenPRM significantly outperforms prior PRMs with only 23K training data from MATH dataset. Through test-time scaling, a 1.5B GenPRM outperforms GPT-4o , and a 7B GenPRM surpasses Qwen2.5-Math-PRM-72B on ProcessBench. Additionally, GenPRM demonstrates strong abilities to serve as a critic model for policy model refinement. This work establishes a new paradigm for process supervision that bridges the gap between PRMs and critic models in LLMs. Our code, model, and data are available in https://ryanliu112.github.io/GenPRM .
Figure 1: GenPRM achieves state-of-the-art performance across multiple benchmarks in two key roles: (a) As a verifier : GenPRM-7B outperforms all classification-based PRMs of comparable size and even surpasses Qwen2.5-Math-PRM-72B via test-time scaling. (b) As a critic : GenPRM-7B demonstrates superior critique capabilities, achieving 3.4ร greater performance gains than DeepSeekR1-Distill-Qwen-7B after 3 refinement iterations.
<details>
<summary>Image 1 Details</summary>

### Visual Description
\n
## Bar Chart & Line Graph: GenPRM as Verifier & Critic
### Overview
The image presents two charts: a bar chart comparing the "Best-of-32 Accuracy" and "ProcessBench" scores for various models when "GenPRM" is used as a verifier, and a line graph showing the "ProcessBench F1 Score" as a function of "# Refinement Turn" for different models when "GenPRM" is used as a critic.
### Components/Axes
**Chart (a): GenPRM as a Verifier (Best-of-N & ProcessBench)**
* **X-axis:** Model names: Skywork-PRM-1.5B, Skywork-PRM-7B, Owen2.5-Math-7B-PRMBOOK, Owen2.5-Math-PRM-7B, Direct GenPRM-7B, GenPRM-7B [Pass@1], GenPRM-7B [Maj@8].
* **Y-axis:** Best-of-32 Accuracy (%) - Scale from 45 to 69.
* **Legend:**
* Best-of-32 (Teal/Green)
* ProcessBench (Coral/Orange)
* **Annotations:**
* Pass@32 (67.6) - positioned above the first bar.
* GPT-4.0 (61.9) - positioned above the third bar.
* Maj@32 (54.1) - positioned above the fourth bar.
**Chart (b): GenPRM as a Critic**
* **X-axis:** # Refinement Turn - Scale from 0 to 3.
* **Y-axis:** ProcessBench F1 Score (%) - Scale from 30 to 90.
* **Legend:**
* GenPRM-7B (Red)
* DeepSeek-R1-Distill-7B (Green)
* Self-Refine (Blue)
* **Annotations:**
* Pass@1 - positioned near the x-axis.
* 3.4x - positioned near the end of the graph.
### Detailed Analysis or Content Details
**Chart (a): GenPRM as a Verifier**
* **Skywork-PRM-1.5B:** Best-of-32 Accuracy โ 36.4%, ProcessBench โ 48.8%
* **Skywork-PRM-7B:** Best-of-32 Accuracy โ 52.5%, ProcessBench โ 54.1%
* **Owen2.5-Math-7B-PRMBOOK:** Best-of-32 Accuracy โ 53.1%, ProcessBench โ 42.1%
* **Owen2.5-Math-PRM-7B:** Best-of-32 Accuracy โ 56.5%, ProcessBench โ 53.8%
* **Direct GenPRM-7B:** Best-of-32 Accuracy โ 73.5%, ProcessBench โ 56.2%
* **GenPRM-7B [Pass@1]:** Best-of-32 Accuracy โ 78.3%, ProcessBench โ 52.2%
* **GenPRM-7B [Maj@8]:** Best-of-32 Accuracy โ 80.5%, ProcessBench โ 59.7%
The Best-of-32 accuracy generally increases with the model size and complexity, peaking at GenPRM-7B [Maj@8]. ProcessBench scores are more variable and do not show a consistent trend.
**Chart (b): GenPRM as a Critic**
* **GenPRM-7B:**
* Refinement Turn 0: โ 48.2%
* Refinement Turn 1: โ 51.1%
* Refinement Turn 2: โ 51.8%
* Refinement Turn 3: โ 52.1%
* **DeepSeek-R1-Distill-7B:**
* Refinement Turn 0: โ 47.5%
* Refinement Turn 1: โ 49.2%
* Refinement Turn 2: โ 50.5%
* Refinement Turn 3: โ 51.1%
* **Self-Refine:**
* Refinement Turn 0: โ 46.5%
* Refinement Turn 1: โ 48.8%
* Refinement Turn 2: โ 50.1%
* Refinement Turn 3: โ 50.8%
All three models show an increasing trend in ProcessBench F1 Score with increasing refinement turns, but the improvement plateaus after Refinement Turn 2. GenPRM-7B consistently achieves the highest F1 score across all refinement turns.
### Key Observations
* GenPRM-7B [Maj@8] achieves the highest Best-of-32 accuracy.
* The ProcessBench scores are generally lower than the Best-of-32 accuracy scores.
* GenPRM-7B consistently outperforms DeepSeek-R1-Distill-7B and Self-Refine as a critic, with a 3.4x improvement.
* The improvement in ProcessBench F1 Score diminishes with each refinement turn.
### Interpretation
The data suggests that GenPRM is an effective verifier, particularly when combined with majority voting ([Maj@8]). The increasing Best-of-32 accuracy with more complex models indicates that GenPRM can effectively identify and validate higher-quality outputs.
As a critic, GenPRM demonstrates a positive impact on the ProcessBench F1 Score through iterative refinement. However, the diminishing returns suggest that there is a limit to the benefits of continued refinement. The consistent outperformance of GenPRM-7B over other models highlights its superior ability to guide the refinement process.
The discrepancy between Best-of-32 accuracy and ProcessBench scores could indicate that the two metrics evaluate different aspects of model performance. Best-of-32 accuracy may focus on overall correctness, while ProcessBench F1 Score may be more sensitive to the quality of reasoning or the ability to follow specific instructions. The annotation "Pass@1" and "Pass@32" suggest that the Best-of-32 metric is based on selecting the best output from multiple generations.
</details>
* Equal contribution
โก Corresponding authors: Biqing Qi (qibiqing@pjlab.org.cn), Xiu Li (li.xiu@sz.tsinghua.edu.cn), and Bowen Zhou (zhoubowen@tsinghua.edu.cn)
โ Project lead & Work done during an internship at Shanghai AI Laboratory
Figure 2: Comparison between GenPRM (right) and previous classification-based PRMs (left).
<details>
<summary>Image 2 Details</summary>

### Visual Description
\n
## Diagram: Program Reasoning Model Comparison
### Overview
The image presents a diagram comparing two approaches to Program Reasoning Models (PRM): a Classification-based PRM and a Generative PRM (labeled as "Ours"). The diagram illustrates the flow of information and the associated confidence scores for each approach. The Generative PRM appears to involve a multi-step verification process.
### Components/Axes
The diagram is divided into two main sections separated by a dashed vertical line.
* **Left Side:** Classification-based PRM, containing two sub-components: "Discriminative PRM" and "Direct Generative PRM".
* **Right Side:** Generative PRM (labeled "GenPRM"), which includes a central "GenPRM" block and a series of interconnected verification steps.
* **Legend:** Located at the bottom-right, the legend defines the shapes and colors used to represent verification steps:
* Light Green Circle: "CoT Verification" (Chain of Thought)
* Light Blue Circle: "Code Verification"
* **Confidence Scores:** Orange rectangles display confidence scores associated with each process.
* **Yes/No Indicators:** Blue rectangles indicate the outcome of verification steps ("Yes" or "No").
* **Red X:** Indicates a failure in the Generative PRM process.
* **Green Checkmark:** Indicates a success in the Classification-based PRM process.
### Detailed Analysis or Content Details
**Classification-based PRM:**
* **Discriminative PRM:** Flows to a "value head" which outputs a confidence score of approximately 0.9. This is indicated by an orange rectangle with the value "0.9" and a green checkmark.
* **Direct Generative PRM:** Flows to a "Yes" indicator, which then flows to an orange rectangle with a confidence score of approximately 0.8. This is also indicated by a green checkmark.
**Generative PRM (Ours):**
* The "GenPRM" block (light blue rectangle) has three outgoing arrows.
* **Path 1:** Flows through three light green "CoT Verification" circles, then to a "Yes" indicator.
* **Path 2:** Flows through three light green "CoT Verification" circles, then to a "No" indicator.
* **Path 3:** Flows through three light blue "Code Verification" circles, then to a "No" indicator.
* The "No" indicators from Paths 2 and 3 converge to a final output with a confidence score of approximately 0.2, indicated by an orange rectangle with the value "0.2" and a red X.
### Key Observations
* The Classification-based PRM consistently achieves higher confidence scores (0.8-0.9) and is always successful (indicated by green checkmarks).
* The Generative PRM has a more complex flow with multiple verification steps.
* The Generative PRM ultimately results in a low confidence score (0.2) and a failure (indicated by the red X), suggesting that the verification process identified issues.
* The Generative PRM's failure is triggered by "No" outcomes from both CoT and Code Verification paths.
### Interpretation
The diagram demonstrates a comparison between two PRM approaches. The Classification-based PRM appears to be more reliable, consistently producing high confidence scores and successful outcomes. The Generative PRM, while potentially more sophisticated due to its multi-step verification process, is shown to be less reliable in this instance, resulting in a low confidence score and a failure. The diagram suggests that the Generative PRM's verification steps are effectively identifying flaws in its reasoning or code generation. The confidence scores provide a quantitative measure of the reliability of each approach. The diagram highlights the challenges of building robust generative models for program reasoning and the importance of effective verification mechanisms. The "Ours" label suggests that the Generative PRM is a novel approach being proposed by the authors of the document. The diagram is a visual representation of an experimental result, showcasing the performance of the two PRM approaches.
</details>
## 1. Introduction
Large Language Models (LLMs) have shown significant advances in recent years (OpenAI, 2023; Anthropic, 2023; OpenAI, 2024a,b; DeepSeek-AI et al., 2025). As OpenAI o1 demonstrates the great effectiveness of scaling test-time compute (OpenAI, 2024a), an increasing number of researches focus on Test-Time Scaling (TTS) methods to improve the reasoning performance of LLMs (Snell et al., 2025; Liu et al., 2025).
Effective TTS requires high-quality verifiers, such as Process Reward Models (PRMs) (Liu et al., 2025). However, existing PRMs face several limitations. They exhibit limited process supervision capabilities and struggle to generalize across different models and tasks (Zheng et al., 2024; Zhang et al., 2025c; Liu et al., 2025). Furthermore, most current approaches train PRMs as classifiers that output scalar values, neglecting the natural language generation abilities of LLMs, which are pre-trained on extensive corpora. This classifier-based modeling inherently prevents PRMs from leveraging test-time scaling methods to enhance process supervision capabilities. These limitations lead us to the following research question: How can generative modeling enhance the process supervision capabilities of PRMs while enabling test-time scaling?
In this work, we address these challenges through a generative process reward model, named GenPRM. Specifically, GenPRM differs from classification-based PRMs in that GenPRM redefines process supervision as a generative task rather than a discriminative scoring task by integrating Chain-of-Thought (CoT) (Wei et al., 2022) reasoning and code verification processes before providing final judgment. To improve conventional hard label estimation, we propose Relative Progress Estimation (RPE), which leverages a relative criterion for label estimation. Additionally, we introduce a rationale synthesis framework with code verification to obtain high-quality process supervision reasoning data. A comparison of our method with previous classification-based methods is presented in Figure 2.
Our contributions can be summarized as follows:
1. We propose a generative process reward model that performs explicit CoT reasoning with code verification and utilizes Relative Progress Estimation to obtain accurate PRM labels.
3. We provide a new perspective on PRMs in this work, fully leveraging their TTS capabilities, reshaping their applications, and opening new directions for future research in process supervision.
2. Empirical results on ProcessBench and common mathematical reasoning tasks demonstrate that GenPRM outperforms prior classification-based PRMs. Additionally, smaller GenPRM models can surpass larger PRMs via TTS.
## 2. Preliminaries
## 2.1. Markov Decision Process
Following Liu et al. (2025), we formulate the test-time scaling process with PRMs as a Markov Decision Process (MDP) defined by ( ๐ฎ , ๐ , ๐, ๐, ๐พ ) , where ๐ฎ is the state space, ๐ is the action space, ๐ represents transition dynamics, ๐ : ๐ฎ ร ๐ โ R is the reward function, and ๐พ โ [0 , 1] is the discount factor. Starting with a prompt set ๐ณ and an initial state ๐ 1 = ๐ฅ โผ ๐ณ , the policy model ๐ ๐ generates an action ๐ 1 โผ ๐ ๐ ( ยท | ๐ 1 ) . 1 Unlike traditional RL methods with stochastic transitions (Liu et al., 2022, 2024), transitions in LLMs are deterministic, i.e., ๐ ๐ก +1 = ๐ ( ยท | ๐ ๐ก , ๐ ๐ก ) = [ ๐ ๐ก , ๐ ๐ก ] , where [ ยท , ยท ] denotes string concatenation. This process continues until the episode terminates (i.e., generating the [EOS] token), obtaining a trajectory of ๐ steps: ๐ = { ๐ 1 , ๐ 2 , ยท ยท ยท , ๐ ๐ } . The goal is to optimize either the reward of each step (as in search-based methods) or the reward over the full response (as in Best-of-N sampling).
## 2.2. Supervised Fine-Tuning
Supervised Fine-Tuning (SFT) trains a model to predict the next token based on prior context. For a dataset ๐ SFT = { ( ๐ฅ ( ๐ ) , ๐ฆ ( ๐ ) ) } ๐ ๐ =1 , the SFT loss is:
<!-- formula-not-decoded -->
where ๐ ๐ represents a model with parameters ๐ .
## 2.3. Test-Time Scaling
In this work, we consider two test-time scaling methods, including majority voting and Best-of-N.
Majority Voting. Majority voting (Wang et al., 2023) selects the answer that appears the most frequently among all solutions.
Best-of-N. Best-of-N (BoN) (Brown et al., 2024; Snell et al., 2025) selects the best answer from ๐ candidate solutions.
## 3. Method
In this section, we first describe how to develop GenPRM and integrate the reasoning process with code verification. We then introduce how to scale test-time compute of policy models using GenPRM and apply TTS for GenPRM. Last, we present the improved label estimation method and data generation and filtering framework of GenPRM.
## 3.1. GenPRM and Test-Time Scaling
## 3.1.1. From Discriminative PRM to Generative PRM
Discriminative PRM. Assume we have a PRM dataset ๐ Disc = { ( ๐ ๐ก , ๐ ๐ก ) , ๐ ๐ก } , where ๐ ๐ก โ { 0 , 1 } for PRMlabels with hard estimation. The discriminative PRM ๐ ๐ is trained via cross-entropy loss (Skywork
1 Following Snell et al. (2025); Liu et al. (2025), we refer to models that generate solutions as policy models.
Figure 3: Overall framework of GenPRM. Our framework consists of six key parts: 1 /bigcircle The policy model generates solution steps, with MC scores estimated from rollout trajectories. 2 /bigcircle Our proposed RPE derives accurate PRM labels. 3 /bigcircle High-quality process supervision data is synthesized through CoT reasoning augmented with code verification. 4 /bigcircle We apply consensus filtering followed by SFT to train GenPRM. 5 /bigcircle The trained GenPRM functions as a verifier or critic, enabling enhanced test-time scaling for policy models. 6 /bigcircle The performance of GenPRM further improves through test-time scaling.
<details>
<summary>Image 3 Details</summary>

### Visual Description
\n
## Diagram: GenPRM System Architecture
### Overview
This diagram illustrates the architecture of the GenPRM (Generative Policy Rollout Model) system, a framework for solving math problems. The system consists of six main stages: Solution Generation & MC Reward Estimation, Relative Progress Estimation, Rationale Synthesis, GenPRM Training, Policy Model Test-Time Scaling, and GenPRM Test-Time Scaling. The diagram depicts the flow of information and processes within the system, highlighting the use of intermediate steps, correct/incorrect answers, reward judgements, and the iterative nature of the GenPRM model.
### Components/Axes
The diagram is segmented into six numbered boxes, each representing a stage in the GenPRM process. A legend at the bottom identifies key symbols:
* **a**: Intermediate Step (light blue)
* **q**: Correct Answer (green)
* **q**: Incorrect Answer (red)
* **r**: Single Reward Judgement (yellow)
* **r**: Aggregated Reward Judgement (dark green)
The diagram uses arrows to indicate the flow of information between stages. Text boxes within each stage describe the processes involved.
### Detailed Analysis or Content Details
**1. Solution Generation & MC Reward Estimation:**
* A "Math Problem" is presented.
* The problem is broken down into a series of steps represented by `a11`, `a21`, `a31`, `a12`, `a22`, `a32`, and `a13`, `a23`, `a33`. These are labeled as "Intermediate Step" (light blue).
* The steps lead to answers `q1`, `q2`, `q3`. These are labeled as "Correct Answer" (green) or "Incorrect Answer" (red).
* An equation `MC(s, a) = P = 1` and `MC(s) = z โ` is shown, representing Monte Carlo (MC) reward estimation.
* A specific example is given: `MC(s12, a12) = 2/3`.
**2. Relative Progress Estimation:**
* The output of the previous stage is fed into this stage.
* A visual comparison is made between `f = 1` (Correct Answer - green) and `f = 0` (Incorrect Answer - red).
**3. Rationale Synthesis:**
* The "Problem" from stage 1 is input.
* A "CoT Analysis" (Chain-of-Thought Analysis) is performed, enclosed in `<analyze>` tags. The text within reads: "Let's analyze the paragraph step by step..." and closes with `</analyze>`.
* This analysis leads to a "Final Label" of "Yes/No".
* "Code Verification" is performed, enclosed in `<verify>` tags. The text within reads: "Let's use python code to find any potential error..." and closes with `</verify>`.
* The code is "Executed", resulting in "[Code Output]...".
* A comparison is made between the analysis and code execution, resulting in either "consistent" (green upward arrow) or "conflict" (red).
**4. GenPRM Training:**
* "Data (23K)" is input.
* This data is used to train "GenPRM" using "SFT" (Supervised Fine-Tuning).
* "Consensus filtering" is applied to the output of GenPRM.
**5. Policy Model Test-Time Scaling:**
* "GenPRM as a Verifier" is used to verify solutions.
* "GenPRM as a Critic" provides critiques.
* The process is repeated "x N" times, with solutions being generated, critiqued, and refined iteratively.
* A small icon of a brain with a star is present.
**6. GenPRM Test-Time Scaling:**
* "GenPRM" is used to "analyze" and "verify" solutions.
* The process is repeated "x N" times.
* The output is represented as `r1`, `r2`, `r3`, `r4`, which are "Single Reward Judgements" (yellow).
* These individual judgements are aggregated into a final "Aggregated Reward Judgement" (dark green).
### Key Observations
* The system is highly iterative, with multiple stages of verification and refinement.
* The use of both Chain-of-Thought analysis and code verification suggests a robust approach to problem-solving.
* The "x N" notation indicates that certain processes are scaled up for increased performance or reliability.
* The color-coding of answers (green for correct, red for incorrect) provides a clear visual indication of success or failure.
### Interpretation
The diagram depicts a sophisticated system for automated math problem solving. The GenPRM framework leverages a combination of generative modeling, reinforcement learning, and verification techniques to achieve high accuracy and reliability. The iterative nature of the system allows it to learn from its mistakes and improve its performance over time. The inclusion of both symbolic reasoning (CoT analysis) and computational verification (code execution) suggests a hybrid approach that combines the strengths of both methods. The scaling factor "x N" implies that the system is designed to handle complex problems efficiently. The diagram highlights the importance of reward signals (r) in guiding the learning process and ensuring that the system converges to optimal solutions. The use of consensus filtering suggests a mechanism for mitigating the effects of noisy or unreliable data. Overall, the diagram presents a compelling vision for the future of automated problem-solving.
</details>
o1 Team, 2024; Zhang et al., 2025c):
<!-- formula-not-decoded -->
Direct Generative PRM. With a dataset ๐ Direct-Gen = { ( ๐ ๐ก , ๐ ๐ก ) , ๐ ๐ก } , where ๐ ๐ก is Yes for a correct step and No otherwise, the direct generative PRM (Xiong et al., 2024) is trained through SFT to predict Yes or No for each step. For step ๐ก , we use the probability of the Yes token as the predicted process reward ห ๐ ๐ก :
<!-- formula-not-decoded -->
Generative PRM. By equipping the direct generative PRM with an explicit reasoning process like CoT (Wei et al., 2022), we obtain a generative PRM. Let ๐ฃ 1: ๐ก -1 denote the rationale from step 1 to ๐ก -1 and ๐ฃ ๐ก denote the rationale for step ๐ก . Assume we have a dataset ๐ Gen = { ( ๐ ๐ก , ๐ ๐ก , ๐ฃ 1: ๐ก -1 ) , ( ๐ฃ ๐ก , ๐ ๐ก ) } . GenPRM learns to reason and verify each step via SFT on this dataset. The generative process reward ห ๐ ๐ก can be obtained via the following equation:
<!-- formula-not-decoded -->
Generative PRM with Code Verification. If we only verify the reasoning step with CoT based on natural language, the process may lack robustness in certain complex scenarios (Zhu et al., 2024; Gou et al., 2024). The difference between the generative PRM and the generative PRM with code verification is that the latter generates code to verify the reasoning step by executing it and provides the judgment based on the execution results. At step ๐ก , after generating the rationale ๐ฃ ๐ก containing CoT and code, we execute the code and obtain feedback ๐ ๐ก . Given the current state ๐ ๐ก , action ๐ ๐ก , previous rationales ๐ฃ 1: ๐ก -1 , and previous corresponding execution feedback ๐ 1: ๐ก -1 , the PRM first generates the
rationale ๐ฃ ๐ก . After execution and obtaining the feedback ๐ ๐ก , we compute the final generative process reward as follows:
<!-- formula-not-decoded -->
In the following sections, we refer to GenPRM as this generative PRM type with code verification. The effectiveness of CoT and code verification can be found in Section 4.4.
## 3.1.2. Test-Time Scaling
Policy Model TTS: GenPRM as a Verifier. To scale the test-time compute of policy models, we can sampling multiple responses from policy models and then use GenPRM as a verifier to select the final answer (Snell et al., 2025) in the way of parallel TTS.
Policy Model TTS: GenPRM as a Critic. By equipping the PRM with generative process supervision abilities, GenPRM can be naturally used as a critic model to refine the outputs of policy models and we can scale the refinement process with multiple turns in a sequential TTS manner.
GenPRM TTS. When evaluating each solution step, we first sample ๐ reasoning verification paths and then use majority voting to obtain the final prediction by averaging the rewards. For GenPRM without code verification, the rewards are computed as follows:
<!-- formula-not-decoded -->
And we can further incorporate code verification and execution feedback into this reasoning process:
<!-- formula-not-decoded -->
Then the rewards can be used for ranking the responses of policy models or be converted into binary labels through a threshold 0 . 5 for judging the correctness of the step. The discussion of using code verification can be found at Table 5.
## 3.2. Synthesizing Data of GenPRM
In this section, we introduce our pipeline for synthesizing training data of GenPRM. The pipeline consists of three stages: (1) generating reasoning paths and obtaining PRM labels via Monte Carlo (MC) estimation; (2) evaluating the progress of each step via Relative Progress Estimation; and (3) synthesizing rationales with CoT and code verification, and inferring LLM-as-a-judge labels with consensus filtering.
## 3.2.1. Solution Generation and Monte Carlo Estimation
Solution Generation with Step Forcing. We utilize the 7.5K problems from the training set of the MATH dataset (Hendrycks et al., 2021) as the problem set. For each problem, we use Qwen2.5-7BInstruct (Yang et al., 2024a) as the generation model to collect multiple solutions. Since using ' \n\n ' for step division does not consider the semantics of each step and may result in overly fine-grained division, we apply a step forcing approach to generate solutions. Specifically, we add ' Step 1: ' as the prefix for the generation model to complete the response. For a response with ๐ reasoning steps, the format is as follows:
## The response format with step forcing
Step 1: {step content}
...
Step T:
{step content}
The proportion of correct paths versus incorrect paths varies significantly depending on the difficulty of the problems. To ensure a sufficient number of correct and incorrect paths, we sample up to 2048 paths for both hard and easy problems. If no correct or incorrect paths are found after sampling 2048 responses, we discard the corresponding problems.
Balancing the Precision and Efficiency of MC Estimation. Following Math-Shepherd (Wang et al., 2024b), we estimate the probability of correctness for each step using completion-based sampling. For each reasoning step ๐ ๐ก , we generate ๐พ completion trajectories using a completion model, specifically Qwen2.5-Math-7B-Instruct (Yang et al., 2024b), and use MC estimation to calculate the probability that the current step ๐ ๐ก is correct (Wang et al., 2024b; Zhang et al., 2025c):
<!-- formula-not-decoded -->
<!-- formula-not-decoded -->
where ๐ ๐ is the answer of the ๐ -th response, ๐ * is the ground-truth answer, and โถ is the indicator function. However, it is difficult for the completion model to reach the correct answer for hard problems even when the original step is correct, leading to incorrect results for MC estimation. To address this and balance the computation cost, we use a dynamic ๐พ based on the estimated Pass@1 ๐๐ถ ( ๐ 1 ) :
## 3.2.2. Relative Progress Estimation
Previous work has shown that hard label estimation is better than soft label estimation for PRMs (Zhang et al., 2025c). However, after MC estimation, we observe that although the MC score of many steps is greater than 0, the steps are incorrect, as also noted by Zhang et al. (2025c). On the other hand, we assume that a positive step should be both correct and beneficial. A reasoning step is considered as a beneficial one if it is easier to reach the correct answer by adding this step as the generation prefix. To address these issues, we propose Relative Progress Estimation (RPE), which shares a similar idea with relative advantage estimation in GRPO (Shao et al., 2024; DeepSeek-AI et al., 2025), to improve conventional hard label estimation.
Specifically , the MC score is an empirical estimation of the current state ๐ ๐ก . To evaluate the quality of the current action ๐ ๐ก , it is natural to compare the MC score of the next state ๐ ๐ก +1 with that of the current state ๐ ๐ก , since ๐ ๐ก +1 = [ ๐ ๐ก , ๐ ๐ก ] . For each response, if the first erroneous step is step ๐ก โฒ (i.e., ๐๐ถ ( ๐ ๐ก โฒ ) = 0 ), we set the MC score of the following steps to 0. Our RPE ๐ ๐ก for step ๐ก is defined as follows:
where ๐๐ถ ( ๐ 1 ) is the estimated Pass@1 computed in the solution generation phase. However, we empirically find that using a strict criterion where progress is always greater than 1 leads to unsatisfactory performance, as shown in Table 3. To address this, we estimate the final reward label ห ๐ ๐ก by
<!-- formula-not-decoded -->
introducing a threshold ๐ :
<!-- formula-not-decoded -->
We also discuss another form of relative progress ๐ ๐ก = ๐๐ถ ( ๐ ๐ก , ๐ ๐ก ) -๐๐ถ ( ๐ ๐ก ) in Table 3 in Section 4.4.
## 3.2.3. Rationale Generation, Verification and Filtering
To obtain high-quality rationale data, we use QwQ-32B (Qwen Team, 2025) as the rationale generation model and introduce a three-step pipeline that automatically generates and verifies the rationale of each reasoning step. Given a problem ๐ฅ with a ground-truth answer ๐ * and candidate steps { ๐ 1 , ยท ยท ยท , ๐ ๐ } , the generation and verification proceed as follows:
Step 1: Code-Based Rationale Generation. To evaluate the correctness of ๐ ๐ก , we synthesize step-by-step CoT analysis. It has been shown that program-based reasoning improves verification outcomes (Zhu et al., 2024). Based on CoT analysis, we continue to synthesize code-based rationales to verify ๐ ๐ก based on the problem and historical steps { ๐ 1 , ยท ยท ยท , ๐ ๐ก -1 } . We prompt the rationale generation model to surround the CoT with <analyze> and </analyze> , and the code with <verify> and </verify> . The prompt for rationale generation is shown in Table A.2.
Step 2: Code Execution and Verification. With the generated code, we execute it and obtain the feedback ๐ ๐ก for step ๐ก . The execution feedback is formatted as [Code output: {execution result}] and is concatenated to the generated CoT and code as the prefix for the subsequent generation. If the execution result is inconsistent with the generated CoT verification, we observe that QwQ-32B performs self-reflection behaviors until reaching a consensus.
Step 3: Label Judgment and Consensus Filtering. After generating and verifying the rationale data of all candidate steps, the rationale generation model finally outputs an number. If all steps are inferred to be correct, the number will be -1, otherwise will be the index of the first erroneous step. For each solution, if there is at least one process label with RPE is not consistent with the labels generated by LLM-as-a-judge (Zheng et al., 2023), we discard the entire solution and only retain the one with all labels consistent. After consensus filtering, we discard approximately 51% of the data and finally obtain a dataset containing 23K problems with reasoning steps and rationale data.
## 4. Experiments
In this section, we aim to answer the following questions:
- Q1: How does GenPRM perform compared with previous PRMs? (ยง4.2, ยง4.3)
- Q2: How does the performance of GenPRM scale with more test-time compute? (ยง4.2, ยง4.3)
- Q3: How does GenPRM benefit policy model test-time scaling? (ยง4.3)
- Q4: How do the components and hyperparameters influence GenPRM? (ยง4.4)
## 4.1. Setup
Benchmarks. We evaluate GenPRM and baseline methods on ProcessBench (Zheng et al., 2024), a benchmark designed to assess process supervision capabilities in mathematical reasoning tasks. 2
2 Our evaluation code is adapted from https://github.com/QwenLM/ProcessBench .
Additionally, we conduct BoN and critic refinement experiments using MATH (Hendrycks et al., 2021), AMC23 (AI-MO, 2024b), AIME24 (AI-MO, 2024a), and Minerva Math (Lewkowycz et al., 2022). For BoN response generation, we employ Qwen2.5-Math-7B-Instruct (Yang et al., 2024b) and Gemma-3-12b-it (Gemma Team and Google DeepMind, 2025) as policy models. For policy model TTS with GenPRM as the critic, we use Gemma-3-12b-it (Gemma Team and Google DeepMind, 2025) and Qwen2.5-7B-Instruct (Yang et al., 2024a) as generators.
Baselines. For ProcessBench and BoN experiments, we compare GenPRM with the following methods:
- Math-Shepherd-PRM-7B (Wang et al., 2024b): This method trains a PRM using hard labels computed based on MC estimation.
- Skywork-PRM series (Skywork o1 Team, 2024): Comprises Skywork-PRM-1.5B and SkyworkPRM-7B.
- RLHFlow series (Xiong et al., 2024): Includes RLHFlow-PRM-Mistral-8B and RLHFlow-PRMDeepseek-8B.
- EurusPRM (Cui et al., 2025): EurusPRM-Stage1 and EurusPRM-Stage2 are trained as implicit PRMs (Yuan et al., 2024).
- RetrievalPRM-7B (Zhu et al., 2025): The method enhances PRM with retrieved questions and corresponding steps.
- Qwen2.5-Math series (Zheng et al., 2024; Zhang et al., 2025c): Qwen2.5-Math-7B-MathShepherd and Qwen2.5-Math-7B-PRM800K are trained with Math-Shepherd (Wang et al., 2024b) and PRM800K (Lightman et al., 2024), respectively. For Qwen2.5-Math-PRM-7B and Qwen2.5-Math-PRM-72B, the training data is applied consensus filtering using LLM-as-ajudge (Zheng et al., 2023).
- Universal-PRM-7B (Tan et al., 2025): The method proposes an automated framework using ensemble prompting and reverse verification.
- Direct Generative PRM-7B : The method trains a direct generative PRM with the original language head via SFT using the same data as GenPRM, but without CoT and code verification.
- Dyve-14B (Zhong et al., 2025): This method dynamically applies fast or slow verification for each reasoning step.
For critic experiments, we use the following methods for comparison:
- Self-Refine (Madaan et al., 2023): This method uses the generator to self-critique and refine the solution.
- DeepSeek-R1-Distill-Qwen-7B (DeepSeek-AI et al., 2025): This model is fine-tuned based on Qwen2.5-Math-7B (Yang et al., 2024a) using high-quality reasoning data generated by DeepSeek-R1 (DeepSeek-AI et al., 2025).
Implementation Details. For RPE, we set ๐ = 0 . 8 across all experiments, with ablation studies presented in Section 4.4. Rationale data is generated using QwQ-32B (Qwen Team, 2025) and the prompt template is shown in Table A.2. Our base models are from the DeepSeek-R1-Distill series (DeepSeek-AI et al., 2025), specifically the 1.5B, 7B, and 32B parameter variants. The training configuration for our method uses a batch size of 64 and a learning rate of 2 . 0 ร 10 -6 . During evaluation, we employ a temperature of 0.6. For critique refinement experiments, we extract content within the <analyze></analyze> tags, focusing exclusively on steps predicted as negative by the policy model. The baseline methods utilize standardized prompt templates (detailed in Table A.2) to ensure consistent critique generation formats.
Table 1: ProcessBench results reported with F1 scores. The results of GenPRM are shaded . For 1.5B PRMs, bold indicates the best Pass@1 or scores superior to GPT-4o. For 7-8B and 14-72B PRMs, bold denotes the best Pass@1 or scores superior to Qwen2.5-Math-PRM-72B.
| Model | # Samples | GSM8K | MATH | Olympiad Bench | Omni- MATH | Avg. |
|-------------------------------|----------------------------|----------------------------|----------------------------|----------------------------|----------------------------|----------------------------|
| Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) |
| GPT-4o-0806 o1-mini | unk unk | 79.2 93.2 | 63.6 88.9 | 51.4 87.2 | 53.5 82.4 | 61.9 87.9 |
| PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) |
| Skywork-PRM-1.5B | unk | 59.0 | 48.0 | 19.3 | 19.2 | 36.4 |
| GenPRM-1.5B (Pass@1) | 23K | 52.8 | 66.6 | 55.1 | 54.5 | 57.3 |
| GenPRM-1.5B (Maj@8) | 23K | 51.3 | 74.4 | 65.3 | 62.5 | 63.4 |
| PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) |
| Math-Shepherd-PRM-7B | 445K | 47.9 | 29.5 | 24.8 | 23.8 | 31.5 |
| RLHFlow-PRM-Mistral-8B | 273K | 50.4 | 33.4 | 13.8 | 15.8 | 28.4 |
| RLHFlow-PRM-Deepseek-8B | 253K | 38.8 | 33.8 | 16.9 | 16.9 | 26.6 |
| Skywork-PRM-7B | unk | 70.8 | 53.6 | 22.9 | 21.0 | 42.1 |
| EurusPRM-Stage1 | 463K | 44.3 | 35.6 | 21.7 | 23.1 | 31.2 |
| EurusPRM-Stage2 | 30K | 47.3 | 35.7 | 21.2 | 20.9 | 31.3 |
| Qwen2.5-Math-7B-Math-Shepherd | 445K | 62.5 | 31.6 | 13.7 | 7.7 | 28.9 |
| Qwen2.5-Math-7B-PRM800K | 264K | 68.2 | 62.6 | 50.7 | 44.3 | 56.5 |
| Qwen2.5-Math-PRM-7B | โผ 344K | 82.4 | 77.6 | 67.5 | 66.3 | 73.5 |
| RetrievalPRM-7B | 404K | 74.6 | 71.1 | 60.2 | 57.3 | 65.8 |
| Universal-PRM-7B | unk | 85.8 | 77.7 | 67.6 | 66.4 | 74.3 |
| Direct Generative PRM-7B | 23K | 63.9 | 65.8 | 54.5 | 55.9 | 60.0 |
| GenPRM-7B (Pass@1) | 23K | 78.7 | 80.3 | 72.2 | 69.8 | 75.2 |
| GenPRM-7B (Maj@8) | 23K | 81.0 | 85.7 | 78.4 | 76.8 | 80.5 |
| PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) |
| Dyve-14B | 117K | 68.5 | 58.3 | 49.0 | 47.2 | 55.8 |
| Qwen2.5-Math-PRM-72B | โผ 344K | 87.3 | 80.6 | 74.3 | 71.1 | 78.3 |
| GenPRM-32B (Pass@1) | 23K | 83.1 | 81.7 | 72.8 | 72.8 | 77.6 |
| GenPRM-32B (Maj@8) | 23K | 85.1 | 86.3 | 78.9 | 80.1 | 82.6 |
## 4.2. ProcessBench Results
GenPRM outperforms classification-based PRMs on ProcessBench. As shown in Table 1, GenPRM7B significantly outperforms direct generative PRM and surpasses all previous PRMs with parameters less than 72B on ProcessBench. Also, GenPRM-1.5B outperforms Skywork-PRM-1.5B by a large margin. It is noteworthy that GenPRM is trained with merely 23K data from MATH (Hendrycks et al., 2021) only. By comparing the detailed results in Table 6, we can find that the performance gain of GenPRM mainly comes from the stronger abilities of finding erroneous steps and we provide concrete cases in Appendix C. These results demonstrating the superiority of generative modeling of PRM.
GenPRM enables smaller PRMs surpass 10 ร larger PRMs and GPT-4o via TTS. We also compare the TTS results of GenPRM in Table 1 and find that GenPRM-1.5B surpasses GPT-4 and GenPRM-7B exceeds Qwen2.5-Math-PRM-72B on ProcessBench via simply majority voting, showing that scaling test-time compute is highly effective for GenPRM. We also find that the performance improvement of scaling the test-time compute on harder problems is larger than that of easier questions.
## 4.3. Policy Model Test-Time Scaling Results
GenPRM as a Verifier. The results in Figure 4 (a)-(d) show that GenPRM outperforms the baselines on MATH, AMC23, AIME24, and Minerva Math with Qwen2.5-Math-7B-Instruct (Yang et al., 2024b) as the generation model. The advantage of GenPRM becomes larger by scaling the test-time compute of GenPRM and the generation model. Figure 4 (e)-(h) demonstrates that GenPRM generalizes well to responses with Gemma-3-12b-it (Gemma Team and Google DeepMind, 2025) as the generation model.
Figure 4: BoN results with different generation models on multiple mathematical benchmarks.
<details>
<summary>Image 4 Details</summary>

### Visual Description
\n
## Line Charts: Accuracy vs. Number of Solutions for Different Models
### Overview
The image presents eight line charts, arranged in a 2x4 grid, displaying the accuracy of various language models (Owen and Gemma) on four different math datasets: MATH, AMC23, AIME24, and Minerva Math. Accuracy is plotted against the number of solutions (N), ranging from 2<sup>2</sup> to 2<sup>5</sup> (4 to 32). Each chart focuses on a specific dataset and compares the performance of different model variants.
### Components/Axes
* **X-axis:** Number of Solutions (N), labeled as "Number of Solutions (N)". Scale: 4, 8, 16, 32.
* **Y-axis:** Accuracy (%), labeled as "Accuracy (%)". Scales vary per chart, but generally range from approximately 32% to 95%.
* **Legend:** Located at the top of the image, spanning all charts. The legend identifies the following models/metrics:
* Pass@K (Dark Blue)
* Maj. (Light Blue)
* GenPRM-7B (Pass@1) (Green)
* GenPRM-7B (Maj@8) (Orange)
* Qwen2.5-Math-7B (Red)
* Qwen2.5-Math-7B-PRM800K (Purple)
* Skywork-7B (Teal)
* Skywork-1.5B (Yellow)
* **Chart Titles:** Each chart is labeled with the dataset name and the model family (Owen or Gemma) in parentheses, e.g., "(a) MATH (Qwen)", "(f) AMC23 (Gemma)".
### Detailed Analysis or Content Details
**Chart (a) MATH (Qwen):**
* Pass@K: Starts at ~86%, increases to ~93% at N=32.
* Maj.: Starts at ~85%, increases to ~91% at N=32.
* GenPRM-7B (Pass@1): Starts at ~85%, increases to ~92% at N=32.
* GenPRM-7B (Maj@8): Starts at ~84%, increases to ~90% at N=32.
* Qwen2.5-Math-7B: Starts at ~86%, increases to ~93% at N=32.
* Qwen2.5-Math-7B-PRM800K: Starts at ~85%, increases to ~92% at N=32.
* Skywork-7B: Starts at ~85%, increases to ~91% at N=32.
* Skywork-1.5B: Starts at ~84%, increases to ~90% at N=32.
**Chart (b) AMC23 (Qwen):**
* Pass@K: Starts at ~72%, increases to ~95% at N=32.
* Maj.: Starts at ~72%, increases to ~94% at N=32.
* GenPRM-7B (Pass@1): Starts at ~73%, increases to ~95% at N=32.
* GenPRM-7B (Maj@8): Starts at ~72%, increases to ~94% at N=32.
* Qwen2.5-Math-7B: Starts at ~73%, increases to ~95% at N=32.
* Qwen2.5-Math-7B-PRM800K: Starts at ~72%, increases to ~94% at N=32.
* Skywork-7B: Starts at ~72%, increases to ~94% at N=32.
* Skywork-1.5B: Starts at ~71%, increases to ~93% at N=32.
**Chart (c) AIME24 (Qwen):**
* Pass@K: Starts at ~10%, increases to ~35% at N=32.
* Maj.: Starts at ~10%, increases to ~33% at N=32.
* GenPRM-7B (Pass@1): Starts at ~10%, increases to ~34% at N=32.
* GenPRM-7B (Maj@8): Starts at ~10%, increases to ~32% at N=32.
* Qwen2.5-Math-7B: Starts at ~12%, increases to ~35% at N=32.
* Qwen2.5-Math-7B-PRM800K: Starts at ~11%, increases to ~33% at N=32.
* Skywork-7B: Starts at ~11%, increases to ~33% at N=32.
* Skywork-1.5B: Starts at ~10%, increases to ~32% at N=32.
**Chart (d) Minerva Math (Qwen):**
* Pass@K: Starts at ~33%, increases to ~50% at N=32.
* Maj.: Starts at ~33%, increases to ~48% at N=32.
* GenPRM-7B (Pass@1): Starts at ~33%, increases to ~49% at N=32.
* GenPRM-7B (Maj@8): Starts at ~33%, increases to ~47% at N=32.
* Qwen2.5-Math-7B: Starts at ~34%, increases to ~50% at N=32.
* Qwen2.5-Math-7B-PRM800K: Starts at ~33%, increases to ~48% at N=32.
* Skywork-7B: Starts at ~33%, increases to ~47% at N=32.
* Skywork-1.5B: Starts at ~32%, increases to ~46% at N=32.
**Chart (e) MATH (Gemma):**
* Pass@K: Starts at ~84%, increases to ~94% at N=32.
* Maj.: Starts at ~83%, increases to ~92% at N=32.
* GenPRM-7B (Pass@1): Starts at ~84%, increases to ~93% at N=32.
* GenPRM-7B (Maj@8): Starts at ~83%, increases to ~91% at N=32.
* Qwen2.5-Math-7B: Starts at ~85%, increases to ~94% at N=32.
* Qwen2.5-Math-7B-PRM800K: Starts at ~84%, increases to ~93% at N=32.
* Skywork-7B: Starts at ~84%, increases to ~92% at N=32.
* Skywork-1.5B: Starts at ~83%, increases to ~91% at N=32.
**Chart (f) AMC23 (Gemma):**
* Pass@K: Starts at ~70%, increases to ~95% at N=32.
* Maj.: Starts at ~69%, increases to ~94% at N=32.
* GenPRM-7B (Pass@1): Starts at ~71%, increases to ~95% at N=32.
* GenPRM-7B (Maj@8): Starts at ~70%, increases to ~94% at N=32.
* Qwen2.5-Math-7B: Starts at ~72%, increases to ~95% at N=32.
* Qwen2.5-Math-7B-PRM800K: Starts at ~71%, increases to ~94% at N=32.
* Skywork-7B: Starts at ~70%, increases to ~94% at N=32.
* Skywork-1.5B: Starts at ~69%, increases to ~93% at N=32.
**Chart (g) AIME24 (Gemma):**
* Pass@K: Starts at ~15%, increases to ~40% at N=32.
* Maj.: Starts at ~15%, increases to ~38% at N=32.
* GenPRM-7B (Pass@1): Starts at ~16%, increases to ~39% at N=32.
* GenPRM-7B (Maj@8): Starts at ~15%, increases to ~37% at N=32.
* Qwen2.5-Math-7B: Starts at ~17%, increases to ~40% at N=32.
* Qwen2.5-Math-7B-PRM800K: Starts at ~16%, increases to ~38% at N=32.
* Skywork-7B: Starts at ~16%, increases to ~38% at N=32.
* Skywork-1.5B: Starts at ~15%, increases to ~37% at N=32.
**Chart (h) Minerva Math (Gemma):**
* Pass@K: Starts at ~30%, increases to ~47% at N=32.
* Maj.: Starts at ~30%, increases to ~45% at N=32.
* GenPRM-7B (Pass@1): Starts at ~31%, increases to ~46% at N=32.
* GenPRM-7B (Maj@8): Starts at ~30%, increases to ~44% at N=32.
* Qwen2.5-Math-7B: Starts at ~32%, increases to ~47% at N=32.
* Qwen2.5-Math-7B-PRM800K: Starts at ~31%, increases to ~45% at N=32.
* Skywork-7B: Starts at ~31%, increases to ~44% at N=32.
* Skywork-1.5B: Starts at ~30%, increases to ~43% at N=32.
### Key Observations
* Accuracy generally increases with the number of solutions (N) across all datasets and models.
* The Qwen2.5-Math-7B model consistently performs at or near the top across all datasets.
* The Skywork-1.5B model generally exhibits the lowest accuracy.
* The AIME24 dataset consistently shows the lowest overall accuracy compared to the other datasets.
* The performance difference between the "Pass@K" and "Maj." metrics is minimal.
* The performance of the models is relatively similar between the Owen and Gemma families.
### Interpretation
The charts demonstrate the impact of increasing the number of solutions (N) on the accuracy of various language models in solving math problems. The consistent upward trend across all datasets suggests that providing more solutions improves the models' ability to arrive at the correct answer. The superior performance of the Qwen2.5-Math-7B model indicates its effectiveness in tackling these types of mathematical challenges. The lower accuracy on the AIME24 dataset suggests that this dataset presents a higher level of difficulty or requires different reasoning skills compared to the others. The relatively small differences between the models within each dataset suggest that the core architecture and training data play a significant role in performance, while the number of solutions acts as a supplementary factor. The consistent performance between the Owen and Gemma families suggests that the underlying model architecture is comparable, and the differences in performance are likely due to variations in training data or fine-tuning strategies.
</details>
GenPRM as a Critic. We also conduct experiments by using GenPRM as a critic to refine the outputs of the policy model. The results in Table 2 and Figure 1 (right) show that GenPRM exhibits strong critique abilities than the baselines, significantly improving the performance of the policy model and the performance continues to increase with more refinement based on the critic feedback.
## 4.4. Analysis
Label Estimation Method and Criterion. To explore how different label estimation influences GenPRM, we conduct experiments with the following methods: (1) hard label (Wang et al., 2024b; Zhang et al., 2025c); (2) RPE in (10); and (3) a RPE variant ( ๐ ๐ก = ๐๐ถ ( ๐ ๐ก , ๐ ๐ก ) -๐๐ถ ( ๐ ๐ก ) ). For the RPE and its variant, we use different thresholds ๐ for evaluation and set the labels as correct by checking whether ๐ ๐ก โฅ ๐ . The results in Table 3 show that RPE and its variant outperforms hard label estimation and RPE with ๐ = 0 . 8 achieves the best result. By scaling test-time compute with majority voting, the results in Table 4 demonstrate that RPE with ๐ = 0 . 8 still reaches the best.
Table 2: Results of critique refinement experiments. The results of GenPRM are shaded . For each refinement turn, the highest values are bolded .
| Critic Model | Gemma-3-12b-it as Generator | Gemma-3-12b-it as Generator | Gemma-3-12b-it as Generator | Gemma-3-12b-it as Generator | Gemma-3-12b-it as Generator | Qwen2.5-7B-Instruct as Generator | Qwen2.5-7B-Instruct as Generator | Qwen2.5-7B-Instruct as Generator | Qwen2.5-7B-Instruct as Generator | Qwen2.5-7B-Instruct as Generator | Avg. |
|------------------------|-------------------------------|-------------------------------|-------------------------------|-------------------------------|-------------------------------|------------------------------------|------------------------------------|------------------------------------|------------------------------------|------------------------------------|--------|
| Critic Model | AMC23 | AIME24 | MATH | Minerva Math | Avg. | AMC23 | AIME24 | MATH | Minerva Math | Avg. | Avg. |
| Zero-shot | 64.1 | 15.8 | 83.8 | 31.9 | 48.9 | 51.6 | 7.1 | 76.2 | 34.5 | 42.4 | 45.7 |
| Turn 1 | Turn 1 | Turn 1 | Turn 1 | Turn 1 | Turn 1 | Turn 1 | Turn 1 | Turn 1 | Turn 1 | Turn 1 | Turn 1 |
| Generator | 66.6 | 15.8 | 84.7 | 33.3 | 50.1 | 50.6 | 8.0 | 76.8 | 34.0 | 42.4 | 46.3 |
| DeepSeek-R1-Distill-7B | 69.1 | 17.9 | 84.6 | 33.0 | 51.2 | 50.6 | 6.3 | 77.7 | 34.7 | 42.3 | 46.8 |
| GenPRM-7B | 74.1 | 19.6 | 86.0 | 35.3 | 53.8 | 57.5 | 8.3 | 80.6 | 36.5 | 45.7 | 49.8 |
| Turn 2 | Turn 2 | Turn 2 | Turn 2 | Turn 2 | Turn 2 | Turn 2 | Turn 2 | Turn 2 | Turn 2 | Turn 2 | Turn 2 |
| Generator | 66.6 | 18.0 | 84.8 | 31.6 | 50.3 | 49.8 | 8.0 | 76.9 | 31.8 | 41.6 | 46.0 |
| DeepSeek-R1-Distill-7B | 70.9 | 18.3 | 85.0 | 33.5 | 51.9 | 51.9 | 7.9 | 78.1 | 32.8 | 42.7 | 47.3 |
| GenPRM-7B | 75.0 | 21.3 | 86.9 | 35.6 | 54.7 | 59.4 | 9.6 | 82.2 | 35.0 | 46.6 | 50.7 |
| Turn 3 | Turn 3 | Turn 3 | Turn 3 | Turn 3 | Turn 3 | Turn 3 | Turn 3 | Turn 3 | Turn 3 | Turn 3 | Turn 3 |
| Generator | 67.8 | 18.1 | 85.0 | 32.1 | 50.8 | 49.7 | 8.1 | 77.1 | 30.8 | 41.4 | 46.1 |
| DeepSeek-R1-Distill-7B | 69.6 | 18.8 | 85.0 | 33.4 | 51.7 | 51.9 | 8.3 | 78.2 | 32.7 | 42.7 | 47.2 |
| GenPRM-7B | 76.2 | 22.8 | 86.7 | 36.0 | 55.4 | 62.7 | 9.3 | 82.9 | 34.9 | 47.5 | 51.5 |
Table 3: Results of GenPRM with different label estimation method and threshold on ProcessBench, reported with Pass@1. The best results are shown in bold .
| Estimation Method | Positive Label Criterion | GSM8K | MATH | Olympiad Bench | Omni- MATH | Avg. |
|------------------------------------|----------------------------|---------|--------|------------------|--------------|--------|
| ๐ ๐ก = ๐๐ถ ( ๐ ๐ก ,๐ ๐ก ) (hard label) | ๐ ๐ก > 0 | 72.9 | 78.9 | 73.2 | 68 | 73.2 |
| ๐ ๐ก = ๐๐ถ ( ๐ ๐ก ,๐ ๐ก ) - ๐๐ถ ( ๐ ๐ก ) | ๐ ๐ก โฅ - 0 . 1 | 77.3 | 79.9 | 70.8 | 68.5 | 74.1 |
| ๐ ๐ก = ๐๐ถ ( ๐ ๐ก ,๐ ๐ก ) - ๐๐ถ ( ๐ ๐ก ) | ๐ ๐ก โฅ - 0 . 3 | 76.8 | 79.6 | 71.1 | 69 | 74.1 |
| ๐ ๐ก = ๐๐ถ ( ๐ ๐ก ,๐ ๐ก ) - ๐๐ถ ( ๐ ๐ก ) | ๐ ๐ก โฅ - 0 . 5 | 75.8 | 80.2 | 72.8 | 68.6 | 74.3 |
| = ๐๐ถ ( ๐ ๐ก ,๐ ๐ก ) ๐๐ถ ( ๐ ๐ก ) | ๐ ๐ก โฅ 0 . 1 | 74.8 | 78.7 | 71.6 | 68.7 | 73.5 |
| = ๐๐ถ ( ๐ ๐ก ,๐ ๐ก ) ๐๐ถ ( ๐ ๐ก ) | ๐ ๐ก โฅ 0 . 5 | 75.7 | 79.2 | 70.4 | 68.5 | 73.5 |
| ๐ ๐ก | ๐ ๐ก โฅ 0 . 8 | 78.7 | 80.3 | 72.2 | 69.8 | 75.2 |
| | ๐ ๐ก โฅ 1 . 0 | 76.4 | 77.4 | 68.1 | 67.2 | 72.3 |
Table 4: Results of GenPRM with different label estimation method and threshold on ProcessBench, reported with Maj@8. The best results are shown in bold .
| Estimation Method | Positive Label Criterion | GSM8K | MATH | Olympiad Bench | Omni- MATH | Avg. |
|------------------------------------|----------------------------|---------|--------|------------------|--------------|--------|
| ๐ ๐ก = ๐๐ถ ( ๐ ๐ก ,๐ ๐ก ) (hard label) | ๐ ๐ก > 0 | 75.1 | 83.8 | 80.6 | 74.4 | 78.5 |
| ๐ ๐ก = ๐๐ถ ( ๐ ๐ก ,๐ ๐ก ) - ๐๐ถ ( ๐ ๐ก ) | ๐ ๐ก โฅ - 0 . 1 | 79.8 | 85.1 | 78 | 74.5 | 79.4 |
| ๐ ๐ก = ๐๐ถ ( ๐ ๐ก ,๐ ๐ก ) - ๐๐ถ ( ๐ ๐ก ) | ๐ ๐ก โฅ - 0 . 3 | 80.9 | 86.5 | 78.1 | 75 | 80.2 |
| ๐ ๐ก = ๐๐ถ ( ๐ ๐ก ,๐ ๐ก ) - ๐๐ถ ( ๐ ๐ก ) | ๐ ๐ก โฅ - 0 . 5 | 78.1 | 85.6 | 79.1 | 73.4 | 79.1 |
| = ๐๐ถ ( ๐ ๐ก ,๐ ๐ก ) ๐๐ถ ( ๐ ๐ก ) | ๐ ๐ก โฅ 0 . 1 | 77 | 84.6 | 78.1 | 75.3 | 78.7 |
| = ๐๐ถ ( ๐ ๐ก ,๐ ๐ก ) ๐๐ถ ( ๐ ๐ก ) | ๐ ๐ก โฅ 0 . 5 | 78 | 85.2 | 78.2 | 74.3 | 78.9 |
| ๐ ๐ก | ๐ ๐ก โฅ 0 . 8 | 81 | 85.7 | 78.4 | 76.8 | 80.5 |
| | ๐ ๐ก โฅ 1 . 0 | 81.1 | 84.1 | 76 | 74.7 | 79 |
Reasoning Components. To understand how each reasoning component influence GenPRM, we conduct experiments by training GenPRM with: (1) CoT data only, (2) code verification data only, and (3) full data. During inference phase, we also compare several variants. For example, GenPRM trained with full data can be used to only verify each step with CoT only by stopping generation at </analyze> token. The results in Table 5 show that: (1) the improvement of GenPRM mainly comes
from CoT reasoning; (2) generating code and reasoning with code execution result improves the process verification performance as well.
Table 5: Results on ProcessBench of GenPRM with different reasoning components, reported with Maj@8. The best results are shown in bold .
| Training | Training | Inference | Inference | Inference | GSM8K | MATH | Olympiad Bench | Omni- MATH | Avg. |
|------------|------------|-------------|-------------|-------------|---------|--------|------------------|--------------|--------|
| CoT | Code | CoT | Code | Code Exec. | GSM8K | MATH | Olympiad Bench | Omni- MATH | Avg. |
| โ | โ | โ | โ | โ | 63.9 | 65.8 | 54.5 | 55.9 | 60.0 |
| โ | โ | โ | โ | โ | 67.0 | 70.8 | 61.6 | 57.4 | 64.2 |
| โ | โ | โ | โ | โ | 70.6 | 76.6 | 67.3 | 63.9 | 69.6 |
| โ | โ | โ | โ | โ | 76.4 | 83.0 | 80.5 | 75.4 | 78.8 |
| | | โ | โ | โ | 60.1 | 66.7 | 59.9 | 59.2 | 61.5 |
| | | โ | โ | โ | 61.3 | 74.7 | 68.1 | 62.0 | 66.5 |
| โ | โ | โ | โ | โ | 78.8 | 85.1 | 78.7 | 74.9 | 79.3 |
| | | โ | โ | โ | 81.0 | 85.1 | 78.1 | 75.5 | 79.9 |
| | | โ | โ | โ | 81.0 | 85.7 | 78.4 | 76.8 | 80.5 |
## 5. Related Work
Process Reward Models. Process reward models have been proved to be effective for providing step-wise scores and are superior to outcome reward models in mathematical reasoning tasks (Uesato et al., 2022; Lightman et al., 2024). However, annotating a process supervision dataset such as PRM800K (Lightman et al., 2024) requires significant human costs. To mitigate this cost, prior works utilize Monte Carlo estimation (Wang et al., 2024b) and binary search (Luo et al., 2024) for automated label generation. Subsequent research improves PRMs through methods such as advantage modeling (Setlur et al., 2025), ๐ -value rankings (Li and Li, 2025), implicit entropy regularization (Zhang et al., 2024a), retrieval-augmented generation (Zhu et al., 2025), and fast-slow verification (Zhong et al., 2025). Furthermore, the community has developed high-quality opensource PRMs, including the RLHFlow series (Xiong et al., 2024), Math-psa (Wang et al., 2024a), Skywork series (Skywork o1 Team, 2024), and Qwen2.5-Math series (Zheng et al., 2024; Zhang et al., 2025c). Recently, a line of works focus on extending PRMs to other tasks, including coding (Zhang et al., 2024b), medical tasks (Jiang et al., 2025), agentic tasks (Choudhury, 2025), general domain tasks (Zhang et al., 2025a; Zeng et al., 2025), and multimodal tasks (Wang et al., 2025). Current studies also focus on benchmarking PRMs (Zheng et al., 2024; Song et al., 2025) to systematically evaluate their performance.
Large Language Model Test-Time Scaling. Scaling test-time computation is an effective method for improving performance during the inference phase (OpenAI, 2024a,b; DeepSeek-AI et al., 2025). TTS is commonly implemented with external verifiers (e.g., ORMs and PRMs) or strategies (e.g., beam search and MCTS) (Wu et al., 2025; Snell et al., 2025; Beeching et al., 2024; Liu et al., 2025). In this work, we scale the test-time computation of a generative PRM with an explicit reasoning process and GenPRM can also serve as a verifier or a critic model in external TTS.
Enhancing the Generative Abilities of Reward Models. Previous research has investigated methods to enhance the generative capabilities of reward models using CoT reasoning (Ankner et al., 2024; Zhang et al., 2025b; Mahan et al., 2024). For instance, CLoud reward models (Ankner et al., 2024) are trained to generate critiques for responses and predict rewards using an additional reward
head. GenRM-CoT (Zhang et al., 2025b) and GenRM (Mahan et al., 2024) train generative reward models that perform CoT reasoning before making final predictions via SFT and preference learning, respectively. CTRL (Xie et al., 2025) demonstrates that critic models exhibit strong discriminative abilities when utilized as generative reward models. Prior to these works, GRM (Yang et al., 2024c) regularizes the hidden states of reward models with a text generation loss.
## 6. Conclusion
In this work, we propose GenPRM, a generative process reward model that performs explicit reasoning and code verification for process supervision and enables scaling the test-time compute of PRMs. Experimental results on ProcessBench and several mathematical datasets show GenPRM outperforms prior PRMs. We also demonstrate that the performance of GenPRM increases via test-time scaling and GenPRM is effective as a critic model. We believe that this work provides perspectives on PRMs by demonstrating the strong TTS abilities of PRMs and extending the applications of PRMs.
Limitations. First, GenPRM provides process supervision by generative reasoning, which introduces additional computation during inference phase. Future work will investigate how to prune the reasoning process dynamically (Zhong et al., 2025). Although GenPRM focuses mainly on mathematical reasoning tasks, it is worth to explore how to apply generative reasoning on coding and general reasoning tasks in the future (Zhang et al., 2025a). Additionally, it would be interesting to leverage RL to incentivize the generative reasoning abilities of GenPRM.
## References
- AI-MO. AIME 2024, 2024a. URL https://huggingface.co/datasets/AI-MO/ aimo-validation-aime .
- AI-MO. AMC 2023, 2024b. URL https://huggingface.co/datasets/AI-MO/ aimo-validation-amc .
- Zachary Ankner, Mansheej Paul, Brandon Cui, Jonathan D Chang, and Prithviraj Ammanabrolu. Critique-out-Loud Reward Models. arXiv preprint arXiv:2408.11791 , 2024.
- Anthropic. Introducing Claude, 2023. URL https://www.anthropic.com/index/ introducing-claude/ .
- Edward Beeching, Lewis Tunstall, and Sasha Rush. Scaling Test-Time Compute with Open Models, 2024. URL https://huggingface.co/spaces/HuggingFaceH4/ blogpost-scaling-test-time-compute .
- Bradley Brown, Jordan Juravsky, Ryan Ehrlich, Ronald Clark, Quoc V Le, Christopher Rรฉ, and Azalia Mirhoseini. Large Language Monkeys: Scaling Inference Compute with Repeated Sampling. arXiv preprint arXiv:2407.21787 , 2024.
- Sanjiban Choudhury. Process Reward Models for LLM Agents: Practical Framework and Directions. arXiv preprint arXiv:2502.10325 , 2025.
- Ganqu Cui, Lifan Yuan, Zefan Wang, Hanbin Wang, Wendi Li, Bingxiang He, Yuchen Fan, Tianyu Yu, Qixin Xu, Weize Chen, et al. Process Reinforcement through Implicit Rewards. arXiv preprint arXiv:2502.01456 , 2025.
- DeepSeek-AI, Daya Guo, Dejian Yang, Haowei Zhang, Junxiao Song, Ruoyu Zhang, Runxin Xu, Qihao Zhu, Shirong Ma, Peiyi Wang, Xiao Bi, Xiaokang Zhang, Xingkai Yu, Yu Wu, Z. F. Wu, Zhibin Gou, Zhihong Shao, Zhuoshu Li, Ziyi Gao, Aixin Liu, Bing Xue, Bingxuan Wang, Bochao Wu, Bei Feng, Chengda Lu, Chenggang Zhao, Chengqi Deng, Chenyu Zhang, Chong Ruan, Damai Dai, Deli Chen, Dongjie Ji, Erhang Li, Fangyun Lin, Fucong Dai, Fuli Luo, Guangbo Hao, Guanting Chen, Guowei Li, H. Zhang, Han Bao, Hanwei Xu, Haocheng Wang, Honghui Ding, Huajian Xin, Huazuo Gao, Hui Qu, Hui Li, Jianzhong Guo, Jiashi Li, Jiawei Wang, Jingchang Chen, Jingyang Yuan, Junjie Qiu, Junlong Li, J. L. Cai, Jiaqi Ni, Jian Liang, Jin Chen, Kai Dong, Kai Hu, Kaige Gao, Kang Guan, Kexin Huang, Kuai Yu, Lean Wang, Lecong Zhang, Liang Zhao, Litong Wang, Liyue Zhang, Lei Xu, Leyi Xia, Mingchuan Zhang, Minghua Zhang, Minghui Tang, Meng Li, Miaojun Wang, Mingming Li, Ning Tian, Panpan Huang, Peng Zhang, Qiancheng Wang, Qinyu Chen, Qiushi Du, Ruiqi Ge, Ruisong Zhang, Ruizhe Pan, Runji Wang, R. J. Chen, R. L. Jin, Ruyi Chen, Shanghao Lu, Shangyan Zhou, Shanhuang Chen, Shengfeng Ye, Shiyu Wang, Shuiping Yu, Shunfeng Zhou, Shuting Pan, S. S. Li, Shuang Zhou, Shaoqing Wu, Shengfeng Ye, Tao Yun, Tian Pei, Tianyu Sun, T. Wang, Wangding Zeng, Wanjia Zhao, Wen Liu, Wenfeng Liang, Wenjun Gao, Wenqin Yu, Wentao Zhang, W. L. Xiao, Wei An, Xiaodong Liu, Xiaohan Wang, Xiaokang Chen, Xiaotao Nie, Xin Cheng, Xin Liu, Xin Xie, Xingchao Liu, Xinyu Yang, Xinyuan Li, Xuecheng Su, Xuheng Lin, X. Q. Li, Xiangyue Jin, Xiaojin Shen, Xiaosha Chen, Xiaowen Sun, Xiaoxiang Wang, Xinnan Song, Xinyi Zhou, Xianzu Wang, Xinxia Shan, Y. K. Li, Y. Q. Wang, Y. X. Wei, Yang Zhang, Yanhong Xu, Yao Li, Yao Zhao, Yaofeng Sun, Yaohui Wang, Yi Yu, Yichao Zhang, Yifan Shi, Yiliang Xiong, Ying He, Yishi Piao, Yisong Wang, Yixuan Tan, Yiyang Ma, Yiyuan Liu, Yongqiang Guo, Yuan Ou, Yuduan Wang, Yue Gong, Yuheng Zou, Yujia He, Yunfan Xiong, Yuxiang Luo, Yuxiang You, Yuxuan Liu, Yuyang Zhou, Y. X. Zhu, Yanhong Xu, Yanping Huang, Yaohui Li, Yi Zheng, Yuchen Zhu, Yunxian Ma, Ying Tang, Yukun Zha, Yuting Yan, Z. Z. Ren, Zehui Ren, Zhangli Sha, Zhe Fu, Zhean Xu, Zhenda Xie, Zhengyan Zhang, Zhewen Hao, Zhicheng Ma, Zhigang Yan, Zhiyu Wu, Zihui Gu, Zijia Zhu, Zijun Liu, Zilin Li, Ziwei Xie, Ziyang Song, Zizheng Pan, Zhen Huang, Zhipeng Xu, Zhongyu Zhang, and Zhen Zhang. DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning. arXiv preprint arXiv:2501.12948 , 2025.
- Gemma Team and Google DeepMind. Introducing Gemma 3: The most capable model you can run on a single GPU or TPU, March 2025. URL https://blog.google/technology/developers/ gemma-3 .
- Zhibin Gou, Zhihong Shao, Yeyun Gong, yelong shen, Yujiu Yang, Nan Duan, and Weizhu Chen. CRITIC: Large Language Models Can Self-Correct with Tool-Interactive Critiquing. In International Conference on Learning Representations (ICLR) , 2024. URL https://openreview.net/forum? id=Sx038qxjek .
- Dan Hendrycks, Collin Burns, Saurav Kadavath, Akul Arora, Steven Basart, Eric Tang, Dawn Song, and Jacob Steinhardt. Measuring Mathematical Problem Solving With the MATH Dataset. In Advances in Neural Information Processing Systems Datasets and Benchmarks Track (Round 2) , 2021. URL https://openreview.net/forum?id=7Bywt2mQsCe .
- Shuyang Jiang, Yusheng Liao, Zhe Chen, Ya Zhang, Yanfeng Wang, and Yu Wang. MedS 3 : Towards Medical Small Language Models with Self-Evolved Slow Thinking. arXiv preprint arXiv:2501.12051 , 2025.
- Aitor Lewkowycz, Anders Andreassen, David Dohan, Ethan Dyer, Henryk Michalewski, Vinay Ramasesh, Ambrose Slone, Cem Anil, Imanol Schlag, Theo Gutman-Solo, Yuhuai Wu, Behnam Neyshabur, Guy Gur-Ari, and Vedant Misra. Solving Quantitative Reasoning Problems with Language Models. In S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh, editors,
- Advances in Neural Information Processing Systems (NeurIPS) , volume 35, pages 3843-3857. Curran Associates, Inc., 2022. URL https://proceedings.neurips.cc/paper\_files/paper/ 2022/file/18abbeef8cfe9203fdf9053c9c4fe191-Paper-Conference.pdf .
- Wendi Li and Yixuan Li. Process Reward Model with Q-value Rankings. In International Conference on Learning Representations (ICLR) , 2025. URL https://openreview.net/forum?id= wQEdh2cgEk .
- Hunter Lightman, Vineet Kosaraju, Yuri Burda, Harrison Edwards, Bowen Baker, Teddy Lee, Jan Leike, John Schulman, Ilya Sutskever, and Karl Cobbe. Let's Verify Step by Step. In International Conference on Learning Representations (ICLR) , 2024. URL https://openreview.net/forum? id=v8L0pN6EOi .
- Runze Liu, Fengshuo Bai, Yali Du, and Yaodong Yang. Meta-Reward-Net: Implicitly Differentiable Reward Learning for Preference-based Reinforcement Learning. In Advances in Neural Information Processing Systems (NeurIPS) , volume 35, pages 22270-22284, 2022.
- Runze Liu, Yali Du, Fengshuo Bai, Jiafei Lyu, and Xiu Li. PEARL: Zero-shot Cross-task Preference Alignment and Robust Reward Learning for Robotic Manipulation. In Ruslan Salakhutdinov, Zico Kolter, Katherine Heller, Adrian Weller, Nuria Oliver, Jonathan Scarlett, and Felix Berkenkamp, editors, International Conference on Machine Learning (ICML) , volume 235 of Proceedings of Machine Learning Research , pages 30946-30964. PMLR, 21-27 Jul 2024. URL https://proceedings. mlr.press/v235/liu24o.html .
- Runze Liu, Junqi Gao, Jian Zhao, Kaiyan Zhang, Xiu Li, Biqing Qi, Wanli Ouyang, and Bowen Zhou. Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling. arXiv preprint arXiv:2502.06703 , 2025.
- Liangchen Luo, Yinxiao Liu, Rosanne Liu, Samrat Phatale, Harsh Lara, Yunxuan Li, Lei Shu, Yun Zhu, Lei Meng, Jiao Sun, et al. Improve Mathematical Reasoning in Language Models by Automated Process Supervision. arXiv preprint arXiv:2406.06592 , 2024.
- Aman Madaan, Niket Tandon, Prakhar Gupta, Skyler Hallinan, Luyu Gao, Sarah Wiegreffe, Uri Alon, Nouha Dziri, Shrimai Prabhumoye, Yiming Yang, Shashank Gupta, Bodhisattwa Prasad Majumder, Katherine Hermann, Sean Welleck, Amir Yazdanbakhsh, and Peter Clark. Self-Refine: Iterative Refinement with Self-Feedback. In A. Oh, T. Naumann, A. Globerson, K. Saenko, M. Hardt, and S. Levine, editors, Advances in Neural Information Processing Systems (NeurIPS) , volume 36, pages 46534-46594. Curran Associates, Inc., 2023. URL https://proceedings.neurips.cc/paper\_files/paper/2023/file/ 91edff07232fb1b55a505a9e9f6c0ff3-Paper-Conference.pdf .
- Dakota Mahan, Duy Van Phung, Rafael Rafailov, Chase Blagden, Nathan Lile, Louis Castricato, JanPhilipp Frรคnken, Chelsea Finn, and Alon Albalak. Generative Reward Models. arXiv preprint arXiv:2410.12832 , 2024.
OpenAI. GPT-4 Technical Report. arXiv preprint arXiv:2303.08774 , 2023.
- OpenAI. Learning to reason with LLMs, 2024a. URL https://openai.com/index/ learning-to-reason-with-llms .
- OpenAI. OpenAI o3-mini, 2024b. URL https://openai.com/index/openai-o3-mini .
- Qwen Team. QwQ-32B: Embracing the Power of Reinforcement Learning, March 2025. URL https://qwenlm.github.io/blog/qwq-32b .
- Amrith Setlur, Chirag Nagpal, Adam Fisch, Xinyang Geng, Jacob Eisenstein, Rishabh Agarwal, Alekh Agarwal, Jonathan Berant, and Aviral Kumar. Rewarding Progress: Scaling Automated Process Verifiers for LLM Reasoning. In International Conference on Learning Representations (ICLR) , 2025. URL https://openreview.net/forum?id=A6Y7AqlzLW .
- Zhihong Shao, Peiyi Wang, Qihao Zhu, Runxin Xu, Junxiao Song, Xiao Bi, Haowei Zhang, Mingchuan Zhang, YK Li, Y Wu, et al. DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models. arXiv preprint arXiv:2402.03300 , 2024.
- Skywork o1 Team. Skywork-o1 Open Series. https://huggingface.co/Skywork , November 2024. URL https://huggingface.co/Skywork .
- Charlie Victor Snell, Jaehoon Lee, Kelvin Xu, and Aviral Kumar. Scaling LLM Test-Time Compute Optimally Can be More Effective than Scaling Parameters for Reasoning. In International Conference on Learning Representations (ICLR) , 2025. URL https://openreview.net/forum?id= 4FWAwZtd2n .
- Mingyang Song, Zhaochen Su, Xiaoye Qu, Jiawei Zhou, and Yu Cheng. PRMBench: A Fine-grained and Challenging Benchmark for Process-Level Reward Models. arXiv preprint arXiv:2501.03124 , 2025.
- Xiaoyu Tan, Tianchu Yao, Chao Qu, Bin Li, Minghao Yang, Dakuan Lu, Haozhe Wang, Xihe Qiu, Wei Chu, Yinghui Xu, et al. AURORA: Automated Training Framework of Universal Process Reward Models via Ensemble Prompting and Reverse Verification. arXiv preprint arXiv:2502.11520 , 2025.
- Jonathan Uesato, Nate Kushman, Ramana Kumar, Francis Song, Noah Siegel, Lisa Wang, Antonia Creswell, Geoffrey Irving, and Irina Higgins. Solving math word problems with process-and outcome-based feedback. arXiv preprint arXiv:2211.14275 , 2022.
- Jun Wang, Meng Fang, Ziyu Wan, Muning Wen, Jiachen Zhu, Anjie Liu, Ziqin Gong, Yan Song, Lei Chen, Lionel M Ni, et al. OpenR: An open source framework for advanced reasoning with large language models. arXiv preprint arXiv:2410.09671 , 2024a.
- Peiyi Wang, Lei Li, Zhihong Shao, Runxin Xu, Damai Dai, Yifei Li, Deli Chen, Yu Wu, and Zhifang Sui. Math-Shepherd: Verify and Reinforce LLMs Step-by-step without Human Annotations. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) , pages 9426-9439, 2024b.
- Weiyun Wang, Zhangwei Gao, Lianjie Chen, Chen Zhe, Jinguo Zhu, Xiangyu Zhao, Yangzhou Liu, Yue Cao, Shenglong Ye, Xizhou Zhu, Lewei Lu, Haodong Duan, Yu Qiao, Jifeng Dai, and Wenhai Wang. VisualPRM: An Effective Process Reward Model for Multimodal Reasoning. arXiv preprint arXiv:2503.10291 , 2025.
- Xuezhi Wang, Jason Wei, Dale Schuurmans, Quoc V Le, Ed H. Chi, Sharan Narang, Aakanksha Chowdhery, and Denny Zhou. Self-Consistency Improves Chain of Thought Reasoning in Language Models. In International Conference on Learning Representations (ICLR) , 2023. URL https: //openreview.net/forum?id=1PL1NIMMrw .
- Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Fei Xia, Ed Chi, Quoc V Le, Denny Zhou, et al. Chain-of-thought prompting elicits reasoning in large language models. In Advances in neural information processing systems (NeurIPS) , volume 35, pages 24824-24837, 2022.
- Yangzhen Wu, Zhiqing Sun, Shanda Li, Sean Welleck, and Yiming Yang. Inference Scaling Laws: An Empirical Analysis of Compute-Optimal Inference for LLM Problem-Solving. In International Conference on Learning Representations (ICLR) , 2025. URL https://openreview.net/forum? id=VNckp7JEHn .
- Zhihui Xie, Liyu Chen, Weichao Mao, Jingjing Xu, Lingpeng Kong, et al. Teaching Language Models to Critique via Reinforcement Learning. arXiv preprint arXiv:2502.03492 , 2025.
- Wei Xiong, Hanning Zhang, Nan Jiang, and Tong Zhang. An Implementation of Generative PRM. https://github.com/RLHFlow/RLHF-Reward-Modeling , 2024.
- An Yang, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu, Chengyuan Li, Dayiheng Liu, Fei Huang, Haoran Wei, Huan Lin, Jian Yang, Jianhong Tu, Jianwei Zhang, Jianxin Yang, Jiaxi Yang, Jingren Zhou, Junyang Lin, Kai Dang, Keming Lu, Keqin Bao, Kexin Yang, Le Yu, Mei Li, Mingfeng Xue, Pei Zhang, Qin Zhu, Rui Men, Runji Lin, Tianhao Li, Tingyu Xia, Xingzhang Ren, Xuancheng Ren, Yang Fan, Yang Su, Yichang Zhang, Yu Wan, Yuqiong Liu, Zeyu Cui, Zhenru Zhang, and Zihan Qiu. Qwen2.5 Technical Report. arXiv preprint arXiv:2412.15115 , 2024a.
- An Yang, Beichen Zhang, Binyuan Hui, Bofei Gao, Bowen Yu, Chengpeng Li, Dayiheng Liu, Jianhong Tu, Jingren Zhou, Junyang Lin, Keming Lu, Mingfeng Xue, Runji Lin, Tianyu Liu, Xingzhang Ren, and Zhenru Zhang. Qwen2.5-Math Technical Report: Toward Mathematical Expert Model via Self-Improvement. arXiv preprint arXiv:2409.12122 , 2024b.
- Rui Yang, Ruomeng Ding, Yong Lin, Huan Zhang, and Tong Zhang. Regularizing Hidden States Enables Learning Generalizable Reward Model for LLMs. In Advances in Neural Information Processing Systems (NeurIPS) , 2024c. URL https://openreview.net/forum?id=jwh9MHEfmY .
- Lifan Yuan, Wendi Li, Huayu Chen, Ganqu Cui, Ning Ding, Kaiyan Zhang, Bowen Zhou, Zhiyuan Liu, and Hao Peng. Free Process Rewards without Process Labels. arXiv preprint arXiv:2412.01981 , 2024.
- Thomas Zeng, Shuibai Zhang, Shutong Wu, Christian Classen, Daewon Chae, Ethan Ewer, Minjae Lee, Heeju Kim, Wonjun Kang, Jackson Kunde, et al. VersaPRM: Multi-Domain Process Reward Model via Synthetic Reasoning Data. arXiv preprint arXiv:2502.06737 , 2025.
- Hanning Zhang, Pengcheng Wang, Shizhe Diao, Yong Lin, Rui Pan, Hanze Dong, Dylan Zhang, Pavlo Molchanov, and Tong Zhang. Entropy-Regularized Process Reward Model. arXiv preprint arXiv:2412.11006 , 2024a.
- Kaiyan Zhang, Jiayuan Zhang, Haoxin Li, Xuekai Zhu, Ermo Hua, Xingtai Lv, Ning Ding, Biqing Qi, and Bowen Zhou. OpenPRM: Building Open-domain Process-based Reward Models with Preference Trees. In International Conference on Learning Representations (ICLR) , 2025a. URL https://openreview.net/forum?id=fGIqGfmgkW .
- Lunjun Zhang, Arian Hosseini, Hritik Bansal, Mehran Kazemi, Aviral Kumar, and Rishabh Agarwal. Generative Verifiers: Reward Modeling as Next-Token Prediction. In International Conference on Learning Representations (ICLR) , 2025b. URL https://openreview.net/forum?id=Ccwp4tFEtE .
- Yuxiang Zhang, Shangxi Wu, Yuqi Yang, Jiangming Shu, Jinlin Xiao, Chao Kong, and Jitao Sang. o1-Coder: an o1 Replication for Coding. arXiv preprint arXiv:2412.00154 , 2024b.
- Zhenru Zhang, Chujie Zheng, Yangzhen Wu, Beichen Zhang, Runji Lin, Bowen Yu, Dayiheng Liu, Jingren Zhou, and Junyang Lin. The Lessons of Developing Process Reward Models in Mathematical Reasoning. arXiv preprint arXiv:2501.07301 , 2025c.
- Chujie Zheng, Zhenru Zhang, Beichen Zhang, Runji Lin, Keming Lu, Bowen Yu, Dayiheng Liu, Jingren Zhou, and Junyang Lin. ProcessBench: Identifying Process Errors in Mathematical Reasoning. arXiv preprint arXiv:2412.06559 , 2024.
- Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, Siyuan Zhuang, Zhanghao Wu, Yonghao Zhuang, Zi Lin, Zhuohan Li, Dacheng Li, Eric Xing, Hao Zhang, Joseph E Gonzalez, and Ion Stoica. Judging LLM-as-a-Judge with MT-Bench and Chatbot Arena. In A. Oh, T. Naumann, A. Globerson, K. Saenko, M. Hardt, and S. Levine, editors, Advances in Neural Information Processing Systems (NeurIPS) , volume 36, pages 46595-46623. Curran Associates, Inc., 2023. URL https://proceedings.neurips.cc/paper\_files/paper/2023/file/ 91f18a1287b398d378ef22505bf41832-Paper-Datasets\_and\_Benchmarks.pdf .
- Jianyuan Zhong, Zeju Li, Zhijian Xu, Xiangyu Wen, and Qiang Xu. Dyve: Thinking Fast and Slow for Dynamic Process Verification. arXiv preprint arXiv:2502.11157 , 2025.
- Jiachen Zhu, Congmin Zheng, Jianghao Lin, Kounianhua Du, Ying Wen, Yong Yu, Jun Wang, and Weinan Zhang. Retrieval-Augmented Process Reward Model for Generalizable Mathematical Reasoning. arXiv preprint arXiv:2502.14361 , 2025.
- Xuekai Zhu, Biqing Qi, Kaiyan Zhang, Xinwei Long, Zhouhan Lin, and Bowen Zhou. PaD: Programaided Distillation Can Teach Small Models Reasoning Better than Chain-of-thought Fine-tuning. In Kevin Duh, Helena Gomez, and Steven Bethard, editors, Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers) , pages 2571-2597, Mexico City, Mexico, June 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.naacl-long.142. URL https://aclanthology.org/2024.naacl-long.142/ .
## A. Experimental Details
## A.1. Scoring and Voting Methods
PRM-Last. PRM-Last considers the process reward of the last step of the entire LLM response as the final score, i.e., score = ๐ ๐ .
PRM-Avg. PRM-Avg computes the mean process reward across all steps as the final score, i.e., score = 1 ๐ โ๏ธ ๐ ๐ก =1 ๐ ๐ก .
PRM-Min. PRM-Min uses the minimum process reward across all steps as the final score, i.e., score = min ๐ { ๐ ๐ก } ๐ ๐ก =1 .
## A.2. Implementation Details
Prompt for CoT and code rationale generation is shown in Table A.2.
## Prompt for CoT and code rationale generation
## [System]:
You are a math teacher. Your task is to review and critique the paragraphs in solution step by step with python code.
## [User]:
The following is the math problem and a solution (split into paragraphs, enclosed with tags and indexed from 1):
[Math Problem] {problem} [ Solution] <paragraph\_1> {solution\_section\_1} </paragraph\_1> ... <paragraph\_n> {solution\_section\_n} </paragraph\_n> Your task is to verify the correctness of paragraph in the solution. Split your verification by '### Paragraph {{ID}}' .
Your verification for each paragraph should be constructed by 2 parts, wrapped by '<analyze></analyze>' and '<verify></verify>' separately.
1. In '<analyze></analyze>' part, you need to analyze the reasoning process and explain why the paragraph is correct or incorrect in detail.
2. In '<verify></verify>' part, you must write **Python code** in the form of '''python\n{{CODE}}\n''' to verify every details that can be verified by code. You can import PyPI (i.e., 'sympy', 'scipy' and so on) to implement complicated calculation. Make sure to print the critic results in the code. Every code will be executed automatically by system. You need to analyze the '[Code Output]' after code executing.
3. >Pay attention that you must follow the format of '''python\n{{CODE}}\n''' when you write the code, otherwise the code will not be executed.
After all verifications, if you identify an error in a paragraph, return the **index of the paragraph where the earliest error occurs**. Otherwise, return the **index of -1 (which typically denotes "not found")**. Please put your final answer (i.e., the index) within box in the form of '$\\boxed{{INDEX}}$' .
Following Zheng et al. (2024); Zhang et al. (2025c), we use the prompt in Table A.2 to evaluate LLM-as-a-judge methods on ProcessBench (Zheng et al., 2024).
## Evaluation prompt for LLM-as-a-judge methods on ProcessBench
I will provide a math problem along with a solution. They will be formatted as follows:
[Math Problem]
```
```
Your task is to review each paragraph of the solution in sequence, analyzing, verifying, and critiquing the reasoning in detail. You need to provide the analyses and the conclusion in the following format:
```
```
... <analysis\_n> ...(analysis of paragraph n)... </analysis\_n> <conclusion> Correct/Incorrect </conclusion> * When you analyze each paragraph, you should use proper verification, recalculation, or reflection to indicate whether it is logically and mathematically valid. Please elaborate on the analysis process carefully. * If an error is detected in any paragraph, you should describe the nature and cause of the error in detail, and suggest how to correct the error or the correct approach. Once a paragraph is found to contain any error, stop further analysis of subsequent paragraphs (as they may depend on the identified error) and directly provide the conclusion of "Incorrect." For instance, given a solution of five paragraphs, if an error is found in the third paragraph, you should reply in the following format: <analysis\_1> ...(analysis of paragraph 1)... </analysis\_1> <analysis\_2> ...(analysis of paragraph 2)... </analysis\_2> <analysis\_3> ...(analysis of paragraph 3; since an error is found here, also provide detailed critique and correction guideline)... </analysis\_3> <conclusion> Incorrect </conclusion> Note that the analyses of paragraphs 4 and 5 should be skipped as the paragraph 3 has been found to contain an error. * Respond with your analyses and conclusion directly. -----------------The following is the math problem and the solution for you task:
[Math Problem]
{tagged\_problem}
[ Solution]
{tagged\_response}
## Prompt for critique generation
## [User]:
The following is a math problem and my solution. Your task is to review and critique the paragraphs in solution step by step. Pay attention that you should not solve the problem and give the final answer. All of your task is to critique. Output your judgement of whether the paragraph is correct in the form of '\\boxed{{Yes|No}}' at the end of each paragraph verification:
```
```
## B. Additional Results
We provide full results of ProcessBench in Table 6.
Model Size. We investigate the impact of model size on GenPRM by evaluating variants with 1.5B, 7B, and 32B parameters. As shown in Table 7, scaling the model from 1.5B to 7B parameters yields substantial performance gains (57.3 โ 75.2 and 63.4 โ 80.5). However, further increasing the model size to 32B provides only marginal improvements, suggesting that the 7B variant offers the best balance between efficiency and effectiveness.
Data Size. To assess the influence of training data volume, we train GenPRM on progressively larger subsets of ProcessBench (25%, 50%, and 100% of the full dataset). Table 8 demonstrates that Pass@1 F1 scores improve rapidly with initial data increases, but the growth rate slows substantially with additional data.
Inference Tokens. We provide statistics of the reasoning tokens per step and per response in Table 9.
Table 6: Full results of critic models and PRMs on ProcessBench.
| Model | GSM8K | GSM8K | GSM8K | MATH | MATH | MATH | OlympiadBench | OlympiadBench | OlympiadBench | Omni-MATH | Omni-MATH | Omni-MATH | Avg. |
|-------------------------------|----------------------------|----------------------------|----------------------------|----------------------------|----------------------------|----------------------------|----------------------------|----------------------------|----------------------------|----------------------------|----------------------------|----------------------------|----------------------------|
| Model | Err. | Corr. | F1 | Err. | Corr. | F1 | Err. | Corr. | F1 | Err. | Corr. | F1 | F1 |
| Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) | Proprietary LLMs ( Critic) |
| GPT-4-0806 | 70.0 | 91.2 | 79.2 | 54.4 | 76.6 | 63.6 | 45.8 | 58.4 | 51.4 | 45.2 | 65.6 | 53.5 | 61.9 |
| o1-mini | 88.9 | 97.9 | 93.2 | 83.5 | 95.1 | 88.9 | 80.2 | 95.6 | 87.2 | 74.8 | 91.7 | 82.4 | 87.9 |
| Open-Source LLMs ( Critic) | Open-Source LLMs ( Critic) | Open-Source LLMs ( Critic) | Open-Source LLMs ( Critic) | Open-Source LLMs ( Critic) | Open-Source LLMs ( Critic) | Open-Source LLMs ( Critic) | Open-Source LLMs ( Critic) | Open-Source LLMs ( Critic) | Open-Source LLMs ( Critic) | Open-Source LLMs ( Critic) | Open-Source LLMs ( Critic) | Open-Source LLMs ( Critic) | Open-Source LLMs ( Critic) |
| Llama-3-8B-Instruct | 42.5 | 7.8 | 13.1 | 28.6 | 9.1 | 13.8 | 27.1 | 2.7 | 4.8 | 26.1 | 8.3 | 12.6 | 11.1 |
| Llama-3-70B-Instruct | 35.7 | 96.9 | 52.2 | 13.0 | 93.3 | 22.8 | 12.0 | 92.0 | 21.2 | 11.2 | 91.7 | 20.0 | 29.1 |
| Llama-3.1-8B-Instruct | 44.4 | 6.2 | 10.9 | 41.9 | 2.7 | 5.1 | 32.4 | 1.5 | 2.8 | 32.0 | 0.8 | 1.6 | 5.1 |
| Llama-3.1-70B-Instruct | 64.3 | 89.6 | 74.9 | 35.4 | 75.6 | 48.2 | 35.1 | 69.9 | 46.7 | 30.7 | 61.8 | 41.0 | 52.7 |
| Llama-3.3-70B-Instruct | 72.5 | 96.9 | 82.9 | 43.3 | 94.6 | 59.4 | 31.0 | 94.1 | 46.7 | 28.2 | 90.5 | 43.0 | 58.0 |
| Qwen2.5-Math-7B-Instruct | 15.5 | 100.0 | 26.8 | 14.8 | 96.8 | 25.7 | 7.7 | 91.7 | 14.2 | 6.9 | 88.0 | 12.7 | 19.9 |
| Qwen2.5-Math-72B-Instruct | 49.8 | 96.9 | 65.8 | 36.0 | 94.3 | 52.1 | 19.5 | 97.3 | 32.5 | 19.0 | 96.3 | 31.7 | 45.5 |
| Qwen2.5-Coder-7B-Instruct | 7.7 | 100.0 | 14.3 | 3.4 | 98.3 | 6.5 | 2.1 | 99.1 | 4.1 | 0.9 | 98.3 | 1.8 | 6.7 |
| Qwen2.5-Coder-14B-Instruct | 33.8 | 96.4 | 50.1 | 25.4 | 92.4 | 39.9 | 20.7 | 94.1 | 34.0 | 15.9 | 94.2 | 27.3 | 37.8 |
| Qwen2.5-Coder-32B-Instruct | 54.1 | 94.8 | 68.9 | 44.9 | 90.6 | 60.1 | 33.4 | 91.2 | 48.9 | 31.5 | 87.6 | 46.3 | 56.1 |
| Qwen2-7B-Instruct | 40.6 | 4.7 | 8.4 | 30.5 | 13.8 | 19.0 | 22.4 | 10.9 | 14.7 | 20.0 | 8.7 | 12.1 | 13.6 |
| Qwen2-72B-Instruct | 57.0 | 82.9 | 67.6 | 37.7 | 70.9 | 49.2 | 34.0 | 55.2 | 42.1 | 32.3 | 53.1 | 40.2 | 49.8 |
| Qwen2.5-7B-Instruct | 40.6 | 33.2 | 36.5 | 30.8 | 45.1 | 36.6 | 26.5 | 33.9 | 29.7 | 26.2 | 28.6 | 27.4 | 32.6 |
| Qwen2.5-14B-Instruct | 54.6 | 94.8 | 69.3 | 38.4 | 87.4 | 53.3 | 31.5 | 78.8 | 45.0 | 28.3 | 76.3 | 41.3 | 52.2 |
| Qwen2.5-32B-Instruct | 49.3 | 97.9 | 65.6 | 36.7 | 95.8 | 53.1 | 25.3 | 95.9 | 40.0 | 24.1 | 92.5 | 38.3 | 49.3 |
| Qwen2.5-72B-Instruct | 62.8 | 96.9 | 76.2 | 46.3 | 93.1 | 61.8 | 38.7 | 92.6 | 54.6 | 36.6 | 90.9 | 52.2 | 61.2 |
| QwQ-32B-Preview | 81.6 | 95.3 | 88.0 | 78.1 | 79.3 | 78.7 | 61.4 | 54.6 | 57.8 | 55.7 | 68.0 | 61.3 | 71.5 |
| PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) | PRMs (1.5B) |
| Skywork-PRM-1.5B | 50.2 | 71.5 | 59.0 | 37.9 | 65.2 | 48.0 | 15.4 | 26.0 | 19.3 | 13.6 | 32.8 | 19.2 | 36.4 |
| GenPRM-1.5B (Pass@1) | 37.0 | 92.7 | 52.8 | 57.1 | 80.1 | 66.6 | 47.0 | 66.5 | 55.1 | 45.2 | 68.7 | 54.5 | 57.3 |
| GenPRM-1.5B (Maj@8) | 34.8 | 97.4 | 51.3 | 64.7 | 87.7 | 74.4 | 57.2 | 76.1 | 65.3 | 51.3 | 80.1 | 62.5 | 63.4 |
| PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) | PRMs (7-8B) |
| Math-Shepherd-PRM-7B | 32.4 | 91.7 | 47.9 | 18.0 | 82.0 | 29.5 | 15.0 | 71.1 | 24.8 | 14.2 | 73.0 | 23.8 | 31.5 |
| RLHFlow-PRM-Mistral-8B | 33.8 | 99.0 | 50.4 | 21.7 | 72.2 | 33.4 | 8.2 | 43.1 | 13.8 | 9.6 | 45.2 | 15.8 | 28.4 |
| RLHFlow-PRM-Deepseek-8B | 24.2 | 98.4 | 38.8 | 21.4 | 80.0 | 33.8 | 10.1 | 51.0 | 16.9 | 10.9 | 51.9 | 16.9 | 26.6 |
| Skywork-PRM-7B | 61.8 | 82.9 | 70.8 | 43.8 | 62.2 | 53.6 | 17.9 | 31.9 | 22.9 | 14.0 | 41.9 | 21.0 | 42.1 |
| EurusPRM-Stage1 | 46.9 | 42.0 | 44.3 | 33.3 | 38.2 | 35.6 | 23.9 | 19.8 | 21.7 | 21.9 | 24.5 | 23.1 | 31.2 |
| EurusPRM-Stage2 | 51.2 | 44.0 | 47.3 | 36.4 | 35.0 | 35.7 | 25.7 | 18.0 | 21.2 | 23.1 | 19.1 | 20.9 | 31.3 |
| Qwen2.5-Math-7B-Math-Shepherd | 46.4 | 95.9 | 62.5 | 18.9 | 96.6 | 31.6 | 7.4 | 93.8 | 13.7 | 4.0 | 95.0 | 7.7 | 28.9 |
| Qwen2.5-Math-7B-PRM800K | 53.1 | 95.3 | 68.2 | 48.0 | 90.1 | 62.6 | 35.7 | 87.3 | 50.7 | 29.8 | 86.1 | 44.3 | 56.5 |
| Qwen2.5-Math-PRM-7B | 72.0 | 96.4 | 82.4 | 68.0 | 90.4 | 77.6 | 55.7 | 85.5 | 67.5 | 55.2 | 83.0 | 66.3 | 73.5 |
| RetrievalPRM-7B | 64.7 | 88.1 | 74.6 | 67.2 | 75.6 | 71.1 | 56.0 | 65.2 | 60.2 | 52.8 | 62.7 | 57.3 | 65.8 |
| Universal-PRM-7B | - | - | 85.8 | - | - | 77.7 | - | - | 67.6 | - | - | 66.4 | 74.3 |
| Direct Generative PRM-7B | 52.7 | 81.4 | 63.9 | 55.9 | 80.0 | 65.8 | 44.8 | 69.6 | 54.5 | 45.5 | 72.6 | 55.9 | 60.0 |
| GenPRM-7B (Pass@1) | 67.7 | 94.0 | 78.7 | 74.6 | 87.0 | 80.3 | 68.3 | 76.6 | 72.2 | 63.5 | 77.4 | 69.8 | 75.2 |
| GenPRM-7B (Maj@8) | 69.6 | 96.9 | 81.0 | 80.5 | 91.6 | 85.7 | 74.0 | 83.5 | 78.4 | 70.0 | 85.1 | 76.8 | 80.5 |
| PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) | PRMs (14-72B) |
| Dyve-14B | - | - | 68.5 | - | - | 58.3 | - | - | 49.0 | - | - | 47.2 | 55.8 |
| Qwen2.5-Math-PRM-72B | 78.7 | 97.9 | 87.3 | 74.2 | 88.2 | 80.6 | 67.9 | 82.0 | 74.3 | 64.8 | 78.8 | 71.1 | 78.3 |
| GenPRM-32B (Pass@1) | 73.1 | 96.4 | 83.1 | 79.4 | 84.1 | 81.7 | 73.4 | 72.2 | 72.8 | 70.3 | 75.5 | 72.8 | 77.6 |
| GenPRM-32B (Maj@8) | 74.9 | 98.5 | 85.1 | 84.0 | 88.7 | 86.3 | 79.0 | 78.8 | 78.9 | 76.3 | 84.2 | 80.1 | 82.6 |
## C. Cases
In this section, we analyze two cases to have a better understanding of GenPRM. The case in Figure 5 shows that the code execution feedback can correct the mistakes in CoT and enhance the process supervision abilities of GenPRM. The case in Figure 6 demonstrates that GenPRM provides accurate process supervision with CoT reasoning only.
Table 7: Evaluation results of GenPRM with different sizes on ProcessBench.
| Model Size | Metric | GSM8K | MATH | OlympiadBench | Omni-MATH | Avg. |
|--------------|----------|---------|--------|-----------------|-------------|--------|
| 1.5B | Pass@1 | 52.8 | 66.6 | 55.1 | 54.5 | 57.3 |
| 1.5B | Maj@8 | 81 | 74.4 | 65.3 | 62.5 | 63.4 |
| 7B | Pass@1 | 78.7 | 80.3 | 72.2 | 69.8 | 75.2 |
| 7B | Maj@8 | 81 | 85.7 | 78.4 | 76.8 | 80.5 |
| 32B | Pass@1 | 83.1 | 81.7 | 72.8 | 72.8 | 77.6 |
| 32B | Maj@8 | 85.1 | 86.3 | 78.9 | 80.1 | 82.6 |
Table 8: Evaluation results of GenPRM with different proportions of training data on ProcessBench.
| Data Size | Metric | GSM8K | MATH | OlympiadBench | Omni-MATH | Avg. |
|-------------|----------|---------|--------|-----------------|-------------|--------|
| 25% | Pass@1 | 71.8 | 78 | 68.9 | 68.2 | 71.7 |
| 50% | Pass@1 | 78.1 | 80.6 | 71.5 | 67.6 | 74.5 |
| 100% | Pass@1 | 78.7 | 80.3 | 72.2 | 69.8 | 75.2 |
Table 9: Statistics of the output tokens of GenPRM.
| | MATH | AMC23 | AIME24 | Minerva Math |
|----------|--------|---------|----------|----------------|
| Step | 344.7 | 416.2 | 432.5 | 503.3 |
| Response | 2771.4 | 3200.2 | 4112.9 | 4877.1 |
## Problem:
Simplify $\cos 54^\circ \cos 4^\circ - \cos 36^\circ \cos 86^\circ.$ Enter your answer is a trigonometric function evaluated at an integer, such as "sin 7". (The system will assume that the angle is in degrees.)
```
```
Figure 5: The first case of GenPRM.
## Problem:
## Compute the distance between the two foci of the ellipse $x^2 + 4y^2 = 9.$
Figure 6: The second case of GenPRM.
<details>
<summary>Image 5 Details</summary>

### Visual Description
\n
## Text-Based Dialogue: Math Problem Solution Review
### Overview
The image presents a dialogue between a "User" and an "Assistant" regarding the step-by-step solution of a mathematical problem involving an ellipse. The dialogue consists of two turns, each containing a user statement, an assistant analysis, and a judgement. The problem involves finding the distance between the foci of an ellipse. The text is primarily in English, with some mathematical notation.
### Components/Axes
The image is structured as a conversation log. Each turn is delineated by "User (Turn X)" and "Assistant (Turn X)". Within each turn:
* **User Statement:** Presents a step in the solution.
* **Assistant Analysis:** Provides a detailed critique of the user's step, breaking it down into numbered points.
* **Judgement:** A visual indicator (boxed "Yes") confirming the correctness of the step.
* Mathematical notation is used throughout, including fractions, square roots, and symbols for variables (e.g., a, b, c).
### Detailed Analysis or Content Details
**Turn 1:**
* **User Statement:** "Since xยฒ + 4yยฒ = 9, we divide both sides by 9 to obtain \frac{xยฒ}{9} + \frac{4yยฒ}{9} = 1."
* **Assistant Analysis:**
1. The original equation is xยฒ + 4yยฒ = 9. Dividing by 9 is correct because the standard form of an ellipse requires the right-hand side to be 1.
2. The result \frac{xยฒ}{9} + \frac{4yยฒ}{9} = 1 is correct.
3. The semi-major axis is identified as 3 because the denominator under xยฒ is 9, so aยฒ = 9, leading to a = 3.
4. The semi-minor axis is calculated as \frac{3}{2} because the coefficient under yยฒ is 4, so bยฒ = \frac{9}{4}, leading to b = \frac{3}{2}.
* **Judgement:** \boxed{Yes}
**Turn 2:**
* **User Statement:** "The distance between the two foci of an ellipse with semi-major axis a and semi-minor axis b is given by 2c, where c = \sqrt{aยฒ - bยฒ} = \sqrt{3ยฒ - (\frac{3}{2})ยฒ} = \sqrt{9 - \frac{9}{4}} = \sqrt{\frac{27}{4}} = \frac{3\sqrt{3}}{2}."
* **Assistant Analysis:**
1. The formula for the distance between foci is 2c, where c = \sqrt{aยฒ - bยฒ}.
2. Substituting a = 3 and b = \frac{3}{2} into the formula yields c = \sqrt{3ยฒ - (\frac{3}{2})ยฒ}.
3. Calculating c = \sqrt{9 - \frac{9}{4}} = \sqrt{\frac{27}{4}} = \frac{3\sqrt{3}}{2} is correct.
4. Therefore, the distance between the foci is 2c = 2 * \frac{3\sqrt{3}}{2} = 3\sqrt{3}.
5. The final answer of 3โ3 is correct.
* **Judgement:** \boxed{Yes}
### Key Observations
Both steps presented by the user are deemed correct by the assistant. The assistant provides a thorough breakdown of each step, verifying the mathematical operations and logic. The problem involves standard ellipse equations and manipulations.
### Interpretation
The dialogue demonstrates a structured approach to solving a mathematical problem. The assistant's analysis highlights the importance of understanding the underlying principles (e.g., the standard form of an ellipse equation) and verifying each step. The use of "boxed Yes" provides a clear visual confirmation of correctness, which is useful for learning and self-assessment. The problem itself is a standard application of ellipse properties, and the solution is straightforward once the correct formulas and procedures are applied. The assistant's detailed explanations are geared towards reinforcing understanding rather than simply providing answers. The consistent positive feedback suggests the user is on the right track and reinforces correct problem-solving techniques.
</details>