2505.14674v1
Model: gemini-2.0-flash
## Reward Reasoning Model
Jiaxin Guo ∗ 1 , 2 Zewen Chi ∗ 1 Li Dong ∗ 1
Qingxiu Dong 1 , 3
Xun Wu 1 Shaohan Huang 1 Furu Wei 1 â‹„
1 Microsoft Research
2 Tsinghua University https://aka.ms/GeneralAI
## Abstract
Reward models play a critical role in guiding large language models toward outputs that align with human expectations. However, an open challenge remains in effectively utilizing test-time compute to enhance reward model performance. In this work, we introduce Reward Reasoning Models (RRMs), which are specifically designed to execute a deliberate reasoning process before generating final rewards. Through chain-of-thought reasoning, RRMs leverage additional test-time compute for complex queries where appropriate rewards are not immediately apparent. To develop RRMs, we implement a reinforcement learning framework that fosters selfevolved reward reasoning capabilities without requiring explicit reasoning traces as training data. Experimental results demonstrate that RRMs achieve superior performance on reward modeling benchmarks across diverse domains. Notably, we show that RRMs can adaptively exploit test-time compute to further improve reward accuracy. The pretrained reward reasoning models are available at https: //huggingface.co/Reward-Reasoning .
Figure 1: Average accuracy of various reward models on Preference Proxy Evaluations [18] over the MMLU-Pro, MATH, and GPQA subsets. The proposed reward reasoning model (RRM) outperforms previous reward models across model sizes. We also conduct reinforcement learning on unlabeled data, using RRM as the reward model. Even without ground-truth answers, reinforcement learning with RRM achieves significant improvements on GPQA, which evaluates general-domain reasoning.
<details>
<summary>Image 1 Details</summary>

### Visual Description
## Scatter Plot and Bar Chart: Frontier Reward Modeling Performance and Reinforcement Learning Results
### Overview
The image presents two charts. The first is a scatter plot comparing the average accuracy of various language models against the number of parameters. The second is a bar chart comparing the accuracy of a model before and after post-training with RRM on unlabeled data.
### Components/Axes
**Scatter Plot (Left)**
* **Title:** Frontier Reward Modeling Performance
* **X-axis:** Number of Parameters (B)
* Scale: 3B, 7B, 70B, 400B (Logarithmic scale implied)
* **Y-axis:** Average Accuracy (%)
* Scale: 55%, 60%, 65%, 70%, 75%, 80%
* **Data Points:** Represent different language models.
* RRM-32B (Star symbol)
* RRM-7B (Star symbol)
* Meta-J1-Llama-70B (Circle)
* Athene-RM-70B (Circle)
* Llama-3.1-70B-Instruct (Circle)
* Meta-J1-Llama-8B (Circle)
* InternLM2-20B-Reward (Circle)
* Nemotron-4-340B-Reward (Circle)
* Armo-8B-v0.1 (Circle)
* DeepSeek-GRM-27B (Circle)
* Llama-3.1-8B-Instruct (Circle)
* **Highlighted Region:** An area shaded in light orange, encompassing the top-left portion of the plot.
**Bar Chart (Right)**
* **Title:** Reinforcement Learning with RRM (Ours) on Unlabeled Data
* **Y-axis:** Accuracy (%)
* Scale: 20%, 30%, 40%
* **X-axis:** GPQA (Single category)
* **Bars:**
* Blue: R1-Distill-Qwen-7B
* Green: Post-trained with RRM
### Detailed Analysis
**Scatter Plot**
* **RRM-32B:** Located at approximately (50B, 83%).
* **RRM-7B:** Located at approximately (7B, 73%).
* **Meta-J1-Llama-70B:** Located at approximately (70B, 72%).
* **Athene-RM-70B:** Located at approximately (70B, 70%).
* **Llama-3.1-70B-Instruct:** Located at approximately (70B, 69%).
* **Meta-J1-Llama-8B:** Located at approximately (8B, 67%).
* **InternLM2-20B-Reward:** Located at approximately (20B, 66%).
* **Nemotron-4-340B-Reward:** Located at approximately (340B, 64%).
* **Armo-8B-v0.1:** Located at approximately (8B, 64%).
* **DeepSeek-GRM-27B:** Located at approximately (27B, 64%).
* **Llama-3.1-8B-Instruct:** Located at approximately (8B, 57%).
**Bar Chart**
* **R1-Distill-Qwen-7B (Blue):** Accuracy of 26.8%.
* **Post-trained with RRM (Green):** Accuracy of 40.9%.
### Key Observations
* The scatter plot shows a general trend of increasing accuracy with the number of parameters, but with significant variance.
* RRM-32B and RRM-7B, indicated by star symbols, appear to outperform other models with similar parameter counts.
* The bar chart demonstrates a significant improvement in accuracy after post-training with RRM.
### Interpretation
The scatter plot suggests that while increasing the number of parameters generally improves model accuracy, the architecture and training method (indicated by the RRM models) play a crucial role in achieving higher performance. The shaded region might represent a target performance area. The bar chart clearly shows the effectiveness of RRM in improving the accuracy of a language model through post-training on unlabeled data. The increase from 26.8% to 40.9% indicates a substantial performance gain.
</details>
∗ Equal contribution. ⋄ Corresponding author.
3 Peking University
## 1 Introduction
Large language models (LLMs) such as GPTs [9, 1] have significantly transformed the field of artificial intelligence. In recent years, the development paradigm of LLMs has evolved from primarily scaling pre-training resources to emphasizing post-training techniques, driven by the dual imperatives of aligning models with human preferences [45] and enhancing specific capabilities like reasoning [6, 56]. This shift reflects a growing recognition that model performance depends not only on scale but also on sophisticated methods to refine model behavior after initial training.
Reinforcement learning has emerged as a fundamental approach in LLM post-training, leveraging supervision signals from either human feedback (RLHF) or verifiable rewards (RLVR) [45, 15, 19, 33, 22]. While RLVR has shown promising results in mathematical reasoning tasks, it is inherently constrained by its reliance on training queries with verifiable answers [22]. This requirement substantially limits RLVR's application to large-scale training on general-domain queries where verification is often intractable [16, 29, 58]. In contrast, RLHF typically employs a reward model as a proxy for human preference, enabling more extensive application across diverse domains [7, 44]. Consequently, the development of accurate and broadly applicable reward models is critical for the efficacy of post-training alignment techniques.
Recent work on reward models can be categorized into scalar reward models [45, 39] and generative reward models [12, 54, 60, 80]. Scalar reward models typically replace the decoding layer with a linear head to predict a single scalar value. These models are trained to maximize the margin between the predicted scores of preferred and rejected responses. Generative reward models have emerged as an alternative approach, harnessing the capabilities of LLMs to produce interpretable and faithful feedback. These models offer enhanced flexibility, enabling them to follow adaptive evaluation instructions to construct synthetic training data, thereby facilitating self-improvement through iterative refinement [21, 78].
Despite the widespread application of current reward models, it remains an open challenge to effectively scale test-time compute for reward estimation. To serve as general-purpose evaluators, reward models should be capable of adapting to a diverse spectrum of queries, ranging from immediately obvious questions to complex tasks that require extensive reasoning [20, 50]. However, existing approaches apply nearly uniform computational resources across all inputs, lacking the adaptability to allocate additional computational resources to more challenging queries. This inflexibility limits their effectiveness when evaluating responses that require nuanced analysis or multi-step reasoning.
To address the aforementioned challenge, we propose Reward Reasoning Models (RRMs). Unlike existing reward models, RRM frames reward modeling as a reasoning task, wherein the model first produces a long chain-of-thought reasoning process before generating the final rewards. Since supervised data providing reward reasoning traces are not readily available, we develop a training framework called Reward Reasoning via Reinforcement Learning, which encourages RRMs to selfevolve their reward reasoning capabilities within a rule-based reward environment. Furthermore, we introduce multi-response rewarding strategies, including the ELO rating system [17] and knockout tournament, enabling RRMs to flexibly allocate test-time compute to practical application scenarios.
Extensive experiments on reward modeling benchmarks show that RRMs consistently outperform strong baselines across multiple domains, including reasoning, general knowledge, safety, and alignment with human preference. Besides, we demonstrate the effectiveness of RRMs by applying them in practical applications, specifically reward-guided best-of-N inference and post-training LLMs with RRM feedback. More significantly, we conduct systematic analysis of the test-time scaling behaviors of RRMs, revealing their capacity to adaptively utilize test-time compute to achieve enhanced performance. Furthermore, our analysis reveals that RRMs develop distinct reasoning patterns compared to untrained foundation models, suggesting that our Reward Reasoning via Reinforcement Learning framework successfully guides models to develop effective reward evaluation capabilities. These insights provide deeper understanding of reward reasoning processes and will likely inspire the development of future reward reasoning models within the research community.
Our main contributions are as follows:
- We propose Reward Reasoning Models (RRMs), which perform explicit reasoning before producing final rewards. This reasoning phase enables RRMs to adaptively allocate additional computational resources when evaluating responses to complex tasks. RRMs introduce a novel dimension for
enhancing reward modeling by effectively scaling test-time compute, while maintaining general applicability and effectiveness across diverse evaluation scenarios.
- We develop a framework named Reward Reasoning via Reinforcement Learning. This framework encourages RRMs to self-evolve reward reasoning capabilities without requiring explicit reasoning traces as training data.
- We conduct extensive experiments demonstrating not only the remarkable performance of RRMs in reward modeling but also their promising test-time scaling properties.
## 2 Related Work
Reward Models Reward models can be characterized along two dimensions: reward formulation and scoring scheme [44, 79]. Formulation strategies include numeric only, which assigns scalar scores to query-response pairs [45, 39, 62, 63], and generative, which produces natural language feedback from which rewards may be extracted [3, 5, 11, 12, 41, 57, 71, 75]. Scoring schemes typically follow either absolute approaches, evaluating individual query-response pairs independently [16, 20, 23, 66, 73, 74], or discriminative methods that compare candidate responses to express relative preferences [28, 35, 40, 47, 54, 59, 80].
Generative Reward Models Generative reward models (GRMs), conceptually aligned with the LLM-as-a-Judge paradigm [67, 77], offer nuanced, interpretable feedback with flexibility for both single-instance evaluation and multi-response comparison [32, 43]. This approach addresses limitations of traditional evaluation methods like ROUGE [38] and BLEU [46], which struggle with open-ended tasks requiring sophisticated judgment [51]. GRMs can support judgment across diverse tasks, including multimodal inputs [31, 35, 80], and contemporaneous work on GRMs demonstrates promising scalability in both model capacity and inference-time compute [14, 41]. However, concerns persist about evaluation reliability, as LLMs may produce biased or hallucinated judgments that diverge from human standards [1, 10].
Inference-Time Scaling Inference-time scaling dynamically adjusts computational resources during model inference based on input complexity, inspired by human adaptive reasoning [30, 55, 68]. Recent approaches include parallel scaling strategies such as multi-sampling [8] and reward modelguided aggregation [37, 55, 76], which combine multiple outputs to enhance quality. Alternative methods utilize horizon-based scaling to extend reasoning traces [64]. Advanced systems like OpenAI's o1 and DeepSeek's R1 series demonstrate spontaneous computational allocation that adjusts 'thinking horizons' in response to task complexity [22, 27]. These approaches collectively underscore the importance of inference-time adaptability in improving model performance, particularly on complex reasoning tasks.
## 3 Reward Reasoning Model
## 3.1 Input Representation
Figure 2 provides an overview of reward reasoning models (RRMs). RRMs utilize the Qwen2 [69] model architecture with a Transformer-decoder as backbone. We formulate the reward modeling task as a text completion problem, wherein RRMs take queries and corresponding responses as input, and autoregressively generate output text consisting of a thinking process followed by a final judgment. Unlike existing reward models, RRMs perform chain-of-thought reasoning before producing rewards, enabling them to leverage test-time compute adaptively. We refer to this process as reward reasoning.
Each input of RRMs contains a query and two corresponding responses. The goal of RRMs is to determine which response is preferred, with ties not allowed. We employ the system prompt from the RewardBench repository 2 , which guides the model to perform a systematic analysis of the two responses according to several evaluation criteria, including instruction fidelity, helpfulness, accuracy, harmlessness, and level of detail. The model is also explicitly instructed to avoid common biases (such as response order or length) and must justify its judgment through structured reasoning before
2 https://github.com/allenai/reward-bench
<details>
<summary>Image 2 Details</summary>

### Visual Description
## Diagram: Reward Model Architectures
### Overview
The image presents three different reward model architectures: a Scaler Reward Model, a Generative Reward Model, and a Reward Reasoning Model. Each model takes a query and response as input, but they differ in how they generate or assign rewards.
### Components/Axes
* **(a) Scaler Reward Model:**
* **Input:** "Q: 3x5=? A: 15." (Query & Response) - displayed in a light red box.
* **Model:** "Scaler Reward Model" - displayed in a light yellow box.
* **Output:** "0.92" (Scaler Reward) - displayed in a light green box.
* **(b) Generative Reward Model:**
* **Input:** "Q: 3x5=? A: 15." (Query & Response) - displayed in a light red box.
* **Model:** "Generative Reward Model" - displayed in a light yellow box.
* **Output:** "9, because..." (Reward with Justification) - displayed in a light green box.
* **(c) Reward Reasoning Model:**
* **Input:** "Q: 3x5=? A: 15. B: 16" (Query & Response) - displayed in a light red box.
* **Model:** "Reward Reasoning Model" - displayed in a light yellow box.
* **Intermediate Steps:** "Okay, so I need to... Looking back, ...", "Given that... Alternatively, ...But if ...", "Let's analyze... Wait, perhaps... Thus..." (Long Reasoning) - displayed in light blue boxes.
* **Response:** "The answer is A.", "The answer is B.", "A is better than B." - displayed in light green boxes.
* **Output:** R1 = +1, R2 = -1, Rn = +1 - displayed in a light gray box.
* **Reinforcement Learning:** An arrow indicates that the output is used for reinforcement learning, feeding back into the "Reward Reasoning Model".
### Detailed Analysis or ### Content Details
* **Scaler Reward Model:** This model directly assigns a scalar reward (0.92) to the given query and response.
* **Generative Reward Model:** This model generates a reward along with a justification ("9, because...").
* **Reward Reasoning Model:** This model involves a longer reasoning process, generating intermediate steps before arriving at a response. The responses are then assigned rewards (R1 = +1, R2 = -1, Rn = +1).
### Key Observations
* The Scaler Reward Model provides a single numerical reward.
* The Generative Reward Model provides a reward with an explanation.
* The Reward Reasoning Model breaks down the reasoning process into multiple steps and assigns rewards to individual responses.
* The Reward Reasoning Model uses reinforcement learning to improve the model.
### Interpretation
The diagram illustrates three different approaches to reward modeling. The Scaler Reward Model is the simplest, providing a direct reward. The Generative Reward Model adds a layer of interpretability by providing a justification for the reward. The Reward Reasoning Model is the most complex, simulating a reasoning process and allowing for more nuanced reward assignment through reinforcement learning. The choice of model depends on the specific application and the desired level of interpretability and control.
</details>
(c) Reward Reasoning Model
Figure 2: An overview of reward reasoning model (RRM). RRM adaptively leverages test-time compute through chain-of-though reasoning before producing rewards.
outputting its final decision in the format '\boxed{Assistant 1}' or '\boxed{Assistant 2}'. The detailed prompt template is provided in Appendix A.1.
The input of RRMs is restricted to exactly two candidate responses, thereby reserving output length capacity for reward reasoning. Section 3.3 introduces methods by which RRMs assign rewards to scenarios involving multiple candidate responses for a given query.
## 3.2 Model Training with Reinforcement Learning
We develop a training framework called Reward Reasoning via Reinforcement Learning to train RRMs. Unlike conventional supervised fine-tuning approaches, which relies on existing reasoning traces, our framework encourages RRMs to self-evolve their reasoning capacities within a rule-based reward environment. The reward function is defined as follows:
$$\mathcal { R } = \begin{cases} + 1 , & R M e c t s \, c o r r e c t \, r e s p o n s e & ( 1 ) \\ - 1 , & o t h e r w i s e \end{cases}$$
Note that the reward R evaluates whether RRM correctly prefers the ground-truth response, rather than scoring its own outputs. Despite the simplicity of the reward signals, such rule-based rewards can effectively supervise the policy models to develop reasoning patterns that lead to correct final judgments.
Weuse Deepseek-R1 distilled models [22] as base models, applying group relative policy optimization (GRPO) [70] for training, implemented with the verl library [53]. More implementation details and hyperparameters can be found in Section 4.1 and Appendix A.2.
## 3.3 Multi-Response Rewarding Strategies
Although the input structure of RRMs strictly accepts two candidate responses, RRMs can adaptively reward multiple responses of a specific query. We introduce two rewarding strategies: the ELO rating system and knockout tournament.
ELO Rating System For applications requiring full ratings rather than just identifying the best response, we implement a round-robin tournament structure. In this approach, each candidate is compared with all others pairwise. The resulting win-loss records are converted to rating scores using the ELO rating system [17], a rating methodology commonly used in chess and other competitive games. While this strategy can process ( n 2 ) = O ( n 2 ) pairwise comparison results, computational cost can be reduced by sampling a subset of the pairwise matchups. The resulting ratings can serve as rewards in reinforcement learning from human feedback (RLHF). Experiments demonstrate that we successfully post-train an LLM using these ratings as rewards in RLHF (See Section 4).
Knockout Tournament Inspired by the knockout tournament structure [40], we design a knockout tournament strategy for RRMs that organizes multiple candidates into a competition bracket. Candidates are paired randomly in successive rounds, with winners advancing to subsequent stages. In each pairwise comparison, RRMs determine a preferred response that will participate in the tournament in the next round. Given n candidates, this requires n -1 pairwise comparisons with O ( n ) complexity and O ( log ( n )) sequential rounds. Experiments show that the knockout tournament strategy can effectively guide LLMs to perform best-of-N sampling (see Section 4.3 and Appendix B.2).
Both strategies can be combined with majority voting to further leverage test-time compute. To integrate majority voting with the aforementioned strategies, we sample RRMs multiple times for each pairwise comparison. Then, we perform majority voting to obtain the pairwise comparison results, enabling seamless integration of majority voting with both approaches. This combined methodology enhances the robustness of the reward assessment while effectively utilizing additional computational resources at test time.
## 4 Experiments
We design our experiments that evaluate RRMs on both reward modeling benchmarks and practical applications, including reward-guided inference and LLM post-training. Additionally, we analyze how RRMs utilize additional test-time compute to achieve better performance and examine the reasoning patterns exhibited by RRM across multiple domains.
## 4.1 Training Details
Training Data Training RRMs require diverse pairwise preference data that covers various capabilities and aligns with human preference. In addition to preference pairs from Skywork-Reward [39], we further synthesize preference pairs from diverse data sources. We randomly sample 80K queries from the Tülu 3 prompt dataset [33], generate two responses for each using Deepseek-R1-Distill-Qwen1.5B [22], and annotate preference labels with GPT-4o [26]. Besides, we synthesize preferences pairs using verifiable question-answer pairs from WebInstruct-verified [42], Skywork-OR1 [24], Big-Math-RL [2], and DAPO-Math [72]. We prompt Deepseek-R1 distilled 1.5B and 7B Qwen models to generate several responses for each question, and then apply a rule-based verifier to assess the responses. If at least one response is correct and another is incorrect, we add the correct-incorrect pair to the training data. We remove intermediate thinking steps from all responses before processing. The final training dataset comprises approximately 420K preferences pairs: 80K each from SkyworkReward, Tülu 3, and our-synthesized data using Tülu 3 prompts, and 180K synthesized from other sources.
RRM Training We use DeepSeek-R1-Distill-Qwen models as the base models for RRMs in all the experiments. The training hyperparameters are detailed in Appendix A.2. The RRM training framework is implemented using the verl library [53], and we train both RRM-7B and RRM-32B models on AMD Instinct MI300X Accelerators. For RRM-32B, we employ a weighted mixture of datasets with a sampling ratio of 5:1:1:1 across Skywork-Reward, Tülu-80K, our GPT-4o-labeled preference pairs, and the other synthetic data. The RRM-7B model is trained on a similar dataset mixture using a 5:1:1 ratio of Skywork-Reward, Tülu-80K, and GPT-4o-labeled preference data.
## 4.2 Evaluating Agreement with Human Preference
## 4.2.1 Setup
Benchmarks We evaluate RRMs on widely-used benchmarks for reward modeling, namely RewardBench [34] and PandaLM Test [60]. (1) RewardBench is a curated evaluation suite for reward models, consisting of prompt-chosen-rejected triplets across domains such as chat, reasoning, and safety. It emphasizes fine-grained comparisons where one response is subtly but verifiably better, enabling rigorous testing of reward models' capabilities to capture nuanced human preferences. (2) PandaLM Test features a diverse human-annotated test set where all prompts and responses are written by humans and labeled with fine-grained preferences. Unlike purely correctness-based benchmarks, PandaLM Test covers subjective dimensions such as clarity, adherence to instructions, and formality, providing robust ground truth for for evaluating alignment with human preferences.
Table 1: Evaluation results on RewardBench benchmark and PandaLM Test. Bold numbers indicate the best performance, Underlined numbers indicate the second best.
| Models | RewardBench | RewardBench | RewardBench | RewardBench | RewardBench | PandaLM Test | PandaLM Test |
|----------------------------------------|---------------|---------------|---------------|---------------|---------------|----------------|----------------|
| | Chat | Chat Hard | Safety | Reasoning | Overall | Agreement | F1 |
| Skywork-Reward-Gemma-2-27B-v0.2 [34] | 96.1 | 89.9 | 93.0 | 98.1 | 94.3 | 76.6 | 76.4 |
| JudgeLM-7B [80] | 87.3 | 43.6 | 74.5 | 48.7 | 63.5 | 65.1 | 61.9 |
| JudgeLM-33B [80] | 92.7 | 54.2 | 85.8 | 58.3 | 72.3 | 75.2 | 69.7 |
| Claude-3.5-Sonnet-20240620 [34] | 96.4 | 74.0 | 81.6 | 84.7 | 84.2 | - | - |
| DeepSeek-R1 [41, 12] | 97.1 | 73,7 | 73.3 | 95.6 | 84.9 | 78.7 | 72.5 |
| DeepSeek-GRM-27B [41] | 94.1 | 78.3 | 88.0 | 83.8 | 86.0 | - | - |
| GPT-4-0125-preview [34] | 95.3 | 74.3 | 87.6 | 86.9 | 86.0 | 66.5 | 61.8 |
| GPT-4o-0806 [34] | 96.1 | 76.1 | 86.6 | 88.1 | 86.7 | - | - |
| RM-R1-DeepSeek-Distilled-Qwen-7B [14] | 88.9 | 66.2 | 78.4 | 87.0 | 80.1 | - | - |
| RM-R1-DeepSeek-Distilled-Qwen-14B [41] | 91.3 | 91.3 | 79.4 | 95.5 | 88.9 | - | - |
| RM-R1-DeepSeek-Distilled-Qwen-32B [41] | 95.3 | 80.3 | 91.1 | 96.8 | 90.9 | - | - |
| DirectJudge-7B | 86.0 | 69.7 | 85.5 | 79.5 | 80.2 | 70.3 | 70.2 |
| DirectJudge-32B | 96.1 | 85.1 | 89.5 | 90.9 | 90.4 | 76.7 | 77.4 |
| RRM-7B | 87.7 | 70.4 | 80.7 | 90.0 | 82.2 | 72.9 | 71.1 |
| RRM-7B (voting@16) | 92.1 | 71.5 | 81.3 | 93.8 | 84.8 | 75.9 | 77.8 |
| RRM-32B | 94.7 | 81.1 | 90.7 | 98.3 | 91.2 | 78.8 | 79.0 |
| RRM-32B (voting@16) | 96.1 | 81.4 | 91.6 | 98.6 | 91.9 | 80.2 | 81.9 |
Baselines We compare RRMs with the following baselines: (1) Skywork-Reward [39], a scalar reward model that uses a regression head to output numerical preference scores without explanations or reasoning traces, (2) Production-grade LLMs , including GPT-4o [26] and Claude 3.5 Sonnet [4], which are prompted in an LLM-as-a-judge [78] manner to determine the preferred response, (3) JudgeLM [80], which is trained to generate fine-grained reward scores along with explanations, using synthetic training data generated by GPT-4 [1], (4) DeepSeek-GRM [41] and RM-R1 [14], two concurrent approaches that also incorporate a reasoning phase prior to producing rewards.
In addition to these existing baselines, we introduce (5) DirectJudge , a pairwise judging model implemented using the same training data and base models as RRMs. DirectJudge models receive the same inputs as RRMs but are trained to directly generate judgment without explicit reasoning.
## 4.2.2 Results
Table 1 presents the evaluation results of baseline reward models and RRMs on the RewardBench benchmark and the PandaLM Test. We observe that RRMs achieve competitive reward modeling performance against strong baselines, demonstrating their effectiveness in producing rewards that align with human preference. Notably, RRM-32B attains an accuracy of 98.6 in the reasoning category of RewardBench. Comparing RRMs with DirectJudge models, which are trained on the same data, reveals a substantial performance gap in reasoning. This difference indicates that RRMs effectively leverage test-time compute, thereby enhancing performance on complex queries that benefit from deliberate reasoning processes.
## 4.3 Evaluating Reward-Guided Best-of-N Inference
## 4.3.1 Setup
Preference Proxy Evaluations Preference Proxy Evaluations (PPE) [18] is a benchmark designed to evaluate reward models through proxy tasks. Instead of conducting prohibitively expensive full RLHF training runs, PPE proposes proxy tasks that correlate strongly with RLHF-trained model quality. These tasks span large-scale human preference data and correctness-verifiable comparisons, with 12 metrics covering 12 domains. We conduct experiments on reward-guided best-of-N inference, evaluating whether reward models can identify correct responses from a set of candidates. Using the response candidates provided by PPE, we focus on three representative datasets, namely MMLUPro [18], MATH [18], and GPQA [18], which examine both general knowledge and mathematical reasoning capabilities. Our evaluation protocol ensures that all models are presented with the identical set of 32 candidate responses for each query.
Table 2: Evaluation results on reward-guided best-of-N inference. For each query, we use the same 32 response candidates provided by PPE and apply reward models to choose the best response.
| Models | MMLU-Pro | MATH | GPQA | Overall |
|---------------------------------|------------|--------|--------|-----------|
| Skywork-Reward-Gemma-2-27B-v0.2 | 67 | 56.3 | 44 | 55.8 |
| GPT-4o-0806 | 64.8 | 56.9 | 46.3 | 56 |
| RRM-7B | 69.1 | 82 | 49.2 | 66.8 |
| RRM-7B (voting@5) | 69.4 | 86.1 | 49 | 68.2 |
| RRM-32B | 81.3 | 89.8 | 61.1 | 77.4 |
| RRM-32B (voting@5) | 83 | 91.8 | 64.3 | 79.7 |
Table 3: Evaluation results on binary preference classification following the protocol from Frick et al. [18]. For each benchmark, we report accuracy over a single random permutation of paired responses.
| Models | MMLU-Pro | MATH | GPQA | Overall |
|--------------------------------------------|------------|--------|--------|-----------|
| Skywork-Reward-Gemma-2-27B [65] | 55 | 46.2 | 44.7 | 48.6 |
| Gemma-2-27B [41] | 66.2 | 66.4 | 51.9 | 61.5 |
| DeepSeek-GRM-27B (voting@32) [41] | 65.5 | 69.4 | 56 | 63.6 |
| DeepSeek-GRM-27B (MetaRM) (voting@32) [41] | 68.1 | 70 | 56.9 | 65 |
| Llama-3.1-8B-Instruct [65] | 56.3 | 62.9 | 51.4 | 56.9 |
| Llama-3.1-70B-Instruct [65] | 72.1 | 73.1 | 61.2 | 68.8 |
| J1-Llama-8B (SC@32) [65] | 67.5 | 76.6 | 55.7 | 66.7 |
| J1-Llama-70B (SC@32) [65] | 79.9 | 88.1 | 66.5 | 78.2 |
| RRM-7B | 66.5 | 88 | 57.9 | 70.3 |
| RRM-7B (voting@5) | 68.3 | 90.5 | 58.3 | 72.4 |
| RRM-32B | 80.5 | 94.3 | 67.4 | 80.7 |
| RRM-32B (voting@5) | 81.3 | 95.4 | 68.4 | 81.7 |
Baselines For the first experiment, we employ the knockout tournament rewarding strategy to identify the best-of-N responses. We compare our method against several strong baselines, including Skywork-Reward-Gemma-2 [54] and GPT-4o [26]. The prompt template for GPT-4o is detailed in Appendix A.1.
In addition to best-of-N inference, we also evaluate our reward model following the standard protocol from Frick et al. [18]. For this evaluation, we compare established baselines including J1-Llama [65], DeepSeek-GRM [41], Skywork-Reward-Gemma-2 [39], and various representative reward models from recent literature. Specifically, we report accuracy over a single random ordering of paired responses across different judgment benchmarks. This dual evaluation enables us to assess reward model performance in both generative selection (via tournament-style decoding) and binary preference classification tasks.
## 4.3.2 Results
Table 2 presents the evaluation results on reward-guided best-of-N inference. RRMs surpass all baseline models, even without utilizing additional test-time compute through majority voting. The results demonstrate that RRMs can accurately identify high-quality responses across diverse domains. Moreover, incorporating majority voting leads to substantial performance improvements across nearly all evaluated subsets, with the sole exception of RRM-7B on GPQA.
To further analyze the capabilities of RRMs across different domains, we provide detailed results on each subset of the MMLU-Pro and GPQA benchmarks. As illustrated in Appendix B.1, we compare RRMs against Skywork-Reward-Gemma-2-27B-v0.2 on each individual domain. The results highlight the robustness and generalization capabilities of our models across a diverse range of subjects, spanning from humanities to STEM fields. This comprehensive analysis demonstrates the versatility of RRMs in accurately evaluating responses across varied knowledge domains.
Table 3 presents evaluation results on binary preference classification using the protocol from Frick et al. [18]. RRMs maintain strong performance across all three benchmarks, consistently outperforming baseline reward models and instruction-tuned LLMs. Notably, RRM-32B achieves state-of-the-art
Figure 3: GPQA accuracy of using RRM for RL post-training.
<details>
<summary>Image 3 Details</summary>

### Visual Description
## Line Chart: GPQA vs RL Training Steps
### Overview
The image is a line chart showing the relationship between GPQA (General Purpose Question Answering) performance and the number of RL (Reinforcement Learning) training steps. The chart indicates an upward trend, suggesting that GPQA performance generally improves with more RL training steps.
### Components/Axes
* **X-axis:** RL training steps, with markers at 0, 200, 400, 600, and 800.
* **Y-axis:** GPQA, with markers at 30, 35, and 40.
* **Line:** A single line in a salmon color represents the GPQA performance at different RL training steps.
### Detailed Analysis
The line starts at approximately 27 GPQA at 0 RL training steps. It then rises sharply to approximately 37 GPQA at 200 RL training steps. The line then decreases slightly to approximately 36 GPQA at 400 RL training steps. The line then increases to approximately 38 GPQA at 600 RL training steps. Finally, the line increases to approximately 40 GPQA at 800 RL training steps.
* **Data Points:**
* (0, 27)
* (200, 37)
* (400, 36)
* (600, 38)
* (800, 40)
### Key Observations
* There is a significant initial increase in GPQA performance from 0 to 200 RL training steps.
* The performance plateaus and even slightly decreases between 200 and 400 steps.
* The performance continues to increase from 400 to 800 steps, but at a slower rate than the initial increase.
### Interpretation
The chart suggests that RL training positively impacts GPQA performance, especially in the early stages of training. The initial rapid increase indicates that the model quickly learns to improve its question-answering abilities with RL. The plateau and subsequent slower increase suggest diminishing returns as the model is trained further. The data implies that there is a benefit to using RL training to improve GPQA performance, but the optimal number of training steps may depend on the specific task and model architecture.
</details>
Figure 4: MMLU-Pro accuracy of using RRM for RL post-training.
<details>
<summary>Image 4 Details</summary>

### Visual Description
## Line Chart: MMLU-Pro vs RL Training Steps
### Overview
The image is a line chart showing the relationship between MMLU-Pro score and RL training steps. The x-axis represents RL training steps, and the y-axis represents the MMLU-Pro score. The chart shows an upward trend, indicating that the MMLU-Pro score generally increases with more RL training steps.
### Components/Axes
* **X-axis:** RL training steps, with markers at 0, 200, 400, 600, and 800.
* **Y-axis:** MMLU-Pro, with markers at 50, 52, 54, and 56.
* **Line:** A single line in light red color represents the data.
### Detailed Analysis
The line starts at approximately (0, 48.5) and increases rapidly to (200, 54.8). It then plateaus slightly, reaching (400, 55.2) and (600, 55.2) before increasing again to (800, 56.8).
* **RL training steps = 0:** MMLU-Pro ≈ 48.5
* **RL training steps = 200:** MMLU-Pro ≈ 54.8
* **RL training steps = 400:** MMLU-Pro ≈ 55.2
* **RL training steps = 600:** MMLU-Pro ≈ 55.2
* **RL training steps = 800:** MMLU-Pro ≈ 56.8
### Key Observations
* The most significant increase in MMLU-Pro score occurs between 0 and 200 RL training steps.
* The MMLU-Pro score plateaus between 400 and 600 RL training steps.
* The highest MMLU-Pro score is achieved at 800 RL training steps.
### Interpretation
The chart suggests that increasing RL training steps generally improves the MMLU-Pro score. However, the rate of improvement decreases as the number of training steps increases. The plateau between 400 and 600 steps indicates diminishing returns, suggesting that there may be an optimal number of training steps beyond which further training provides minimal benefit. The data demonstrates a positive correlation between RL training and MMLU-Pro performance, but also highlights the importance of considering the efficiency of training.
</details>
accuracy on MMLU-Pro, MATH, and GPQA, even when compared against significantly larger models such as J1-Llama-70B. Furthermore, incorporating majority voting (voting@5) further boosts performance, with RRM-32B (voting@5) reaching peak results across all benchmarks. These findings further validate the effectiveness of RRMs in classifying reason quality under diverse and challenging evaluation settings.
## 4.4 Post-Training with RRM Feedback
In addition to directly evaluating RRMs on reward model benchmarks, we further assess RRMs by post-training LLMs with reinforcement learning or direct preference optimization, supervised by the RRM-generated rewards. This approach allows the downstream performance of the post-trained LLMs to reflect the quality of the reward signals. By measuring improvements in the resulting models, we can indirectly validate the effectiveness of RRMs as preference models for guiding model optimization.
## 4.4.1 Reinforcement Learning with Unlabeled Data
We train Deepseek-R1-Distill-Qwen-7B on WebInstruct [42] queries using group relative policy optimization (GRPO) [52]. Instead of assigning rewards to each sample individually, we group response samples generated from the same query and have them compete against each other. In each group containing 8 responses, we construct 4 × 8 pairwise matches by randomly selecting 4 competitors for each response, and then obtain the pairwise preference results using RRM-32B. Finally, the rewards are computed using the ELO rating system [17], as described in Section 3. Notably, this approach utilizes only unlabeled queries without requiring any answers or reference responses.
Following the evaluation protocols established by Ma et al. [42], we evaluate the post-trained models on MMLU-Pro and GPQA using greedy decoding with a maximum response length of 8K tokens. As shown in Figure 3 and Figure 4, the downstream performance of the post-traineded models improves steadily throughout the training process. These results demonstrate that RRMs can effectively guide post-training with reinforcement learning, despite most prior work relying exclusively on scalar reward models. This underscores the practical viability of RRMs as a compelling alternative to traditional scalar reward models in post-training pipelines.
## 4.4.2 Direct Preference Optimization
To further explore the utility of RRMs in post-training pipelines, we apply Direct Preference Optimization (DPO) [49] on Qwen2.5-7B [48] using preference labels annotated by different reward models. Specifically, we construct preference datasets from Tülu [34] with 80K queries and responses, and obtain preference annotations from three different verifiers: RRM-7B, RRM-32B, and GPT-4o. Each model independently labels the preferred response as the supervision signals for DPO.
The trained models are evaluated on the Arena-Hard benchmark [36], which contains challenging instructions designed to test comprehensive model capabilities. As shown in Table 4, all post-trained models outperform the original Qwen2.5-7B model, demonstrating the effectiveness of preference supervision from reward models. Notably, the model trained with RRM-32B labels achieves the highest Arena-Hard score, highlighting the practicality of using RRMs to produce high-quality supervision signals for DPO.
Table 4: Performance of DPO post-trained Qwen2.5-7B models on Arena-Hard.
| Arena-Hard Score | Arena-Hard Score | CI |
|-----------------------------------------------------|-----------------------------------------------------|-----------------------------------------------------|
| Before Post-Training | Before Post-Training | |
| Base Model | 18.3 | (-1.61, +1.66) |
| DPO with Preference Data Annotated by Reward Models | DPO with Preference Data Annotated by Reward Models | DPO with Preference Data Annotated by Reward Models |
| GPT-4o | 51.9 | (-2.96, +2.93) |
| RRM-7B | 53.8 | (-1.72, +1.85) |
| RRM-32B | 55.4 | (-2.60, +2.67) |
## 4.5 Scaling Test-Time Compute
## 4.5.1 Parallel Scaling
We conduct parallel test-time compute scaling experiments on MATH [25] reasoning candidate responses. We use Qwen2.5-Math-7B-Instruct [70] to generate 8 candidate responses for each question, and then employ RRMs to perform reward-guided best-of-N inference. This experimental setup allows us to systematically study the scaling behaviors of RRMs under increased test-time computational resources.
Scaling Properties As illustrated in Figure 5, increasing the number of pairwise comparisons steadily improves best-of-N performance on MATH for both RRM-7B and RRM-32B. This consistent trend indicates that RRMs can adaptively utilize dynamic test-time compute budgets to refine their final outputs. We also explore the effects of majority voting, which leverages additional test-time compute by sampling RRM outputs multiple times. Table 5 compares the performance on MATH, where RRMs are prompted on each comparison pair either a single time or eight times, with the latter followed by majority voting. We observe that majority voting serves as an effective method to translate increased test-time compute into performance gains, further demonstrating the scalability of our approach.
Figure 5: MATH accuracy with varying number of pairwise comparisons.
<details>
<summary>Image 5 Details</summary>

### Visual Description
## Line Chart: MATH Performance vs. Number of Pairs
### Overview
The image is a line chart comparing the performance of two models, RRM-7B and RRM-32B, on the MATH dataset as the number of training pairs increases. The x-axis represents the number of pairs (log scale), and the y-axis represents the MATH score.
### Components/Axes
* **X-axis:** Number of pairs (log scale). Values: 10<sup>1</sup>, 10<sup>2</sup>.
* **Y-axis:** MATH score. Values range from approximately 87 to 90.5.
* **Legend:** Located in the center-right of the chart.
* RRM-7B (coral color)
* RRM-32B (blue color)
### Detailed Analysis
* **RRM-7B (coral):**
* Trend: The line slopes upward, indicating increasing MATH score with more training pairs.
* Data Points:
* At 10<sup>1</sup> pairs, the MATH score is approximately 87.2.
* At 10<sup>2</sup> pairs, the MATH score is approximately 88.8.
* **RRM-32B (blue):**
* Trend: The line slopes upward, indicating increasing MATH score with more training pairs.
* Data Points:
* At 10<sup>1</sup> pairs, the MATH score is approximately 88.6.
* At 10<sup>2</sup> pairs, the MATH score is approximately 90.4.
### Key Observations
* RRM-32B consistently outperforms RRM-7B at both data points.
* Both models show improvement in MATH score as the number of training pairs increases.
* The improvement in MATH score appears to diminish as the number of pairs increases from 10<sup>1</sup> to 10<sup>2</sup>, suggesting diminishing returns.
### Interpretation
The chart demonstrates that increasing the number of training pairs improves the performance of both RRM-7B and RRM-32B models on the MATH dataset. The RRM-32B model, which presumably has a larger capacity, consistently achieves higher MATH scores than the RRM-7B model. The diminishing returns observed suggest that there may be a point beyond which adding more training pairs yields only marginal improvements in performance. This could be due to the models reaching their capacity or the dataset having inherent limitations.
</details>
Comparing Rewarding Strategies Table 5 compares the scoring strategies, specifically using RRMs to evaluate candidates through either knockout tournament or ELO rating systems. Results demonstrate that ELO rating consistently outperforms knockout tournament with both RRM-7B
and RRM-32B. Nonetheless, the knockout tournament yields only slightly lower performance while requiring fewer computational resources-only O ( n ) comparisons. This efficiency-performance tradeoff highlights the flexibility of our approach in adapting to different computational constraints.
Table 5: Comparison of scoring strategies using RRM verifiers. ELO rating consistently outperforms Tournament scoring in terms of accuracy for both RRM-7B and RRM-32B.
| | RRM-7B | RRM-7B | RRM-32B | RRM-32B |
|-----------------|----------|----------|-----------|-----------|
| Majority Voting | No | Yes | No | Yes |
| Tournament | 88.2 | 88.7 | 90.0 | 90.4 |
| ELO rating | 88.5 | 88.8 | 90.3 | 90.5 |
## 4.5.2 Sequential Scaling
We study the impact of enabling longer chains of thought [64] before finalizing an answer. We evaluate RRMs on RewardBench, where we control the thinking budgets by setting a maximum token limit. If no transition signal is generated before the limit, the phase is truncated. We also set a small post-thinking budget to prevent compute hacking, i.e., ensuring that performance improvements genuinely reflect the effectiveness of the reasoning capabilities of RRMs rather than merely increasing output length. The detailed design of the post-thinking budget can be found in Appendix C.
Results Experiments on 7B, 14B, and 32B RRMs show that longer thinking horizons consistently improve output accuracy across all model sizes (Figure 6). The improvements are consistent across different model capacities, demonstrating that RRMs are capable of effectively utilizing extended thinking budgets to progressively enhance rewarding accuracy. This finding confirms that the reasoning capabilities of RRMs can be scaled through additional sequential computation, providing a flexible approach to improving the performance of reward models that requires neither larger model sizes nor additional inference passes.
## 4.6 Scaling RRM Training Compute
We investigate how model size and training duration affect the performance of RRMs, exploring the scaling properties of our reward reasoning approach across different compute dimensions. Figure 6 compares RRMs with model sizes of 7B, 14B, and 32B on RewardBench, showing consistent performance gains with increased model size.
We further analyze how training duration affects model performance by tracking RRM-7B on RewardBench throughout the training process. Figure 7 illustrates the performance trajectory across different evaluation domains. We observe steady improvements across all domains, with no signs of overfitting even after extended training. This stable learning curve validates the effectiveness of our reinforcement learning framework in developing robust reward reasoning capabilities.
Figure 6: Results on RewardBench varying thinking budgets.
<details>
<summary>Image 6 Details</summary>

### Visual Description
## Chart: RewardBench vs. Thinking Budget for Different Model Sizes
### Overview
The image is a line chart comparing the performance of three different model sizes (7B, 14B, and 32B) on the RewardBench benchmark as a function of the "thinking budget" (measured in tokens). The chart shows how performance improves with increasing thinking budget for each model size.
### Components/Axes
* **Y-axis:** RewardBench (%), ranging from 75% to 90%.
* **X-axis:** Thinking budget (tokens), with values 1e3, 2e3, 4e3, and 8e3.
* **Legend:** Located on the right side of the chart.
* Blue circle: 32B
* Orange triangle: 14B
* Green square: 7B
### Detailed Analysis
* **32B (Blue):** The line starts at approximately 85% at 1e3 tokens, increases to approximately 89% at 2e3 tokens, reaches approximately 90% at 4e3 tokens, and remains around 90% at 8e3 tokens.
* **14B (Orange):** The line starts at approximately 83% at 1e3 tokens, increases to approximately 88% at 2e3 tokens, reaches approximately 89% at 4e3 tokens, and remains around 89% at 8e3 tokens.
* **7B (Green):** The line starts at approximately 75% at 1e3 tokens, increases to approximately 81% at 2e3 tokens, reaches approximately 82% at 4e3 tokens, and remains around 82% at 8e3 tokens.
### Key Observations
* The 32B model consistently outperforms the 14B and 7B models across all thinking budget values.
* The 14B model consistently outperforms the 7B model across all thinking budget values.
* All models show diminishing returns in performance as the thinking budget increases beyond 2e3 tokens. The performance increase from 1e3 to 2e3 is more significant than the increase from 4e3 to 8e3.
### Interpretation
The chart demonstrates that larger models (32B) achieve higher performance on the RewardBench benchmark compared to smaller models (14B and 7B). Increasing the thinking budget (number of tokens) generally improves performance, but the gains diminish as the budget increases. This suggests that there is a point of diminishing returns where increasing the thinking budget provides less significant improvements in performance. The data indicates that model size is a significant factor in achieving higher RewardBench scores, and that increasing computational resources (thinking budget) can improve performance, but only up to a certain point.
</details>
Figure 7: Results on RewardBench throughout RRM7B training.
<details>
<summary>Image 7 Details</summary>

### Visual Description
## Line Chart: RewardBench Performance vs. Training Steps
### Overview
The image is a line chart comparing the performance of different categories (Chat, Chat Hard, Safety, Reasoning, and Overall) on the RewardBench benchmark as a function of training steps. The x-axis represents training steps, ranging from 0 to 600. The y-axis represents the RewardBench score in percentage, ranging from 60% to 90%.
### Components/Axes
* **X-axis:** Training steps, ranging from 0 to 600 in increments of 200.
* **Y-axis:** RewardBench (%), ranging from 60% to 90% in increments of 10%.
* **Legend (Top-Right):**
* Blue: Chat
* Brown: Chat Hard
* Green: Safety
* Orange: Reasoning
* Pink: Overall
### Detailed Analysis
* **Chat (Blue):** The line starts at approximately 80% and increases to around 88% by 200 training steps. It fluctuates between 85% and 92% from 200 to 600 training steps.
* **Chat Hard (Brown):** The line starts at approximately 60% and increases to around 65% by 200 training steps. It fluctuates between 62% and 68% from 200 to 600 training steps.
* **Safety (Green):** The line starts at approximately 75% and increases to around 78% by 200 training steps. It fluctuates between 77% and 81% from 200 to 600 training steps.
* **Reasoning (Orange):** The line starts at approximately 80% and increases to around 88% by 200 training steps. It fluctuates between 87% and 90% from 200 to 600 training steps.
* **Overall (Pink):** The line starts at approximately 75% and increases to around 79% by 200 training steps. It fluctuates between 78% and 82% from 200 to 600 training steps.
### Key Observations
* Reasoning and Chat categories achieve the highest RewardBench scores, consistently performing above 85% after 200 training steps.
* Chat Hard consistently performs the worst, with scores fluctuating around 65%.
* Safety and Overall categories show similar performance, staying between 75% and 82%.
* All categories show an initial increase in performance during the first 200 training steps, followed by a period of fluctuation.
### Interpretation
The data suggests that the "Reasoning" and "Chat" categories are the most successful in terms of RewardBench performance, while "Chat Hard" is the least successful. The initial increase in performance across all categories indicates a learning phase during the first 200 training steps. The subsequent fluctuations suggest that the models reach a point where further training does not lead to significant improvements, or that the models are experiencing some degree of overfitting or instability. The "Overall" score likely represents an average or aggregate of the other categories, which explains its intermediate performance. The "Safety" category performs slightly better than the "Overall" category.
</details>
## 4.7 Reward Reasoning Pattern Analysis
Following Wang et al. [61] and Chen et al. [13], we analyze the reasoning patterns of RRM-32B by statistically measuring the proportion of model responses containing keywords such as 'wait' and 'alternatively'. We categorize the reasoning patterns into four categories: transition (switching perspectives or strategies), reflection (self-checking or revisiting earlier steps), comparison (evaluating multiple options), and breakdown (decomposing the problem).
As illustrated in Figure 8, compared to the Deepseek-R1-Distill-Qwen-32B model, RRM-32B demonstrates a greater overall utilization of reasoning patterns when judging the superiority of two answers, particularly in analyzing from different perspectives and conducting in-depth comparisons. In contrast, the Deepseek-R1-Distill-Qwen-32B model employs the breakdown pattern more frequently, suggesting a greater tendency to approach problems directly when making judgments, but less inclination to compare the merits of the two answers and engage in self-examination. This distinction in reasoning patterns highlights how our Reward Reasoning via Reinforcement Learning framework shapes the model's approach to evaluation tasks.
Figure 8: Reward reasoning pattern analysis results. Compared to DeepSeek-R1-Distilled-Qwen-32B, RRM-32B exhibits more transition patterns (40.63% vs. 33.73%), reflection patterns (63.28% vs. 52.75%), and comparison patterns (89.84% vs. 85.29%), but fewer direct problem decomposition (8.40% vs. 16.86%).
<details>
<summary>Image 8 Details</summary>

### Visual Description
## Bar Chart: Percentage of Examples by Category
### Overview
The image is a bar chart comparing the percentage of examples for different categories using two methods: "R1-distilled" and "RRM". The categories are "Transition", "Reflection", "Comparison", and "Breakdown". The chart displays the percentage of examples on the y-axis, ranging from 0% to 80%.
### Components/Axes
* **X-axis:** Categories: Transition, Reflection, Comparison, Breakdown.
* **Y-axis:** Percentage of Examples (%), ranging from 0 to 80, with gridlines at intervals of 20.
* **Legend:** Located at the top of the chart.
* R1-distilled: Represented by hatched bars.
* RRM: Represented by solid color bars.
### Detailed Analysis
* **Transition:**
* R1-distilled (hatched): Approximately 34%.
* RRM (yellow): Approximately 41%.
* **Reflection:**
* R1-distilled (hatched): Approximately 53%.
* RRM (light blue): Approximately 63%.
* **Comparison:**
* R1-distilled (hatched): Approximately 84%.
* RRM (orange): Approximately 89%.
* **Breakdown:**
* R1-distilled (hatched): Approximately 17%.
* RRM (green): Approximately 9%.
### Key Observations
* For all categories, the RRM method generally shows a higher percentage of examples compared to the R1-distilled method, except for the "Breakdown" category.
* The "Comparison" category has the highest percentage of examples for both methods.
* The "Breakdown" category has the lowest percentage of examples for both methods.
### Interpretation
The bar chart compares the performance of two methods, "R1-distilled" and "RRM", across four categories: "Transition", "Reflection", "Comparison", and "Breakdown". The data suggests that the RRM method generally performs better than the R1-distilled method, as it yields a higher percentage of examples in most categories. However, the R1-distilled method shows a higher percentage in the "Breakdown" category. The "Comparison" category stands out as having the highest percentage of examples for both methods, indicating that both methods are particularly effective in this area. Conversely, both methods struggle with the "Breakdown" category, as indicated by the low percentages.
</details>
Table 6 presents case studies by comparing the outputs of Deepseek-R1-Distill-Qwen-32B and RRM-32B. Within the responses to a single example, RRM-32B exhibits a stronger propensity to iteratively apply reasoning patterns for more thorough and in-depth comparisons. The reasoning process of the Deepseek-R1-Distill-Qwen-32B model tends to be simple and monolithic-involving separate analyses of the two answers followed by a summary. Conversely, the thinking process of RRM-32B involves multiple iterations of self-reflection, comparative analysis, and consideration of various perspectives, incorporating more detailed analysis and interpretation of the instructions. Consequently, RRM-32B demonstrates an enhanced ability to follow instructions and provide more accurate comparative judgments. This iterative approach to reasoning appears to be a key factor in the superior performance of RRMs across diverse evaluation tasks.
Table 6: Case study of reasoning patterns of RRM-32B and DeepSeek-R1-Distilled-Qwen-32B.
| DeepSeek-R1-Distilled-Qwen-32B | RRM-32B |
|---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| Okay, I need to evaluate which assistant provided a bet- ter response to the given problem...Looking at the two responses: Assistant 1's code starts by... Assistant 2's solution uses a bitwise AND approach.... In conclusion, while Assistant 1's code has some issues, Assistant 2's solution is fundamentally incorrect because the bitwise approach doesn't apply to powers of three. Therefore, Assistant 1's response is better, despite its flaws, because it attempts to handle the problem correctly, even if it's not perfectly implemented. </think> \boxed{Assistant 1} | Okay, so I need to figure out which assistant is better for the given problem... Looking at Assistant 1's response... Now, looking at Assistant 2's response... Wait, that's a problem.... But then, what about the code from Assistant 1?... Let me test it with some examples... Wait, perhaps... But neither assistant provided that solution. So, given the two options, which is better?... Alternatively, perhaps I'm missing something. Let me think again. Wait, the problem requires a solution without loops or recursion... Alternatively... Looking back, Assistant 1's code is ... Assistant 2's code is ... Alternatively, perhaps I'm missing something. Let me think again. Wait, ... Wait, the user's instruction says: "Do NOT say both / neither are good." So I have to choose between Assistant 1 or 2. Given that, I think Assistant 2's code is better because ... </think> The better response is... |
| Alright, let's evaluate the problem step by step... First, I look at Assistant 1's response... Now, Assistant 2's response is much simpler... Therefore, Assistant 1 correctly addressed the exponentia- tion but made a minor error in the final subtraction... </think> The correct evaluation requires...\boxed{Assistant 1} | Okay, so I need to figure out which assistant is better for the given query. ... Assistant 1 starts by ... Assistant 2, on the other hand, ... Looking at the instruction, the user wrote... However, the problem is that the instruction might be am- biguous. If ..., then Assistant 2 is correct. But if ..., then Assistant 1 is correct. Since the instruction is...So I think the intended operation is exponentiation. Therefore, Assistant 1's approach is more accurate for the likely intended problem. However, if the problem was meant to be multiplication, then Assistant 2 is correct. But given the notation, exponentiation is more probable. So blue Assistant 1's answer is better in that context. But wait, the user's instruction is... Hmm, this is a bit confusing. I need to decide based on the instruction. ... But if ... Given that, Assistant 2's interpretation is more likely cor- rect... So, considering the instruction's ambiguity, but leaning towards multiplication, Assistant 2 is better. </think> The instruction is ambiguous, but Assistant 2 correctly interpreted it as multiplication, providing the accurate re- sult. \boxed{Assistant 2} |
## 5 Conclusion
In this paper, we introduced Reward Reasoning Models (RRMs), incorporating explicit reasoning processes prior to reward assignment. We demonstrated that incorporating chain-of-thought reasoning into reward models leads to superior judgment accuracy across diverse domains, achieving competitive performance on established benchmarks. Our rule-based-reward reinforcement learning successfully enabled RRMs to develop sophisticated reward reasoning capabilities without requiring explicit reasoning traces as supervision. Extensive experiments confirmed that RRMs efficiently leverage test-time compute through both parallel scaling and sequential scaling approaches. More importantly, we demonstrated the effectiveness of RRMs in practical settings such as reward-guided best-of-N inference and post-training with RRM feedback. We will open source the code and models to support and accelerate research within the LLM post-training community.
## References
- [1] Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al. GPT-4 technical report. arXiv preprint arXiv:2303.08774 , 2023.
- [2] Alon Albalak, Duy Phung, Nathan Lile, Rafael Rafailov, Kanishk Gandhi, Louis Castricato, Anikait Singh, Chase Blagden, Violet Xiang, Dakota Mahan, and Nick Haber. Big-math: A large-scale, high-quality math dataset for reinforcement learning in language models, 2025. URL https://arxiv.org/abs/2502.17387 .
- [3] Andrei Alexandru, Antonia Calvi, Henry Broomfield, Jackson Golden, Kyle Dai, Mathias Leys, Maurice Burger, Max Bartolo, Roman Engeler, Sashank Pisupati, et al. Atla selene mini: A general purpose evaluation model. arXiv preprint arXiv:2501.17195 , 2025.
- [4] AI Anthropic. The claude 3 model family: Opus, sonnet, haiku. Claude-3 Model Card , 1:1, 2024.
- [5] Negar Arabzadeh, Siqing Huo, Nikhil Mehta, Qingyun Wu, Chi Wang, Ahmed Hassan Awadallah, Charles L. A. Clarke, and Julia Kiseleva. Assessing and verifying task utility in LLM-powered applications. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen, editors, Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing , pages 21868-21888, Miami, Florida, USA, November 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.emnlp-main.1219. URL https: //aclanthology.org/2024.emnlp-main.1219/ .
- [6] Yuntao Bai, Andy Jones, Kamal Ndousse, Amanda Askell, Anna Chen, Nova DasSarma, Dawn Drain, Stanislav Fort, Deep Ganguli, Tom Henighan, et al. Training a helpful and harmless assistant with reinforcement learning from human feedback. arXiv preprint arXiv:2204.05862 , 2022.
- [7] Yuntao Bai, Saurav Kadavath, Sandipan Kundu, Amanda Askell, Jackson Kernion, Andy Jones, Anna Chen, Anna Goldie, Azalia Mirhoseini, Cameron McKinnon, et al. Constitutional ai: Harmlessness from ai feedback. arXiv preprint arXiv:2212.08073 , 2022.
- [8] Bradley Brown, Jordan Juravsky, Ryan Ehrlich, Ronald Clark, Quoc V Le, Christopher Ré, and Azalia Mirhoseini. Large language monkeys: Scaling inference compute with repeated sampling. arXiv preprint arXiv:2407.21787 , 2024.
- [9] Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Language models are few-shot learners. Advances in neural information processing systems , 33:1877-1901, 2020.
- [10] Sébastien Bubeck, Varun Chadrasekaran, Ronen Eldan, Johannes Gehrke, Eric Horvitz, Ece Kamar, Peter Lee, Yin Tat Lee, Yuanzhi Li, Scott Lundberg, et al. Sparks of artificial general intelligence: Early experiments with GPT-4, 2023.
- [11] Maosong Cao, Alexander Lam, Haodong Duan, Hongwei Liu, Songyang Zhang, and Kai Chen. Compassjudger-1: All-in-one judge model helps model evaluation and evolution. arXiv preprint arXiv:2410.16256 , 2024.
- [12] Nuo Chen, Zhiyuan Hu, Qingyun Zou, Jiaying Wu, Qian Wang, Bryan Hooi, and Bingsheng He. Judgelrm: Large reasoning models as a judge. arXiv preprint arXiv:2504.00050 , 2025.
- [13] Runjin Chen, Zhenyu (Allen) Zhang, Junyuan Hong, Souvik Kundu, and Zhangyang Wang. Seal: Steerable reasoning calibration of large language models for free. 2025. URL https: //api.semanticscholar.org/CorpusID:277741244 .
- [14] Xiusi Chen, Gaotang Li, Ziqi Wang, Bowen Jin, Cheng Qian, Yu Wang, Hongru Wang, Yu Zhang, Denghui Zhang, Tong Zhang, et al. Rm-r1: Reward modeling as reasoning. arXiv preprint arXiv:2505.02387 , 2025.
- [15] Paul F. Christiano, Jan Leike, Tom B. Brown, Miljan Martic, Shane Legg, and Dario Amodei. Deep reinforcement learning from human preferences. In Proceedings of the 31st International Conference on Neural Information Processing Systems , NIPS'17, page 4302-4310, Red Hook, NY, USA, 2017. Curran Associates Inc. ISBN 9781510860964.
- [16] Karl Cobbe, Vineet Kosaraju, Mohammad Bavarian, Mark Chen, Heewoo Jun, Lukasz Kaiser, Matthias Plappert, Jerry Tworek, Jacob Hilton, Reiichiro Nakano, et al. Training verifiers to solve math word problems. arXiv preprint arXiv:2110.14168 , 2021.
- [17] Arpad E. Elo. The Rating of Chessplayers, Past and Present . Arco Pub., New York, 1978. ISBN 0668047216. URL http://www.amazon.com/Rating-Chess-Players-Past-Present/ dp/0668047216 .
- [18] Evan Frick, Tianle Li, Connor Chen, Wei-Lin Chiang, Anastasios Nikolas Angelopoulos, Jiantao Jiao, Banghua Zhu, Joseph E. Gonzalez, and Ion Stoica. How to evaluate reward models for RLHF. In The Thirteenth International Conference on Learning Representations , 2025. URL https://openreview.net/forum?id=cbttLtO94Q .
- [19] Jiaxuan Gao, Shusheng Xu, Wenjie Ye, Weilin Liu, Chuyi He, Wei Fu, Zhiyu Mei, Guangju Wang, and Yi Wu. On designing effective rl reward at training time for llm reasoning. arXiv preprint arXiv:2410.15115 , 2024.
- [20] Leo Gao, John Schulman, and Jacob Hilton. Scaling laws for reward model overoptimization. In Proceedings of the 40th International Conference on Machine Learning , ICML'23. JMLR.org, 2023.
- [21] Jiawei Gu, Xuhui Jiang, Zhichao Shi, Hexiang Tan, Xuehao Zhai, Chengjin Xu, Wei Li, Yinghan Shen, Shengjie Ma, Honghao Liu, et al. A survey on llm-as-a-judge. arXiv preprint arXiv:2411.15594 , 2024.
- [22] Daya Guo, Dejian Yang, Haowei Zhang, Junxiao Song, Ruoyu Zhang, Runxin Xu, Qihao Zhu, Shirong Ma, Peiyi Wang, Xiao Bi, et al. Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning. arXiv preprint arXiv:2501.12948 , 2025.
- [23] Fang Guo, Wenyu Li, Honglei Zhuang, Yun Luo, Yafu Li, Le Yan, Qi Zhu, and Yue Zhang. Mcranker: Generating diverse criteria on-the-fly to improve pointwise llm rankers. In Proceedings of the Eighteenth ACM International Conference on Web Search and Data Mining , WSDM '25, page 944-953, New York, NY, USA, 2025. Association for Computing Machinery. ISBN 9798400713293. doi: 10.1145/3701551.3703583. URL https: //doi.org/10.1145/3701551.3703583 .
- [24] Jujie He, Jiacai Liu, Chris Yuhao Liu, Rui Yan, Chaojie Wang, Peng Cheng, Xiaoyu Zhang, Fuxiang Zhang, Jiacheng Xu, Wei Shen, Siyuan Li, Liang Zeng, Tianwen Wei, Cheng Cheng, Yang Liu, and Yahui Zhou. Skywork open reasoner series. https://capricious-hydrogen-41c.notion.site/ Skywork-Open-Reaonser-Series-1d0bc9ae823a80459b46c149e4f51680 , 2025. Notion Blog.
- [25] Dan Hendrycks, Collin Burns, Saurav Kadavath, Akul Arora, Steven Basart, Eric Tang, Dawn Song, and Jacob Steinhardt. Measuring mathematical problem solving with the MATH dataset. In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2) , 2021. URL https://openreview.net/forum?id=7Bywt2mQsCe .
- [26] Aaron Hurst, Adam Lerer, Adam P Goucher, Adam Perelman, Aditya Ramesh, Aidan Clark, AJ Ostrow, Akila Welihinda, Alan Hayes, Alec Radford, et al. Gpt-4o system card. arXiv preprint arXiv:2410.21276 , 2024.
- [27] Aaron Jaech, Adam Kalai, Adam Lerer, Adam Richardson, Ahmed El-Kishky, Aiden Low, Alec Helyar, Aleksander Madry, Alex Beutel, Alex Carney, et al. Openai o1 system card. arXiv preprint arXiv:2412.16720 , 2024.
- [28] Dongfu Jiang, Xiang Ren, and Bill Yuchen Lin. LLM-blender: Ensembling large language models with pairwise ranking and generative fusion. In Anna Rogers, Jordan Boyd-Graber, and Naoaki Okazaki, editors, Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) , pages 14165-14178, Toronto, Canada, July 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023.acl-long.792. URL https://aclanthology.org/2023.acl-long.792/ .
- [29] Carlos E Jimenez, John Yang, Alexander Wettig, Shunyu Yao, Kexin Pei, Ofir Press, and Karthik R Narasimhan. SWE-bench: Can language models resolve real-world github issues? In The Twelfth International Conference on Learning Representations , 2024. URL https: //openreview.net/forum?id=VTF8yNQM66 .
- [30] Jared Kaplan, Sam McCandlish, Tom Henighan, Tom B Brown, Benjamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeffrey Wu, and Dario Amodei. Scaling laws for neural language models. arXiv preprint arXiv:2001.08361 , 2020.
- [31] Muhammad Uzair Khattak, Hanoona Rasheed, Muhammad Maaz, Salman Khan, and Fahad Shahbaz Khan. Maple: Multi-modal prompt learning. In 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) , pages 19113-19122, 2023. doi: 10.1109/CVPR52729.2023.01832.
- [32] Seungone Kim, Juyoung Suk, Shayne Longpre, Bill Yuchen Lin, Jamin Shin, Sean Welleck, Graham Neubig, Moontae Lee, Kyungjae Lee, and Minjoon Seo. Prometheus 2: An open source language model specialized in evaluating other language models. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen, editors, Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing , pages 4334-4353, Miami, Florida, USA, November 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.emnlp-main.248. URL https://aclanthology.org/2024.emnlp-main.248/ .
- [33] Nathan Lambert, Jacob Morrison, Valentina Pyatkin, Shengyi Huang, Hamish Ivison, Faeze Brahman, Lester James V Miranda, Alisa Liu, Nouha Dziri, Shane Lyu, et al. T \ " ulu 3: Pushing frontiers in open language model post-training. arXiv preprint arXiv:2411.15124 , 2024.
- [34] Nathan Lambert, Valentina Pyatkin, Jacob Morrison, LJ Miranda, Bill Yuchen Lin, Khyathi Chandu, Nouha Dziri, Sachin Kumar, Tom Zick, Yejin Choi, et al. Rewardbench: Evaluating reward models for language modeling. arXiv preprint arXiv:2403.13787 , 2024.
- [35] Junlong Li, Shichao Sun, Weizhe Yuan, Run-Ze Fan, hai zhao, and Pengfei Liu. Generative judge for evaluating alignment. In The Twelfth International Conference on Learning Representations , 2024. URL https://openreview.net/forum?id=gtkFw6sZGS .
- [36] Tianle Li, Wei-Lin Chiang, Evan Frick, Lisa Dunlap, Tianhao Wu, Banghua Zhu, Joseph E. Gonzalez, and Ion Stoica. From crowdsourced data to high-quality benchmarks: Arena-hard and benchbuilder pipeline, 2024. URL https://arxiv.org/abs/2406.11939 .
- [37] Hunter Lightman, Vineet Kosaraju, Yuri Burda, Harrison Edwards, Bowen Baker, Teddy Lee, Jan Leike, John Schulman, Ilya Sutskever, and Karl Cobbe. Let's verify step by step. In The Twelfth International Conference on Learning Representations , 2024. URL https: //openreview.net/forum?id=v8L0pN6EOi .
- [38] Chin-Yew Lin. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out , pages 74-81, 2004.
- [39] Chris Yuhao Liu, Liang Zeng, Jiacai Liu, Rui Yan, Jujie He, Chaojie Wang, Shuicheng Yan, Yang Liu, and Yahui Zhou. Skywork-reward: Bag of tricks for reward modeling in llms. arXiv preprint arXiv:2410.18451 , 2024.
- [40] Yantao Liu, Zijun Yao, Rui Min, Yixin Cao, Lei Hou, and Juanzi Li. Pairwise rm: Perform best-of-n sampling with knockout tournament. arXiv preprint arXiv:2501.13007 , 2025.
- [41] Zijun Liu, Peiyi Wang, Runxin Xu, Shirong Ma, Chong Ruan, Peng Li, Yang Liu, and Yu Wu. Inference-time scaling for generalist reward modeling. arXiv preprint arXiv:2504.02495 , 2025.
- [42] Xueguang Ma, Qian Liu, Dongfu Jiang, Ge Zhang, Zejun Ma, and Wenhu Chen. General-reasoner: Advancing llm reasoning across all domains. https://github.com/ TIGER-AI-Lab/General-Reasoner/blob/main/General\_Reasoner.pdf , 2025.
- [43] Dakota Mahan, Duy Van Phung, Rafael Rafailov, Chase Blagden, Nathan Lile, Louis Castricato, Jan-Philipp Fränken, Chelsea Finn, and Alon Albalak. Generative reward models. arXiv preprint arXiv:2410.12832 , 2024.
- [44] Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Gray, John Schulman, Jacob Hilton, Fraser Kelton, Luke Miller, Maddie Simens, Amanda Askell, Peter Welinder, Paul Christiano, Jan Leike, and Ryan Lowe. Training language models to follow instructions with human feedback. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems , 2022. URL https://openreview.net/ forum?id=TG8KACxEON .
- [45] Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, et al. Training language models to follow instructions with human feedback. Advances in neural information processing systems , 35:27730-27744, 2022.
- [46] Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics , pages 311-318, 2002.
- [47] Junsoo Park, Seungyeon Jwa, Ren Meiying, Daeyoung Kim, and Sanghyuk Choi. OffsetBias: Leveraging debiased data for tuning evaluators. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen, editors, Findings of the Association for Computational Linguistics: EMNLP 2024 , pages 1043-1067, Miami, Florida, USA, November 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.findings-emnlp.57. URL https://aclanthology.org/ 2024.findings-emnlp.57/ .
- [48] Qwen, :, An Yang, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu, Chengyuan Li, Dayiheng Liu, Fei Huang, Haoran Wei, Huan Lin, Jian Yang, Jianhong Tu, Jianwei Zhang, Jianxin Yang, Jiaxi Yang, Jingren Zhou, Junyang Lin, Kai Dang, Keming Lu, Keqin Bao, Kexin Yang, Le Yu, Mei Li, Mingfeng Xue, Pei Zhang, Qin Zhu, Rui Men, Runji Lin, Tianhao Li, Tianyi Tang, Tingyu Xia, Xingzhang Ren, Xuancheng Ren, Yang Fan, Yang Su, Yichang Zhang, Yu Wan, Yuqiong Liu, Zeyu Cui, Zhenru Zhang, and Zihan Qiu. Qwen2.5 technical report, 2025. URL https://arxiv.org/abs/2412.15115 .
- [49] Rafael Rafailov, Archit Sharma, Eric Mitchell, Christopher D Manning, Stefano Ermon, and Chelsea Finn. Direct preference optimization: Your language model is secretly a reward model. In Thirty-seventh Conference on Neural Information Processing Systems , 2023. URL https://openreview.net/forum?id=HPuSIXJaa9 .
- [50] Rafael Rafailov, Yaswanth Chittepu, Ryan Park, Harshit Sikchi, Joey Hejna, W. Bradley Knox, Chelsea Finn, and Scott Niekum. Scaling laws for reward model overoptimization in direct alignment algorithms. In The Thirty-eighth Annual Conference on Neural Information Processing Systems , 2024. URL https://openreview.net/forum?id=pf4OuJyn4Q .
- [51] Natalie Schluter. The limits of automatic summarisation according to rouge. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics , pages 41-45. Association for Computational Linguistics, 2017.
- [52] Zhihong Shao, Peiyi Wang, Qihao Zhu, Runxin Xu, Junxiao Song, Xiao Bi, Haowei Zhang, Mingchuan Zhang, Y. K. Li, Y. Wu, and Daya Guo. Deepseekmath: Pushing the limits of mathematical reasoning in open language models, 2024. URL https://arxiv.org/abs/ 2402.03300 .
- [53] Guangming Sheng, Chi Zhang, Zilingfeng Ye, Xibin Wu, Wang Zhang, Ru Zhang, Yanghua Peng, Haibin Lin, and Chuan Wu. HybridFlow: A flexible and efficient RLHF framework. arXiv preprint arXiv: 2409.19256 , 2024.
- [54] Tu Shiwen, Zhao Liang, Chris Yuhao Liu, Liang Zeng, and Yang Liu. Skywork critic model series. https://huggingface.co/Skywork , September 2024. URL https://huggingface. co/Skywork .
- [55] Charlie Victor Snell, Jaehoon Lee, Kelvin Xu, and Aviral Kumar. Scaling LLM testtime compute optimally can be more effective than scaling parameters for reasoning. In The Thirteenth International Conference on Learning Representations , 2025. URL https: //openreview.net/forum?id=4FWAwZtd2n .
- [56] Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, et al. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 , 2023.
- [57] Tu Vu, Kalpesh Krishna, Salaheddin Alzubi, Chris Tar, Manaal Faruqui, and Yun-Hsuan Sung. Foundational autoraters: Taming large language models for better automatic evaluation. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen, editors, Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing , pages 17086-17105, Miami, Florida, USA, November 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024. emnlp-main.949. URL https://aclanthology.org/2024.emnlp-main.949/ .
- [58] Binghai Wang, Rui Zheng, Lu Chen, Yan Liu, Shihan Dou, Caishuang Huang, Wei Shen, Senjie Jin, Enyu Zhou, Chenyu Shi, et al. Secrets of rlhf in large language models part ii: Reward modeling. arXiv preprint arXiv:2401.06080 , 2024.
- [59] Tianlu Wang, Ilia Kulikov, Olga Golovneva, Ping Yu, Weizhe Yuan, Jane Dwivedi-Yu, Richard Yuanzhe Pang, Maryam Fazel-Zarandi, Jason Weston, and Xian Li. Self-taught evaluators. arXiv preprint arXiv:2408.02666 , 2024.
- [60] Yidong Wang, Zhuohao Yu, Wenjin Yao, Zhengran Zeng, Linyi Yang, Cunxiang Wang, Hao Chen, Chaoya Jiang, Rui Xie, Jindong Wang, Xing Xie, Wei Ye, Shikun Zhang, and Yue Zhang. Pandalm: An automatic evaluation benchmark for llm instruction tuning optimization. In ICLR , 2024. URL https://openreview.net/forum?id=5Nn2BLV7SB .
- [61] Yiping Wang, Qing Yang, Zhiyuan Zeng, Liliang Ren, Lucas Liu, Baolin Peng, Hao Cheng, Xuehai He, Kuan Wang, Jianfeng Gao, Weizhu Chen, Shuohang Wang, Simon Shaolei Du, and Yelong Shen. Reinforcement learning for reasoning in large language models with one training example. 2025. URL https://api.semanticscholar.org/CorpusID:278171513 .
- [62] Zhilin Wang, Yi Dong, Jiaqi Zeng, Virginia Adams, Makesh Narsimhan Sreedhar, Daniel Egert, Olivier Delalleau, Jane Polak Scowcroft, Neel Kant, Aidan Swope, et al. Helpsteer: Multi-attribute helpfulness dataset for steerlm. arXiv preprint arXiv:2311.09528 , 2023.
- [63] Zhilin Wang, Yi Dong, Olivier Delalleau, Jiaqi Zeng, Gerald Shen, Daniel Egert, Jimmy J. Zhang, Makesh Narsimhan Sreedhar, and Oleksii Kuchaiev. Helpsteer 2: Open-source dataset for training top-performing reward models. In The Thirty-eight Conference on Neural Information Processing Systems Datasets and Benchmarks Track , 2024. URL https: //openreview.net/forum?id=PvVKUFhaNy .
- [64] Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, brian ichter, Fei Xia, Ed H. Chi, Quoc V Le, and Denny Zhou. Chain of thought prompting elicits reasoning in large language models. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems , 2022. URL https://openreview.net/ forum?id=\_VjQlMeSB\_J .
- [65] Chenxi Whitehouse, Tianlu Wang, Ping Yu, Xian Li, Jason Weston, Ilia Kulikov, and Swarnadeep Saha. J1: Incentivizing thinking in llm-as-a-judge via reinforcement learning. arXiv preprint arXiv:2505.10320 , 2025.
- [66] Genta Indra Winata, David Anugraha, Lucky Susanto, Garry Kuwanto, and Derry Tanti Wijaya. Metametrics: Calibrating metrics for generation tasks using human preferences. In The Thirteenth International Conference on Learning Representations , 2025. URL https://openreview.net/forum?id=slO3xTt4CG .
- [67] Tianhao Wu, Weizhe Yuan, Olga Golovneva, Jing Xu, Yuandong Tian, Jiantao Jiao, Jason Weston, and Sainbayar Sukhbaatar. Meta-rewarding language models: Self-improving alignment with llm-as-a-meta-judge. arXiv preprint arXiv:2407.19594 , 2024.
- [68] Yangzhen Wu, Zhiqing Sun, Shanda Li, Sean Welleck, and Yiming Yang. Inference scaling laws: An empirical analysis of compute-optimal inference for LLM problem-solving. In The Thirteenth International Conference on Learning Representations , 2025. URL https: //openreview.net/forum?id=VNckp7JEHn .
- [69] An Yang, Baosong Yang, Binyuan Hui, Bo Zheng, Bowen Yu, Chang Zhou, Chengpeng Li, Chengyuan Li, Dayiheng Liu, Fei Huang, Guanting Dong, Haoran Wei, Huan Lin, Jialong Tang, Jialin Wang, Jian Yang, Jianhong Tu, Jianwei Zhang, Jianxin Ma, Jin Xu, Jingren Zhou, Jinze Bai, Jinzheng He, Junyang Lin, Kai Dang, Keming Lu, Keqin Chen, Kexin Yang, Mei Li, Mingfeng Xue, Na Ni, Pei Zhang, Peng Wang, Ru Peng, Rui Men, Ruize Gao, Runji Lin, Shijie Wang, Shuai Bai, Sinan Tan, Tianhang Zhu, Tianhao Li, Tianyu Liu, Wenbin Ge, Xiaodong Deng, Xiaohuan Zhou, Xingzhang Ren, Xinyu Zhang, Xipin Wei, Xuancheng Ren, Yang Fan, Yang Yao, Yichang Zhang, Yu Wan, Yunfei Chu, Yuqiong Liu, Zeyu Cui, Zhenru Zhang, and Zhihao Fan. Qwen2 technical report. arXiv preprint arXiv:2407.10671 , 2024.
- [70] An Yang, Beichen Zhang, Binyuan Hui, Bofei Gao, Bowen Yu, Chengpeng Li, Dayiheng Liu, Jianhong Tu, Jingren Zhou, Junyang Lin, Keming Lu, Mingfeng Xue, Runji Lin, Tianyu Liu, Xingzhang Ren, and Zhenru Zhang. Qwen2.5-math technical report: Toward mathematical expert model via self-improvement. arXiv preprint arXiv:2409.12122 , 2024.
- [71] Ziyi Ye, Xiangsheng Li, Qiuchi Li, Qingyao Ai, Yujia Zhou, Wei Shen, Dong Yan, and Yiqun LIU. Learning LLM-as-a-judge for preference alignment. In The Thirteenth International Conference on Learning Representations , 2025. URL https://openreview.net/forum? id=HZVIQE1MsJ .
- [72] Qiying Yu, Zheng Zhang, Ruofei Zhu, Yufeng Yuan, Xiaochen Zuo, Yu Yue, Tiantian Fan, Gaohong Liu, Lingjun Liu, Xin Liu, Haibin Lin, Zhiqi Lin, Bole Ma, Guangming Sheng, Yuxuan Tong, Chi Zhang, Mofan Zhang, Wang Zhang, Hang Zhu, Jinhua Zhu, Jiaze Chen, Jiangjie Chen, Chengyi Wang, Hongli Yu, Weinan Dai, Yuxuan Song, Xiangpeng Wei, Hao Zhou, Jingjing Liu, Wei-Ying Ma, Ya-Qin Zhang, Lin Yan, Mu Qiao, Yonghui Wu, and Mingxuan Wang. Dapo: An open-source llm reinforcement learning system at scale, 2025. URL https://arxiv.org/abs/2503.14476 .
- [73] Yue Yu, Zhengxing Chen, Aston Zhang, Liang Tan, Chenguang Zhu, Richard Yuanzhe Pang, Yundi Qian, Xuewei Wang, Suchin Gururangan, Chao Zhang, Melanie Kambadur, Dhruv Mahajan, and Rui Hou. Self-generated critiques boost reward modeling for language models. In Luis Chiruzzo, Alan Ritter, and Lu Wang, editors, Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers) , pages 11499-11514, Albuquerque, New Mexico, April 2025. Association for Computational Linguistics. ISBN 979-8-89176-189-6. URL https://aclanthology.org/2025.naacl-long.573/ .
- [74] Weizhe Yuan, Richard Yuanzhe Pang, Kyunghyun Cho, Xian Li, Sainbayar Sukhbaatar, Jing Xu, and Jason Weston. Self-rewarding language models. In Proceedings of the 41st International Conference on Machine Learning , ICML'24. JMLR.org, 2024.
- [75] Lunjun Zhang, Arian Hosseini, Hritik Bansal, Mehran Kazemi, Aviral Kumar, and Rishabh Agarwal. Generative verifiers: Reward modeling as next-token prediction. In The 4th Workshop on Mathematical Reasoning and AI at NeurIPS'24 , 2024. URL https://openreview.net/ forum?id=CxHRoTLmPX .
- [76] Zhenru Zhang, Chujie Zheng, Yangzhen Wu, Beichen Zhang, Runji Lin, Bowen Yu, Dayiheng Liu, Jingren Zhou, and Junyang Lin. The lessons of developing process reward models in mathematical reasoning. arXiv preprint arXiv:2501.07301 , 2025.
- [77] Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, Siyuan Zhuang, Zhanghao Wu, Yonghao Zhuang, Zi Lin, Zhuohan Li, Dacheng Li, Eric Xing, Hao Zhang, Joseph E. Gonzalez, and Ion Stoica. Judging LLM-as-a-judge with MT-bench and chatbot arena. In Thirty-seventh Conference on Neural Information Processing Systems Datasets and Benchmarks Track , 2023. URL https://openreview.net/forum?id=uccHPGDlao .
- [78] Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, Siyuan Zhuang, Zhanghao Wu, Yonghao Zhuang, Zi Lin, Zhuohan Li, Dacheng Li, Eric Xing, et al. Judging llm-as-a-judge with mt-bench and chatbot arena. Advances in Neural Information Processing Systems , 36:46595-46623, 2023.
- [79] Jialun Zhong, Wei Shen, Yanzeng Li, Songyang Gao, Hua Lu, Yicheng Chen, Yang Zhang, Wei Zhou, Jinjie Gu, and Lei Zou. A comprehensive survey of reward models: Taxonomy, applications, challenges, and future. arXiv preprint arXiv:2504.12328 , 2025.
- [80] Lianghui Zhu, Xinggang Wang, and Xinlong Wang. JudgeLM: Fine-tuned large language models are scalable judges. In The Thirteenth International Conference on Learning Representations , 2025. URL https://openreview.net/forum?id=xsELpEPn4A .
## A RRM Implementation Details
## A.1 Prompt Template
Here we provide the detailed prompt template used to guide our Reward Reasoning Models (RRMs) in performing systematic comparative analysis between two responses. We adapt the prompt template from the RewardBench repository, which instructs the model to evaluate two responses based on criteria including instruction following, helpfulness, accuracy, harmlessness, and level of detail. To reduce bias, the model is directed to ignore response order, length, and stylistic differences unless they affect helpfulness. The model must output its final verdict in the format '\boxed{Assistant 1}' or '\boxed{Assistant 2}', with no ties allowed.
<details>
<summary>Image 9 Details</summary>

### Visual Description
## Text Document: RRM Prompt Template
### Overview
The image is a text document titled "RRM Prompt Template". It provides instructions for evaluating the quality of responses from two AI assistants for a given instruction. The document outlines rules for evaluation, potential sources of bias, and the expected output format. It also includes sections for the query, assistant responses, and analysis.
### Components/Axes
The document is structured into the following sections:
1. **Title:** RRM Prompt Template
2. **Introduction:** A paragraph explaining the purpose of the document, which is to guide the selection of the best response from two AI assistants.
3. **Rules of Evaluation:** A numbered list of rules to follow when evaluating the responses.
4. **Potential Sources of Bias:** A list of potential biases to avoid.
5. **Output Format:** Instructions on how to format the output based on which assistant is better.
6. **Query Section:** A placeholder for the query.
7. **Assistant Responses Section:** Placeholders for the responses from Assistant 1 and Assistant 2.
8. **Analysis Section:** Instructions to analyze the responses and select the better assistant.
### Detailed Analysis or ### Content Details
**Introduction:**
* The document instructs the user to select the best response from Assistant 1 or Assistant 2.
* The responses are generated by two different AI assistants.
* The user is instructed not to say both or neither are good.
**Rules of Evaluation:**
1. If the instruction does not contain harmful content, prioritize evaluating whether the output honestly/precisely/closely executes the instruction, then consider its helpfulness, accuracy, level of detail, harmlessness, etc.
2. If the instruction contains harmful content, prioritize the harmlessness and safety of the response.
3. Responses should not contain more/less than what the instruction asks for, as such responses do not precisely execute the instruction.
4. Avoid potential bias and ensure judgment is as objective as possible.
**Potential Sources of Bias:**
* The order in which the responses were presented should not affect judgment, as Response A and Response B are equally likely to be the better.
* The length of the responses should not affect judgment, as a longer response does not necessarily correspond to a better response. Evaluate if the response length is appropriate for the given instruction.
**Output Format:**
* The output should consist of "\boxed{Assistant 1}" if Assistant 1 is better, or "\boxed{Assistant 2}" if Assistant 2 is better.
* Omit any other output.
**Sections:**
* `## Query`: Contains the placeholder `{Query}`.
* `## Assistant responses`: Contains the following subsections:
* `### Assistant 1`: Contains the placeholder `{Response 1}`.
* `### Assistant 2`: Contains the placeholder `{Response 2}`.
* `## Analysis`: Contains the text "Let's analyze this step by step and decide which assistant is better, and then answer \boxed{Assistant 1} or \boxed{Assistant 2}."
### Key Observations
* The document provides a structured approach to evaluating AI assistant responses.
* It emphasizes the importance of objectivity and avoiding bias.
* The document uses placeholders for the query and responses, indicating it is a template.
### Interpretation
The document serves as a template for evaluating the responses of two AI assistants to a given query. It provides a set of guidelines to ensure a fair and objective evaluation process. The rules of evaluation prioritize factors such as accuracy, helpfulness, harmlessness, and safety. The document also highlights potential sources of bias, such as the order and length of the responses, to help the evaluator make an informed decision. The structured format, with placeholders for the query, responses, and analysis, makes it easy to use and adapt for different evaluation tasks. The final output format is clearly defined, ensuring consistency in the evaluation results.
</details>
In addition to the training prompt used for RRM models, we also include the evaluation prompt used for querying GPT-4o on the PPE benchmark. We follow the prompt format from Liu et al. [41], which instructs GPT-4o to select the best response from a set of candidates based on several criteria. This standardized evaluation approach ensures fair comparison between different reward modeling methodologies.
## LLM-as-a-Judge Prompt Template
You are a skilled little expert at scoring responses. You should evaluate given responses based on the given judging criteria.\nGiven the context of the conversation (the last round is the User's query) and multiple responses from the Assistant, you need to refer to the [General Evaluation Criteria] to score the responses. Based on the general evaluation criteria, state potential other specific criteria to the query, the weights of different criteria, and then select the best response among all candidates.\nBefore judging, please analyze step by step. Your judgement needs to be as strict as possible.
## #### Evaluation Criteria ####
1. Instruction Adherence:\n - Fully Adhered: The response fully complies with all instructions and requirements of the question.\n - Partially Adhered: The response meets most of the instructions but has some omissions or misunderstandings.\n - Basically Adhered: The response meets some instructions, but the main requirements are not fulfilled.\n - Not Adhered: The response does not meet any instructions.\n Example: If the question requires three examples and the response provides only one, it falls under 'Partially Adhered.' 2. Usefulness:\n - Highly Useful: The response provides comprehensive and accurate information, fully addressing the issue.\n - Useful but Incomplete: The response provides some useful information, but lacks details or accuracy.\n - Limited Usefulness: The response offers little useful information, with most content being irrelevant or incorrect.\n - Useless or Incorrect: The response is completely irrelevant or incorrect.\n Example: If there are factual errors in the response but the overall direction is correct, it falls under 'Useful but Incomplete.' 3. Level of Detail:\n - Very Detailed: The response includes ample details covering all aspects of the issue.\n - Detailed but Slightly Lacking: The response is fairly detailed but misses some important details.\n - Basically Detailed: The response provides some details but is not thorough enough overall.\n - Not Detailed: The response is very brief and lacks necessary details.\n Example: If the response provides only a simple conclusion without an explanation, it falls under 'Not Detailed.' 4. Relevance:\n - Highly Relevant: The response is highly relevant to the question, with information closely aligned with the topic.\n - Generally Relevant: The response is generally relevant but includes some unnecessary information.\n Partially Relevant: The response has a lot of content that deviates from the topic.\n - Not Relevant: The response is completely irrelevant.\n Example: If the response strays from the topic but still provides some relevant information, it falls under 'Partially Relevant.'
#### Conversation Context #### {conversation context & query}
#### Responses to be Scored #### [The Begin of Response] {the response} [The End of Response]
#### Output Format Requirements ####
## Output with three lines
Specific Criteria: <Other potential criteria specific to the query and the context, and the weights of each criteria>.
Analysis: <Compare different responses based on given Criteria>.
Scores: <the index of the best response based on the judgement, in the format of <\boxed{x}>.
## A.2 Hyperparameters for Training RRM
Table 7 presents the key hyperparameters used for training RRMs. These parameters were carefully selected to optimize the reinforcement learning process and ensure effective development of reasoning capabilities in our models.
Table 7: Hyperparameters used for training RRMs.
| Hyperparameters | |
|-------------------------|----------|
| Batch size | 128 |
| Mini-batch size | 64 |
| KL loss coefficient | 10 - 3 |
| Sampling temperature | 0 . 6 |
| Maximum prompt length | 4096 |
| Maximum response length | 8192 |
| GRPO group size | 16 |
| Learning rate (RRM-32B) | × 10 - 7 |
| Learning rate (RRM-7B) | 10 - 6 |
## B Reward-Guided Best-of-N Inference
## B.1 Detailed Results on Subsets of MMLU-Pro and GPQA
We present detailed results on the constituent subsets of MMLU-Pro and GPQA benchmarks. Figure 9 illustrates the performance comparison among our RRM-7B, RRM-32B models and Skywork-RewardGemma-2-27B-v0.2 across the 14 subsets of MMLU-Pro. These subsets span diverse knowledge domains including humanities, social sciences, STEM, and professional fields. The results reveal interesting patterns in model performance. Notably, RRM-7B outperforms Skywork-Reward-Gemma2-27B-v0.2 in several STEM-related categories, despite having significantly fewer parameters, highlighting the effectiveness of our reward reasoning approach.
Figure 9: Performance comparison of Skywork-Reward-Gemma-2-27B-v0.2, RRM-7B, and RRM32B on the 14 subsets of MMLU-Pro.
<details>
<summary>Image 10 Details</summary>

### Visual Description
## Bar Chart: Model Accuracy by Subject
### Overview
The image is a bar chart comparing the accuracy of three models (Skywork-Reward, RRM-7B, and RRM-32B) across various subjects. The y-axis represents accuracy, ranging from 0.0 to 1.0. The x-axis represents the subjects: Chemistry, Engineering, Math, Law, Biology, Health, Computer Science, Other, History, Physics, Business, Philosophy, Psychology, and Economics. Each subject has three bars representing the accuracy of each model.
### Components/Axes
* **Title:** (Inferred) Model Accuracy by Subject
* **Y-axis:**
* **Label:** Accuracy
* **Scale:** 0.0 to 1.0, with tick marks at 0.0, 0.2, 0.4, 0.6, 0.8, and 1.0.
* **X-axis:**
* **Label:** Subjects
* **Categories:** Chemistry, Engineering, Math, Law, Biology, Health, Computer Science, Other, History, Physics, Business, Philosophy, Psychology, Economics
* **Legend:** Located in the top-left corner.
* **Skywork-Reward:** Blue
* **RRM-7B:** Orange
* **RRM-32B:** Green
### Detailed Analysis
Here's a breakdown of the accuracy for each model across the subjects:
* **Chemistry:**
* Skywork-Reward (Blue): ~0.65
* RRM-7B (Orange): ~0.75
* RRM-32B (Green): ~0.80
* **Engineering:**
* Skywork-Reward (Blue): ~0.72
* RRM-7B (Orange): ~0.70
* RRM-32B (Green): ~0.78
* **Math:**
* Skywork-Reward (Blue): ~0.58
* RRM-7B (Orange): ~0.78
* RRM-32B (Green): ~0.82
* **Law:**
* Skywork-Reward (Blue): ~0.64
* RRM-7B (Orange): ~0.64
* RRM-32B (Green): ~0.83
* **Biology:**
* Skywork-Reward (Blue): ~0.64
* RRM-7B (Orange): ~0.47
* RRM-32B (Green): ~0.83
* **Health:**
* Skywork-Reward (Blue): ~0.62
* RRM-7B (Orange): ~0.70
* RRM-32B (Green): ~0.82
* **Computer Science:**
* Skywork-Reward (Blue): ~0.66
* RRM-7B (Orange): ~0.62
* RRM-32B (Green): ~0.82
* **Other:**
* Skywork-Reward (Blue): ~0.56
* RRM-7B (Orange): ~0.58
* RRM-32B (Green): ~0.66
* **History:**
* Skywork-Reward (Blue): ~0.75
* RRM-7B (Orange): ~0.47
* RRM-32B (Green): ~0.74
* **Physics:**
* Skywork-Reward (Blue): ~0.68
* RRM-7B (Orange): ~0.74
* RRM-32B (Green): ~0.81
* **Business:**
* Skywork-Reward (Blue): ~0.80
* RRM-7B (Orange): ~0.82
* RRM-32B (Green): ~0.90
* **Philosophy:**
* Skywork-Reward (Blue): ~0.84
* RRM-7B (Orange): ~0.90
* RRM-32B (Green): ~0.94
* **Psychology:**
* Skywork-Reward (Blue): ~0.64
* RRM-7B (Orange): ~0.72
* RRM-32B (Green): ~0.81
* **Economics:**
* Skywork-Reward (Blue): ~0.77
* RRM-7B (Orange): ~0.67
* RRM-32B (Green): ~0.67
### Key Observations
* RRM-32B (Green) generally has the highest accuracy across most subjects.
* RRM-7B (Orange) often has lower accuracy than Skywork-Reward (Blue) and RRM-32B (Green).
* The accuracy varies significantly across subjects for all models.
* Philosophy and Business tend to have the highest accuracy scores for all models.
* Biology and History have the lowest accuracy scores for RRM-7B.
### Interpretation
The bar chart provides a comparative analysis of the accuracy of three different models across a range of subjects. The consistent outperformance of RRM-32B suggests it is the most effective model overall. The varying performance across subjects indicates that the models' effectiveness is subject-dependent, possibly due to the nature of the subject matter or the training data used for each model. The relatively lower performance of RRM-7B in certain subjects like Biology and History could indicate specific weaknesses in its architecture or training. The high accuracy in Philosophy and Business might reflect the nature of the questions or the availability of relevant training data. Further investigation into the models' architectures, training data, and the specific questions used for evaluation would be necessary to understand the underlying reasons for these performance differences.
</details>
More impressively, our RRM-32B model demonstrates consistently superior or comparable performance across all subsets compared to Skywork-Reward-Gemma-2-27B-v0.2. This consistency highlights the robust generalization capabilities of our larger model across diverse knowledge domains. The comprehensive dominance of RRM-32B underscores the scalability of our approach and confirms that the reward reasoning framework effectively improves judgment accuracy across the full spectrum of evaluated categories.
Similarly, Figure 10 presents the performance breakdown across the GPQA subsets. The pattern remains consistent, with RRM-7B showing stronger performance in certain technical categories while occasionally lagging behind Skywork-Reward-Gemma-2-27B-v0.2 in more general knowledge areas. Meanwhile, RRM-32B maintains excellent performance across all subsets. This comprehensive
analysis further validates the effectiveness of our reward reasoning approach in handling complex scientific and technical queries that require sophisticated judgment capabilities.
Figure 10: Performance comparison of Skywork-Reward-Gemma-2-27B-v0.2, RRM-7B, and RRM32B on the 16 subsets of GPQA.
<details>
<summary>Image 11 Details</summary>

### Visual Description
## Bar Chart: Model Accuracy by Subject
### Overview
The image is a bar chart comparing the accuracy of three different models (Skywork-Reward, RRM-7B, and RRM-32B) across various subjects. The x-axis represents the subjects, and the y-axis represents the accuracy, ranging from 0.0 to 1.0. Each subject has three bars representing the accuracy of each model.
### Components/Axes
* **X-axis:** Subjects (listed below)
* **Y-axis:** Accuracy, ranging from 0.0 to 1.0 in increments of 0.2.
* **Legend:** Located in the top-left corner.
* Skywork-Reward (Blue)
* RRM-7B (Orange)
* RRM-32B (Green)
### Detailed Analysis
The chart displays the accuracy of three models across the following subjects:
1. **Quantum Mechanics:**
* Skywork-Reward: ~0.55
* RRM-7B: ~0.60
* RRM-32B: ~0.72
2. **Chemistry (General):**
* Skywork-Reward: ~0.45
* RRM-7B: ~0.50
* RRM-32B: ~0.62
3. **Organic Chemistry:**
* Skywork-Reward: ~0.35
* RRM-7B: ~0.35
* RRM-32B: ~0.45
4. **Molecular Biology:**
* Skywork-Reward: ~0.45
* RRM-7B: ~0.55
* RRM-32B: ~0.60
5. **Physics (General):**
* Skywork-Reward: ~0.47
* RRM-7B: ~0.55
* RRM-32B: ~0.70
6. **Electromagnetism And Photonics:**
* Skywork-Reward: ~0.65
* RRM-7B: ~0.72
* RRM-32B: ~0.78
7. **Genetics:**
* Skywork-Reward: ~0.42
* RRM-7B: ~0.65
* RRM-32B: ~0.79
8. **Astrophysics:**
* Skywork-Reward: ~0.35
* RRM-7B: ~0.42
* RRM-32B: ~0.67
9. **High-Energy Particle Physics:**
* Skywork-Reward: ~0.55
* RRM-7B: ~0.58
* RRM-32B: ~0.60
10. **Relativistic Mechanics:**
* Skywork-Reward: ~0.75
* RRM-7B: ~0.82
* RRM-32B: ~0.85
11. **Physical Chemistry:**
* Skywork-Reward: ~0.98
* RRM-7B: ~0.98
* RRM-32B: ~0.98
12. **Condensed Matter Physics:**
* Skywork-Reward: ~0.98
* RRM-7B: ~0.75
* RRM-32B: ~0.98
13. **Inorganic Chemistry:**
* Skywork-Reward: ~0.50
* RRM-7B: ~0.35
* RRM-32B: ~0.68
14. **Statistical Mechanics:**
* Skywork-Reward: ~0.34
* RRM-7B: ~0.34
* RRM-32B: ~0.34
15. **Optics And Acoustics:**
* Skywork-Reward: ~0.34
* RRM-7B: ~0.67
* RRM-32B: ~0.67
16. **Analytical Chemistry:**
* Skywork-Reward: ~0.87
* RRM-7B: ~0.67
* RRM-32B: ~0.92
### Key Observations
* The RRM-32B model generally performs better than the other two models across most subjects.
* The Skywork-Reward model and RRM-7B model have similar performance in many subjects.
* The models achieve near-perfect accuracy (1.0) in Physical Chemistry and Condensed Matter Physics.
* The models perform poorly in Statistical Mechanics, with accuracy around 0.34 for all three.
### Interpretation
The bar chart provides a comparative analysis of the accuracy of three different models across a range of subjects. The RRM-32B model appears to be the most accurate overall, suggesting it may be a more robust or better-trained model. The consistent low performance in Statistical Mechanics indicates that this subject may be particularly challenging for all three models, potentially due to the complexity or nature of the subject matter. The near-perfect accuracy in Physical Chemistry and Condensed Matter Physics suggests that these subjects are relatively easier for the models to understand or predict. The data suggests that the choice of model can significantly impact accuracy depending on the subject.
</details>
## B.2 Go Further into Knockout Tournament
To gain a deeper understanding of the knockout tournament strategy described in Section 3, we conduct additional experiments following the setup in Section 4.5.1. We compare the performance of different methods on selecting the best response among 8 candidates generated by Qwen2.5-Math7B-Instruct for each MATH question. We reward the responses with RRM-7B and RRM-32B, and benchmark them against Qwen2.5-Math-PRM-7B and Qwen2.5-Math-PRM-72B [76]. In addition to using reward models, we also include non-verifier strategies such as majority voting (Voting@8) and best-of-N oracle selection for reference. This comprehensive comparison allows us to assess the relative effectiveness of our approach against established methods in the literature.
The numerical results are summarized in Table 8. When compared with baselines, RRM-7B surpasses all comparison methods, including voting@8 and PRM judges. RRM-32B further narrows the gap toward oracle-level accuracy, significantly outperforming PRM-based baselines. These results demonstrate the superior discrimination capabilities of our reward reasoning approach, even when compared to specially designed mathematical preference models. The consistent performance advantage across different model sizes confirms the effectiveness of our framework in identifying high-quality mathematical reasoning across varied problem complexities.
Table 8: Comparison between RRM and Qwen2.5-Math-PRM models on MATH.
| Models | Accuracy |
|----------------------|------------|
| Voting@8 | 86.8 |
| Best-of-8 Oracle | 91.7 |
| Qwen2.5-Math-PRM-7B | 87.8 |
| Qwen2.5-Math-PRM-72B | 88.5 |
| RRM-7B | 88.7 |
| RRM-32B | 90.4 |
As shown in Figure 11, as the knockout tournament progresses through successive elimination rounds, we observe a consistent improvement in accuracy, demonstrating the benefits of iterative comparison. Notably, the knockout tournament achieves this consistent accuracy improvement with only O ( n ) pairwise comparisons. This efficient scaling behavior highlights the practical advantage of our
approach in scenarios where computational resources may be constrained, providing an effective balance between performance gains and computational requirements.
Figure 11: Accuracy progression of the knockout tournament strategy on MATH as elimination rounds proceed.
<details>
<summary>Image 12 Details</summary>

### Visual Description
## Line Chart: MATH Accuracy vs. Tournament Round
### Overview
The image is a line chart comparing the MATH accuracy of different models (RRM-7B and RRM-32B) with and without voting, across tournament rounds (0 to 3). The chart also includes horizontal dashed lines indicating the performance of other models like RRM-32B Elo, RRM-7B Elo, Qwen2.5-PRM-70B, Qwen2.5-PRM-7B, and Voting@8.
### Components/Axes
* **X-axis:** Tournament round, with values 0, 1, 2, and 3.
* **Y-axis:** MATH accuracy, ranging from 82 to 90.
* **Legend (bottom-right):**
* Red: RRM-7B with voting
* Orange: RRM-7B without voting
* Blue: RRM-32B with voting
* Light Blue: RRM-32B without voting
* **Horizontal dashed lines (top-left):**
* RRM-32B Elo (at approximately 90.5)
* RRM-7B Elo, Qwen2.5-PRM-70B (at approximately 88.8)
* Qwen2.5-PRM-7B (at approximately 87.8)
* Voting@8 (at approximately 87)
### Detailed Analysis
* **RRM-7B with voting (Red):**
* Trend: Slopes upward.
* Data points: (0, 82), (1, 85.5), (2, 87.5), (3, 88.5)
* **RRM-7B without voting (Orange):**
* Trend: Slopes upward.
* Data points: (0, 82), (1, 85), (2, 87), (3, 88.2)
* **RRM-32B with voting (Blue):**
* Trend: Slopes upward.
* Data points: (0, 82.2), (1, 86.5), (2, 88.5), (3, 90)
* **RRM-32B without voting (Light Blue):**
* Trend: Slopes upward.
* Data points: (0, 82.2), (1, 86), (2, 88), (3, 89.8)
### Key Observations
* Both RRM-32B models (with and without voting) consistently outperform the RRM-7B models.
* For both model sizes (7B and 32B), using voting generally results in slightly higher MATH accuracy.
* The performance gap between the models appears to narrow as the tournament round increases, but RRM-32B maintains the lead.
* The RRM-32B with voting model approaches the performance level of "RRM-32B Elo" by round 3.
### Interpretation
The chart demonstrates the impact of model size and voting on MATH accuracy in a tournament setting. The RRM-32B models, being larger, achieve higher accuracy than the RRM-7B models. The use of voting enhances the performance of both models, although the effect seems more pronounced in earlier rounds. The convergence of the lines suggests that the benefit of additional tournament rounds diminishes as the models approach their performance ceiling. The horizontal lines provide a benchmark against other models, indicating the relative performance of the RRM models in comparison to established models like RRM-32B Elo, RRM-7B Elo, Qwen2.5-PRM-70B, Qwen2.5-PRM-7B, and Voting@8. The data suggests that increasing model size and incorporating voting are effective strategies for improving MATH accuracy.
</details>
## C Post-thinking Token Length Distribution
To evaluate the impact of thinking budget on model performance, we need to establish an appropriate token budget for the response phase that follows the thinking phase. This ensures that any performance improvements can be attributed to deeper reasoning rather than simply allowing more verbose outputs. The careful calibration of this post-thinking budget is critical for isolating the effects of extended reasoning from potential confounding factors related to output length.
We analyze the token length distribution of responses generated by RRM-32B on the RewardBench dataset after the thinking phase concluded. Figure 12 shows the distribution of post-thinking token length across various samples. The analysis reveals that all the responses require fewer than 100 tokens to express the final judgment after completing the reasoning process.
Figure 12: Post-thinking token length distribution of RRM-32B.
<details>
<summary>Image 13 Details</summary>

### Visual Description
## Histogram: Post-thinking Token Length Frequency
### Overview
The image is a histogram showing the frequency distribution of "Post-thinking token length". The x-axis represents the token length, and the y-axis represents the frequency. The histogram bars are light blue.
### Components/Axes
* **X-axis:** "Post-thinking token length"
* Scale: 25, 50, 75
* **Y-axis:** "Frequency"
* Scale: 0, 100, 200, 300, 400
### Detailed Analysis
The histogram shows the distribution of token lengths. The frequency is highest around token lengths of 30-40, and it decreases as the token length increases or decreases from this range.
* **Token Length 10:** Frequency is approximately 5.
* **Token Length 20:** Frequency is approximately 10.
* **Token Length 30:** Frequency is approximately 400.
* **Token Length 40:** Frequency is approximately 300.
* **Token Length 50:** Frequency is approximately 280.
* **Token Length 60:** Frequency is approximately 70.
* **Token Length 70:** Frequency is approximately 20.
* **Token Length 80:** Frequency is approximately 5.
### Key Observations
* The distribution is unimodal, with a peak around token lengths of 30-40.
* The distribution is skewed to the right, indicating that there are more tokens with lengths less than the mean than tokens with lengths greater than the mean.
* The frequency drops off sharply after a token length of 50.
### Interpretation
The histogram suggests that the "Post-thinking" process tends to produce tokens with lengths between 30 and 40 most frequently. Shorter and longer tokens are less common. The right skew indicates that while longer tokens exist, they are relatively rare compared to the shorter ones. This could be due to the nature of the "Post-thinking" process, which might favor generating tokens of a certain length range.
</details>
Based on this observation, we set a fixed post-thinking budget of 100 tokens for all our sequential scaling experiments. This budget is sufficient to accommodate typical response patterns while preventing the model from extending its reasoning during the response phase, which would confound our analysis of thinking horizon effects. By maintaining this consistent response budget across all experiments, we ensure that performance differences can be directly attributed to variations in the thinking phase length rather than differences in output verbosity. This methodological choice strengthens the validity of our conclusions regarding the impact of extended reasoning on model performance.
## D Reasoning Pattern Analysis
Table 9 presents the pattern groups and keywords applied in reasoning pattern analysis.
Table 9: Pattern groups and keywords applied in reasoning pattern analysis.
| Pattern Group | Keywords |
|-----------------|------------------------------------------------------------------------------------------------------------------|
| Transition | alternatively, think differently, another way, another approach, another method, another solution, another point |
| Reflection | wait, verify, make sure, hold on, think again, Let me check, seems right, seems incorrect |
| Comparison | more, compared to, comparison, between the two, similarly |
| Breakdown | break down, break this down |