## Reward Reasoning Model
Jiaxin Guo ∗ 1 , 2 Zewen Chi ∗ 1 Li Dong ∗ 1
Qingxiu Dong 1 , 3
Xun Wu 1 Shaohan Huang 1 Furu Wei 1 â‹„
1 Microsoft Research
2 Tsinghua University https://aka.ms/GeneralAI
## Abstract
Reward models play a critical role in guiding large language models toward outputs that align with human expectations. However, an open challenge remains in effectively utilizing test-time compute to enhance reward model performance. In this work, we introduce Reward Reasoning Models (RRMs), which are specifically designed to execute a deliberate reasoning process before generating final rewards. Through chain-of-thought reasoning, RRMs leverage additional test-time compute for complex queries where appropriate rewards are not immediately apparent. To develop RRMs, we implement a reinforcement learning framework that fosters selfevolved reward reasoning capabilities without requiring explicit reasoning traces as training data. Experimental results demonstrate that RRMs achieve superior performance on reward modeling benchmarks across diverse domains. Notably, we show that RRMs can adaptively exploit test-time compute to further improve reward accuracy. The pretrained reward reasoning models are available at https: //huggingface.co/Reward-Reasoning .
Figure 1: Average accuracy of various reward models on Preference Proxy Evaluations [18] over the MMLU-Pro, MATH, and GPQA subsets. The proposed reward reasoning model (RRM) outperforms previous reward models across model sizes. We also conduct reinforcement learning on unlabeled data, using RRM as the reward model. Even without ground-truth answers, reinforcement learning with RRM achieves significant improvements on GPQA, which evaluates general-domain reasoning.
<details>
<summary>Image 1 Details</summary>

### Visual Description
## Scatter Plot: Frontier Reward Modeling Performance
### Overview
The image presents a scatter plot comparing the average accuracy of various reward modeling approaches against the number of parameters used in the models. A secondary bar chart shows the accuracy of a reinforcement learning approach with and without RRM (Reward-based Reinforcement Modeling) on unlabeled data.
### Components/Axes
* **X-axis:** Number of Parameters (B) - Scale ranges from approximately 3B to 400B. Marked values are 3B, 7B, 70B, and 400B.
* **Y-axis:** Average Accuracy (%) - Scale ranges from approximately 55% to 85%. Marked values are 55%, 60%, 65%, 70%, 75%, 80%, and 85%.
* **Scatter Plot Data Points:** Represent different reward modeling approaches.
* **Shaded Region:** A light orange shaded region in the top-left corner, indicating a performance frontier.
* **Legend (Scatter Plot):**
* RRM-32B (Star symbol, dark orange)
* RRM-7B (Star symbol, dark orange)
* Other models (Circle symbol, gray)
* **Bar Chart:** Compares accuracy with and without RRM.
* **Legend (Bar Chart):**
* R1-Distill-Qwen-7B (Blue bar)
* Post-trained with RRM (Green bar)
* **Bar Chart X-axis:** GPQA
* **Bar Chart Y-axis:** Accuracy (%) - Scale ranges from 0% to 50%. Marked values are 0%, 20%, 30%, 40%, and 50%.
### Detailed Analysis or Content Details
**Scatter Plot:**
* **RRM-32B:** Located at approximately (320B, 82%).
* **RRM-7B:** Located at approximately (7B, 74%).
* **Meta-J1-Llama-70B:** Located at approximately (70B, 79%).
* **Athene-RM-70B:** Located at approximately (70B, 72%).
* **Llama-3.1-70B-Instruct:** Located at approximately (70B, 69%).
* **Meta-J1-Llama-8B:** Located at approximately (8B, 68%).
* **InternLM2-20B-Reward:** Located at approximately (20B, 66%).
* **Nемоtron-4-340B-Reward:** Located at approximately (340B, 64%).
* **Armo-8B-v0.1:** Located at approximately (8B, 64%).
* **DeepSeek-GRM-27B:** Located at approximately (27B, 63%).
* **Llama-3.1-8B-Instruct:** Located at approximately (8B, 58%).
**Bar Chart:**
* **R1-Distill-Qwen-7B:** Accuracy is approximately 26.8%.
* **Post-trained with RRM:** Accuracy is approximately 40.9%.
### Key Observations
* The scatter plot shows a general trend of increasing accuracy with increasing model size (number of parameters).
* RRM-32B and RRM-7B models achieve higher accuracy compared to other models with similar parameter counts.
* The bar chart demonstrates a significant improvement in accuracy when using RRM for post-training on the GPQA dataset.
* The shaded region suggests a performance frontier, with RRM-32B and RRM-7B models approaching or exceeding it.
### Interpretation
The data suggests that RRM is an effective technique for improving the performance of reward modeling, particularly when combined with larger models. The scatter plot illustrates a positive correlation between model size and accuracy, but RRM appears to enhance this relationship, allowing models to achieve higher accuracy for a given number of parameters. The bar chart provides concrete evidence of RRM's effectiveness in a reinforcement learning context, showing a substantial increase in accuracy on the GPQA dataset. The positioning of RRM-32B and RRM-7B near the performance frontier indicates that these models represent state-of-the-art performance in reward modeling. The outlier is the significant jump in accuracy when using RRM, suggesting it is a key component for achieving high performance. The data implies that RRM is a valuable tool for developing more effective and efficient reward models, which can lead to improvements in reinforcement learning and other AI applications.
</details>
∗ Equal contribution. ⋄ Corresponding author.
3 Peking University
## 1 Introduction
Large language models (LLMs) such as GPTs [9, 1] have significantly transformed the field of artificial intelligence. In recent years, the development paradigm of LLMs has evolved from primarily scaling pre-training resources to emphasizing post-training techniques, driven by the dual imperatives of aligning models with human preferences [45] and enhancing specific capabilities like reasoning [6, 56]. This shift reflects a growing recognition that model performance depends not only on scale but also on sophisticated methods to refine model behavior after initial training.
Reinforcement learning has emerged as a fundamental approach in LLM post-training, leveraging supervision signals from either human feedback (RLHF) or verifiable rewards (RLVR) [45, 15, 19, 33, 22]. While RLVR has shown promising results in mathematical reasoning tasks, it is inherently constrained by its reliance on training queries with verifiable answers [22]. This requirement substantially limits RLVR's application to large-scale training on general-domain queries where verification is often intractable [16, 29, 58]. In contrast, RLHF typically employs a reward model as a proxy for human preference, enabling more extensive application across diverse domains [7, 44]. Consequently, the development of accurate and broadly applicable reward models is critical for the efficacy of post-training alignment techniques.
Recent work on reward models can be categorized into scalar reward models [45, 39] and generative reward models [12, 54, 60, 80]. Scalar reward models typically replace the decoding layer with a linear head to predict a single scalar value. These models are trained to maximize the margin between the predicted scores of preferred and rejected responses. Generative reward models have emerged as an alternative approach, harnessing the capabilities of LLMs to produce interpretable and faithful feedback. These models offer enhanced flexibility, enabling them to follow adaptive evaluation instructions to construct synthetic training data, thereby facilitating self-improvement through iterative refinement [21, 78].
Despite the widespread application of current reward models, it remains an open challenge to effectively scale test-time compute for reward estimation. To serve as general-purpose evaluators, reward models should be capable of adapting to a diverse spectrum of queries, ranging from immediately obvious questions to complex tasks that require extensive reasoning [20, 50]. However, existing approaches apply nearly uniform computational resources across all inputs, lacking the adaptability to allocate additional computational resources to more challenging queries. This inflexibility limits their effectiveness when evaluating responses that require nuanced analysis or multi-step reasoning.
To address the aforementioned challenge, we propose Reward Reasoning Models (RRMs). Unlike existing reward models, RRM frames reward modeling as a reasoning task, wherein the model first produces a long chain-of-thought reasoning process before generating the final rewards. Since supervised data providing reward reasoning traces are not readily available, we develop a training framework called Reward Reasoning via Reinforcement Learning, which encourages RRMs to selfevolve their reward reasoning capabilities within a rule-based reward environment. Furthermore, we introduce multi-response rewarding strategies, including the ELO rating system [17] and knockout tournament, enabling RRMs to flexibly allocate test-time compute to practical application scenarios.
Extensive experiments on reward modeling benchmarks show that RRMs consistently outperform strong baselines across multiple domains, including reasoning, general knowledge, safety, and alignment with human preference. Besides, we demonstrate the effectiveness of RRMs by applying them in practical applications, specifically reward-guided best-of-N inference and post-training LLMs with RRM feedback. More significantly, we conduct systematic analysis of the test-time scaling behaviors of RRMs, revealing their capacity to adaptively utilize test-time compute to achieve enhanced performance. Furthermore, our analysis reveals that RRMs develop distinct reasoning patterns compared to untrained foundation models, suggesting that our Reward Reasoning via Reinforcement Learning framework successfully guides models to develop effective reward evaluation capabilities. These insights provide deeper understanding of reward reasoning processes and will likely inspire the development of future reward reasoning models within the research community.
Our main contributions are as follows:
- We propose Reward Reasoning Models (RRMs), which perform explicit reasoning before producing final rewards. This reasoning phase enables RRMs to adaptively allocate additional computational resources when evaluating responses to complex tasks. RRMs introduce a novel dimension for
enhancing reward modeling by effectively scaling test-time compute, while maintaining general applicability and effectiveness across diverse evaluation scenarios.
- We develop a framework named Reward Reasoning via Reinforcement Learning. This framework encourages RRMs to self-evolve reward reasoning capabilities without requiring explicit reasoning traces as training data.
- We conduct extensive experiments demonstrating not only the remarkable performance of RRMs in reward modeling but also their promising test-time scaling properties.
## 2 Related Work
Reward Models Reward models can be characterized along two dimensions: reward formulation and scoring scheme [44, 79]. Formulation strategies include numeric only, which assigns scalar scores to query-response pairs [45, 39, 62, 63], and generative, which produces natural language feedback from which rewards may be extracted [3, 5, 11, 12, 41, 57, 71, 75]. Scoring schemes typically follow either absolute approaches, evaluating individual query-response pairs independently [16, 20, 23, 66, 73, 74], or discriminative methods that compare candidate responses to express relative preferences [28, 35, 40, 47, 54, 59, 80].
Generative Reward Models Generative reward models (GRMs), conceptually aligned with the LLM-as-a-Judge paradigm [67, 77], offer nuanced, interpretable feedback with flexibility for both single-instance evaluation and multi-response comparison [32, 43]. This approach addresses limitations of traditional evaluation methods like ROUGE [38] and BLEU [46], which struggle with open-ended tasks requiring sophisticated judgment [51]. GRMs can support judgment across diverse tasks, including multimodal inputs [31, 35, 80], and contemporaneous work on GRMs demonstrates promising scalability in both model capacity and inference-time compute [14, 41]. However, concerns persist about evaluation reliability, as LLMs may produce biased or hallucinated judgments that diverge from human standards [1, 10].
Inference-Time Scaling Inference-time scaling dynamically adjusts computational resources during model inference based on input complexity, inspired by human adaptive reasoning [30, 55, 68]. Recent approaches include parallel scaling strategies such as multi-sampling [8] and reward modelguided aggregation [37, 55, 76], which combine multiple outputs to enhance quality. Alternative methods utilize horizon-based scaling to extend reasoning traces [64]. Advanced systems like OpenAI's o1 and DeepSeek's R1 series demonstrate spontaneous computational allocation that adjusts 'thinking horizons' in response to task complexity [22, 27]. These approaches collectively underscore the importance of inference-time adaptability in improving model performance, particularly on complex reasoning tasks.
## 3 Reward Reasoning Model
## 3.1 Input Representation
Figure 2 provides an overview of reward reasoning models (RRMs). RRMs utilize the Qwen2 [69] model architecture with a Transformer-decoder as backbone. We formulate the reward modeling task as a text completion problem, wherein RRMs take queries and corresponding responses as input, and autoregressively generate output text consisting of a thinking process followed by a final judgment. Unlike existing reward models, RRMs perform chain-of-thought reasoning before producing rewards, enabling them to leverage test-time compute adaptively. We refer to this process as reward reasoning.
Each input of RRMs contains a query and two corresponding responses. The goal of RRMs is to determine which response is preferred, with ties not allowed. We employ the system prompt from the RewardBench repository 2 , which guides the model to perform a systematic analysis of the two responses according to several evaluation criteria, including instruction fidelity, helpfulness, accuracy, harmlessness, and level of detail. The model is also explicitly instructed to avoid common biases (such as response order or length) and must justify its judgment through structured reasoning before
2 https://github.com/allenai/reward-bench
<details>
<summary>Image 2 Details</summary>

### Visual Description
\n
## Diagram: Reward Model Comparison
### Overview
The image is a diagram illustrating three different reward models used in reinforcement learning: a Scalar Reward Model, a Generative Reward Model, and a Reward Reasoning Model. It visually compares how each model processes a query and response to generate a reward signal. The diagram highlights the differences in complexity and the level of justification provided by each model.
### Components/Axes
The diagram is divided into three main sections, labeled (a), (b), and (c), each representing a different reward model. Each section includes:
* **Query & Response:** Represented by an orange box.
* **Reward Model:** Represented by a light green box.
* **Reward Output:** The output of the reward model, varying in format depending on the model type.
* **Reinforcement Learning:** A curved arrow indicating the feedback loop to reinforcement learning.
* **Long Reasoning:** A section in (c) showing intermediate reasoning steps.
### Detailed Analysis or Content Details
**(a) Scalar Reward Model:**
* **Query & Response:** "Q: 3x5=? A: 15."
* **Scalar Reward Model:** Labeled as "Scalar Reward Model".
* **Scalar Reward:** Output is "0.92".
**(b) Generative Reward Model:**
* **Query & Response:** "Q: 3x5=? A: 15."
* **Generative Reward Model:** Labeled as "Generative Reward Model".
* **Reward with Justification:** Output is "9, because ...". The "because..." indicates a textual justification is provided.
**(c) Reward Reasoning Model:**
* **Query & Response:** "Q: 3x5=? A: 15: 16"
* **Reward Reasoning Model:** Labeled as "Reward Reasoning Model".
* **Long Reasoning:** Contains several intermediate reasoning steps represented by green boxes with text:
* "Okay, so I need to..." with a thinking face emoji.
* "Looking back, ..." with a thinking face emoji.
* "Given that..." with a thinking face emoji.
* "Alternatively,...But if..." with a thinking face emoji.
* "Let's analyze..." with a thinking face emoji.
* "Wait, perhaps...Thus..." with a thinking face emoji.
* **Response:** Contains several response options represented by green boxes with text:
* "The answer is A."
* "The answer is B."
* "...".
* **Reinforcement Learning:** A curved arrow connects the "Response" section back to the "Reward Reasoning Model", indicating a feedback loop.
* **Rewards:** Represented by equations:
* R<sub>1</sub> = +1
* R<sub>2</sub> = -1
* R<sub>n</sub> = +1
### Key Observations
* The Scalar Reward Model provides a single numerical reward.
* The Generative Reward Model provides a numerical reward *and* a textual justification.
* The Reward Reasoning Model demonstrates a multi-step reasoning process before arriving at a reward, and provides multiple possible responses with associated rewards.
* The use of emojis in the "Long Reasoning" section suggests a simulation of thought processes.
* The rewards (R1, R2, Rn) are simple binary rewards (+1 or -1), indicating a basic reward structure.
### Interpretation
The diagram illustrates a progression in the complexity of reward models used in reinforcement learning. The Scalar Reward Model is the simplest, offering a direct numerical assessment. The Generative Reward Model adds interpretability by providing a justification for the reward. The Reward Reasoning Model is the most sophisticated, simulating a reasoning process and offering a more nuanced evaluation of the response.
The diagram suggests that more complex reward models can provide richer feedback signals to the reinforcement learning agent, potentially leading to more effective learning. The inclusion of intermediate reasoning steps in the Reward Reasoning Model highlights the importance of explainability and transparency in AI systems. The use of emojis is a stylistic choice to emphasize the "thinking" aspect of the model. The diagram demonstrates a shift from simple reward signals to more human-like reasoning and justification in reward mechanisms.
</details>
(c) Reward Reasoning Model
Figure 2: An overview of reward reasoning model (RRM). RRM adaptively leverages test-time compute through chain-of-though reasoning before producing rewards.
outputting its final decision in the format '\boxed{Assistant 1}' or '\boxed{Assistant 2}'. The detailed prompt template is provided in Appendix A.1.
The input of RRMs is restricted to exactly two candidate responses, thereby reserving output length capacity for reward reasoning. Section 3.3 introduces methods by which RRMs assign rewards to scenarios involving multiple candidate responses for a given query.
## 3.2 Model Training with Reinforcement Learning
We develop a training framework called Reward Reasoning via Reinforcement Learning to train RRMs. Unlike conventional supervised fine-tuning approaches, which relies on existing reasoning traces, our framework encourages RRMs to self-evolve their reasoning capacities within a rule-based reward environment. The reward function is defined as follows:
$$\mathcal { R } = \begin{cases} + 1 , & R M e c t s \, c o r r e c t \, r e s p o n s e & ( 1 ) \\ - 1 , & o t h e r w i s e \end{cases}$$
Note that the reward R evaluates whether RRM correctly prefers the ground-truth response, rather than scoring its own outputs. Despite the simplicity of the reward signals, such rule-based rewards can effectively supervise the policy models to develop reasoning patterns that lead to correct final judgments.
Weuse Deepseek-R1 distilled models [22] as base models, applying group relative policy optimization (GRPO) [70] for training, implemented with the verl library [53]. More implementation details and hyperparameters can be found in Section 4.1 and Appendix A.2.
## 3.3 Multi-Response Rewarding Strategies
Although the input structure of RRMs strictly accepts two candidate responses, RRMs can adaptively reward multiple responses of a specific query. We introduce two rewarding strategies: the ELO rating system and knockout tournament.
ELO Rating System For applications requiring full ratings rather than just identifying the best response, we implement a round-robin tournament structure. In this approach, each candidate is compared with all others pairwise. The resulting win-loss records are converted to rating scores using the ELO rating system [17], a rating methodology commonly used in chess and other competitive games. While this strategy can process ( n 2 ) = O ( n 2 ) pairwise comparison results, computational cost can be reduced by sampling a subset of the pairwise matchups. The resulting ratings can serve as rewards in reinforcement learning from human feedback (RLHF). Experiments demonstrate that we successfully post-train an LLM using these ratings as rewards in RLHF (See Section 4).
Knockout Tournament Inspired by the knockout tournament structure [40], we design a knockout tournament strategy for RRMs that organizes multiple candidates into a competition bracket. Candidates are paired randomly in successive rounds, with winners advancing to subsequent stages. In each pairwise comparison, RRMs determine a preferred response that will participate in the tournament in the next round. Given n candidates, this requires n -1 pairwise comparisons with O ( n ) complexity and O ( log ( n )) sequential rounds. Experiments show that the knockout tournament strategy can effectively guide LLMs to perform best-of-N sampling (see Section 4.3 and Appendix B.2).
Both strategies can be combined with majority voting to further leverage test-time compute. To integrate majority voting with the aforementioned strategies, we sample RRMs multiple times for each pairwise comparison. Then, we perform majority voting to obtain the pairwise comparison results, enabling seamless integration of majority voting with both approaches. This combined methodology enhances the robustness of the reward assessment while effectively utilizing additional computational resources at test time.
## 4 Experiments
We design our experiments that evaluate RRMs on both reward modeling benchmarks and practical applications, including reward-guided inference and LLM post-training. Additionally, we analyze how RRMs utilize additional test-time compute to achieve better performance and examine the reasoning patterns exhibited by RRM across multiple domains.
## 4.1 Training Details
Training Data Training RRMs require diverse pairwise preference data that covers various capabilities and aligns with human preference. In addition to preference pairs from Skywork-Reward [39], we further synthesize preference pairs from diverse data sources. We randomly sample 80K queries from the Tülu 3 prompt dataset [33], generate two responses for each using Deepseek-R1-Distill-Qwen1.5B [22], and annotate preference labels with GPT-4o [26]. Besides, we synthesize preferences pairs using verifiable question-answer pairs from WebInstruct-verified [42], Skywork-OR1 [24], Big-Math-RL [2], and DAPO-Math [72]. We prompt Deepseek-R1 distilled 1.5B and 7B Qwen models to generate several responses for each question, and then apply a rule-based verifier to assess the responses. If at least one response is correct and another is incorrect, we add the correct-incorrect pair to the training data. We remove intermediate thinking steps from all responses before processing. The final training dataset comprises approximately 420K preferences pairs: 80K each from SkyworkReward, Tülu 3, and our-synthesized data using Tülu 3 prompts, and 180K synthesized from other sources.
RRM Training We use DeepSeek-R1-Distill-Qwen models as the base models for RRMs in all the experiments. The training hyperparameters are detailed in Appendix A.2. The RRM training framework is implemented using the verl library [53], and we train both RRM-7B and RRM-32B models on AMD Instinct MI300X Accelerators. For RRM-32B, we employ a weighted mixture of datasets with a sampling ratio of 5:1:1:1 across Skywork-Reward, Tülu-80K, our GPT-4o-labeled preference pairs, and the other synthetic data. The RRM-7B model is trained on a similar dataset mixture using a 5:1:1 ratio of Skywork-Reward, Tülu-80K, and GPT-4o-labeled preference data.
## 4.2 Evaluating Agreement with Human Preference
## 4.2.1 Setup
Benchmarks We evaluate RRMs on widely-used benchmarks for reward modeling, namely RewardBench [34] and PandaLM Test [60]. (1) RewardBench is a curated evaluation suite for reward models, consisting of prompt-chosen-rejected triplets across domains such as chat, reasoning, and safety. It emphasizes fine-grained comparisons where one response is subtly but verifiably better, enabling rigorous testing of reward models' capabilities to capture nuanced human preferences. (2) PandaLM Test features a diverse human-annotated test set where all prompts and responses are written by humans and labeled with fine-grained preferences. Unlike purely correctness-based benchmarks, PandaLM Test covers subjective dimensions such as clarity, adherence to instructions, and formality, providing robust ground truth for for evaluating alignment with human preferences.
Table 1: Evaluation results on RewardBench benchmark and PandaLM Test. Bold numbers indicate the best performance, Underlined numbers indicate the second best.
| Models | RewardBench | RewardBench | RewardBench | RewardBench | RewardBench | PandaLM Test | PandaLM Test |
|----------------------------------------|---------------|---------------|---------------|---------------|---------------|----------------|----------------|
| | Chat | Chat Hard | Safety | Reasoning | Overall | Agreement | F1 |
| Skywork-Reward-Gemma-2-27B-v0.2 [34] | 96.1 | 89.9 | 93.0 | 98.1 | 94.3 | 76.6 | 76.4 |
| JudgeLM-7B [80] | 87.3 | 43.6 | 74.5 | 48.7 | 63.5 | 65.1 | 61.9 |
| JudgeLM-33B [80] | 92.7 | 54.2 | 85.8 | 58.3 | 72.3 | 75.2 | 69.7 |
| Claude-3.5-Sonnet-20240620 [34] | 96.4 | 74.0 | 81.6 | 84.7 | 84.2 | - | - |
| DeepSeek-R1 [41, 12] | 97.1 | 73,7 | 73.3 | 95.6 | 84.9 | 78.7 | 72.5 |
| DeepSeek-GRM-27B [41] | 94.1 | 78.3 | 88.0 | 83.8 | 86.0 | - | - |
| GPT-4-0125-preview [34] | 95.3 | 74.3 | 87.6 | 86.9 | 86.0 | 66.5 | 61.8 |
| GPT-4o-0806 [34] | 96.1 | 76.1 | 86.6 | 88.1 | 86.7 | - | - |
| RM-R1-DeepSeek-Distilled-Qwen-7B [14] | 88.9 | 66.2 | 78.4 | 87.0 | 80.1 | - | - |
| RM-R1-DeepSeek-Distilled-Qwen-14B [41] | 91.3 | 91.3 | 79.4 | 95.5 | 88.9 | - | - |
| RM-R1-DeepSeek-Distilled-Qwen-32B [41] | 95.3 | 80.3 | 91.1 | 96.8 | 90.9 | - | - |
| DirectJudge-7B | 86.0 | 69.7 | 85.5 | 79.5 | 80.2 | 70.3 | 70.2 |
| DirectJudge-32B | 96.1 | 85.1 | 89.5 | 90.9 | 90.4 | 76.7 | 77.4 |
| RRM-7B | 87.7 | 70.4 | 80.7 | 90.0 | 82.2 | 72.9 | 71.1 |
| RRM-7B (voting@16) | 92.1 | 71.5 | 81.3 | 93.8 | 84.8 | 75.9 | 77.8 |
| RRM-32B | 94.7 | 81.1 | 90.7 | 98.3 | 91.2 | 78.8 | 79.0 |
| RRM-32B (voting@16) | 96.1 | 81.4 | 91.6 | 98.6 | 91.9 | 80.2 | 81.9 |
Baselines We compare RRMs with the following baselines: (1) Skywork-Reward [39], a scalar reward model that uses a regression head to output numerical preference scores without explanations or reasoning traces, (2) Production-grade LLMs , including GPT-4o [26] and Claude 3.5 Sonnet [4], which are prompted in an LLM-as-a-judge [78] manner to determine the preferred response, (3) JudgeLM [80], which is trained to generate fine-grained reward scores along with explanations, using synthetic training data generated by GPT-4 [1], (4) DeepSeek-GRM [41] and RM-R1 [14], two concurrent approaches that also incorporate a reasoning phase prior to producing rewards.
In addition to these existing baselines, we introduce (5) DirectJudge , a pairwise judging model implemented using the same training data and base models as RRMs. DirectJudge models receive the same inputs as RRMs but are trained to directly generate judgment without explicit reasoning.
## 4.2.2 Results
Table 1 presents the evaluation results of baseline reward models and RRMs on the RewardBench benchmark and the PandaLM Test. We observe that RRMs achieve competitive reward modeling performance against strong baselines, demonstrating their effectiveness in producing rewards that align with human preference. Notably, RRM-32B attains an accuracy of 98.6 in the reasoning category of RewardBench. Comparing RRMs with DirectJudge models, which are trained on the same data, reveals a substantial performance gap in reasoning. This difference indicates that RRMs effectively leverage test-time compute, thereby enhancing performance on complex queries that benefit from deliberate reasoning processes.
## 4.3 Evaluating Reward-Guided Best-of-N Inference
## 4.3.1 Setup
Preference Proxy Evaluations Preference Proxy Evaluations (PPE) [18] is a benchmark designed to evaluate reward models through proxy tasks. Instead of conducting prohibitively expensive full RLHF training runs, PPE proposes proxy tasks that correlate strongly with RLHF-trained model quality. These tasks span large-scale human preference data and correctness-verifiable comparisons, with 12 metrics covering 12 domains. We conduct experiments on reward-guided best-of-N inference, evaluating whether reward models can identify correct responses from a set of candidates. Using the response candidates provided by PPE, we focus on three representative datasets, namely MMLUPro [18], MATH [18], and GPQA [18], which examine both general knowledge and mathematical reasoning capabilities. Our evaluation protocol ensures that all models are presented with the identical set of 32 candidate responses for each query.
Table 2: Evaluation results on reward-guided best-of-N inference. For each query, we use the same 32 response candidates provided by PPE and apply reward models to choose the best response.
| Models | MMLU-Pro | MATH | GPQA | Overall |
|---------------------------------|------------|--------|--------|-----------|
| Skywork-Reward-Gemma-2-27B-v0.2 | 67 | 56.3 | 44 | 55.8 |
| GPT-4o-0806 | 64.8 | 56.9 | 46.3 | 56 |
| RRM-7B | 69.1 | 82 | 49.2 | 66.8 |
| RRM-7B (voting@5) | 69.4 | 86.1 | 49 | 68.2 |
| RRM-32B | 81.3 | 89.8 | 61.1 | 77.4 |
| RRM-32B (voting@5) | 83 | 91.8 | 64.3 | 79.7 |
Table 3: Evaluation results on binary preference classification following the protocol from Frick et al. [18]. For each benchmark, we report accuracy over a single random permutation of paired responses.
| Models | MMLU-Pro | MATH | GPQA | Overall |
|--------------------------------------------|------------|--------|--------|-----------|
| Skywork-Reward-Gemma-2-27B [65] | 55 | 46.2 | 44.7 | 48.6 |
| Gemma-2-27B [41] | 66.2 | 66.4 | 51.9 | 61.5 |
| DeepSeek-GRM-27B (voting@32) [41] | 65.5 | 69.4 | 56 | 63.6 |
| DeepSeek-GRM-27B (MetaRM) (voting@32) [41] | 68.1 | 70 | 56.9 | 65 |
| Llama-3.1-8B-Instruct [65] | 56.3 | 62.9 | 51.4 | 56.9 |
| Llama-3.1-70B-Instruct [65] | 72.1 | 73.1 | 61.2 | 68.8 |
| J1-Llama-8B (SC@32) [65] | 67.5 | 76.6 | 55.7 | 66.7 |
| J1-Llama-70B (SC@32) [65] | 79.9 | 88.1 | 66.5 | 78.2 |
| RRM-7B | 66.5 | 88 | 57.9 | 70.3 |
| RRM-7B (voting@5) | 68.3 | 90.5 | 58.3 | 72.4 |
| RRM-32B | 80.5 | 94.3 | 67.4 | 80.7 |
| RRM-32B (voting@5) | 81.3 | 95.4 | 68.4 | 81.7 |
Baselines For the first experiment, we employ the knockout tournament rewarding strategy to identify the best-of-N responses. We compare our method against several strong baselines, including Skywork-Reward-Gemma-2 [54] and GPT-4o [26]. The prompt template for GPT-4o is detailed in Appendix A.1.
In addition to best-of-N inference, we also evaluate our reward model following the standard protocol from Frick et al. [18]. For this evaluation, we compare established baselines including J1-Llama [65], DeepSeek-GRM [41], Skywork-Reward-Gemma-2 [39], and various representative reward models from recent literature. Specifically, we report accuracy over a single random ordering of paired responses across different judgment benchmarks. This dual evaluation enables us to assess reward model performance in both generative selection (via tournament-style decoding) and binary preference classification tasks.
## 4.3.2 Results
Table 2 presents the evaluation results on reward-guided best-of-N inference. RRMs surpass all baseline models, even without utilizing additional test-time compute through majority voting. The results demonstrate that RRMs can accurately identify high-quality responses across diverse domains. Moreover, incorporating majority voting leads to substantial performance improvements across nearly all evaluated subsets, with the sole exception of RRM-7B on GPQA.
To further analyze the capabilities of RRMs across different domains, we provide detailed results on each subset of the MMLU-Pro and GPQA benchmarks. As illustrated in Appendix B.1, we compare RRMs against Skywork-Reward-Gemma-2-27B-v0.2 on each individual domain. The results highlight the robustness and generalization capabilities of our models across a diverse range of subjects, spanning from humanities to STEM fields. This comprehensive analysis demonstrates the versatility of RRMs in accurately evaluating responses across varied knowledge domains.
Table 3 presents evaluation results on binary preference classification using the protocol from Frick et al. [18]. RRMs maintain strong performance across all three benchmarks, consistently outperforming baseline reward models and instruction-tuned LLMs. Notably, RRM-32B achieves state-of-the-art
Figure 3: GPQA accuracy of using RRM for RL post-training.
<details>
<summary>Image 3 Details</summary>

### Visual Description
\n
## Line Chart: GPQA vs. RL Training Steps
### Overview
This image presents a line chart illustrating the relationship between GPQA scores and RL (Reinforcement Learning) training steps. The chart shows how the GPQA score changes as the number of RL training steps increases.
### Components/Axes
* **X-axis:** "RL training steps" ranging from 0 to 800, with gridlines at intervals of 100.
* **Y-axis:** "GPQA" ranging from 25 to 45, with gridlines at intervals of 5.
* **Data Series:** A single line representing GPQA scores over RL training steps. The line is colored a light coral (#F08080).
* **Data Points:** Marked with circular points along the line.
### Detailed Analysis
The line generally slopes upward, indicating a positive correlation between RL training steps and GPQA scores. However, there is a slight dip in the GPQA score between 200 and 400 RL training steps.
Here's a breakdown of the approximate data points:
* **0 RL training steps:** GPQA ≈ 21
* **200 RL training steps:** GPQA ≈ 38
* **400 RL training steps:** GPQA ≈ 36
* **600 RL training steps:** GPQA ≈ 39
* **800 RL training steps:** GPQA ≈ 41
The line starts at approximately 21 GPQA at 0 RL training steps. It rises sharply to around 38 GPQA at 200 steps. It then decreases slightly to around 36 GPQA at 400 steps, before increasing again to approximately 39 GPQA at 600 steps, and finally reaching around 41 GPQA at 800 steps.
### Key Observations
* The most significant increase in GPQA occurs within the first 200 RL training steps.
* There's a temporary decrease in GPQA between 200 and 400 RL training steps, suggesting a potential plateau or temporary setback in learning.
* The overall trend is positive, indicating that increasing RL training steps generally leads to improved GPQA scores.
### Interpretation
The chart suggests that reinforcement learning is effective in improving GPQA scores, but the learning process isn't perfectly linear. The initial rapid increase likely represents the model quickly learning basic patterns. The dip between 200 and 400 steps could indicate the model encountering more complex scenarios or needing to adjust its strategy. The subsequent increase suggests the model eventually overcomes these challenges and continues to improve. The data suggests that continued training beyond 800 steps might yield further improvements, but the rate of improvement may diminish. The chart demonstrates the iterative nature of reinforcement learning, where progress isn't always consistent but generally trends upward with continued training.
</details>
Figure 4: MMLU-Pro accuracy of using RRM for RL post-training.
<details>
<summary>Image 4 Details</summary>

### Visual Description
\n
## Line Chart: MMLU-Pro vs. RL Training Steps
### Overview
This image presents a line chart illustrating the relationship between "RL training steps" and "MMLU-Pro" scores. The chart shows how the MMLU-Pro score changes as the number of RL training steps increases.
### Components/Axes
* **X-axis:** "RL training steps" ranging from 0 to 800, with gridlines at intervals of 200.
* **Y-axis:** "MMLU-Pro" ranging from 50 to 57, with gridlines at intervals of 2.
* **Data Series:** A single line, colored in a light coral/salmon shade.
* **No Legend:** The chart does not have a separate legend, as there is only one data series.
### Detailed Analysis
The line representing MMLU-Pro exhibits a generally upward trend.
* At 0 RL training steps, the MMLU-Pro score is approximately 50.2.
* The line rises sharply between 0 and 200 RL training steps, reaching a value of approximately 54.8.
* Between 200 and 400 RL training steps, the line continues to increase, but at a slower rate, reaching approximately 55.4.
* From 400 to 600 RL training steps, the line plateaus and slightly decreases, reaching approximately 55.2.
* Finally, from 600 to 800 RL training steps, the line resumes an upward trend, reaching a final value of approximately 56.7.
### Key Observations
* The most significant increase in MMLU-Pro score occurs during the initial 200 RL training steps.
* There is a period of relative stagnation between 400 and 600 RL training steps.
* The final increase between 600 and 800 RL training steps suggests continued improvement with further training.
### Interpretation
The chart suggests a positive correlation between the number of RL training steps and the MMLU-Pro score. Initially, the model benefits significantly from each additional training step. However, as training progresses, the gains diminish, indicating a potential point of diminishing returns around 400-600 RL training steps. The final increase suggests that continued training beyond 600 steps can still yield improvements, although at a slower rate. This could indicate that the model is still learning and refining its performance, or that the MMLU-Pro metric is sensitive to further optimization. The plateau between 400 and 600 steps could be due to the model converging towards a local optimum, or the need for a different training strategy to overcome a performance barrier.
</details>
accuracy on MMLU-Pro, MATH, and GPQA, even when compared against significantly larger models such as J1-Llama-70B. Furthermore, incorporating majority voting (voting@5) further boosts performance, with RRM-32B (voting@5) reaching peak results across all benchmarks. These findings further validate the effectiveness of RRMs in classifying reason quality under diverse and challenging evaluation settings.
## 4.4 Post-Training with RRM Feedback
In addition to directly evaluating RRMs on reward model benchmarks, we further assess RRMs by post-training LLMs with reinforcement learning or direct preference optimization, supervised by the RRM-generated rewards. This approach allows the downstream performance of the post-trained LLMs to reflect the quality of the reward signals. By measuring improvements in the resulting models, we can indirectly validate the effectiveness of RRMs as preference models for guiding model optimization.
## 4.4.1 Reinforcement Learning with Unlabeled Data
We train Deepseek-R1-Distill-Qwen-7B on WebInstruct [42] queries using group relative policy optimization (GRPO) [52]. Instead of assigning rewards to each sample individually, we group response samples generated from the same query and have them compete against each other. In each group containing 8 responses, we construct 4 × 8 pairwise matches by randomly selecting 4 competitors for each response, and then obtain the pairwise preference results using RRM-32B. Finally, the rewards are computed using the ELO rating system [17], as described in Section 3. Notably, this approach utilizes only unlabeled queries without requiring any answers or reference responses.
Following the evaluation protocols established by Ma et al. [42], we evaluate the post-trained models on MMLU-Pro and GPQA using greedy decoding with a maximum response length of 8K tokens. As shown in Figure 3 and Figure 4, the downstream performance of the post-traineded models improves steadily throughout the training process. These results demonstrate that RRMs can effectively guide post-training with reinforcement learning, despite most prior work relying exclusively on scalar reward models. This underscores the practical viability of RRMs as a compelling alternative to traditional scalar reward models in post-training pipelines.
## 4.4.2 Direct Preference Optimization
To further explore the utility of RRMs in post-training pipelines, we apply Direct Preference Optimization (DPO) [49] on Qwen2.5-7B [48] using preference labels annotated by different reward models. Specifically, we construct preference datasets from Tülu [34] with 80K queries and responses, and obtain preference annotations from three different verifiers: RRM-7B, RRM-32B, and GPT-4o. Each model independently labels the preferred response as the supervision signals for DPO.
The trained models are evaluated on the Arena-Hard benchmark [36], which contains challenging instructions designed to test comprehensive model capabilities. As shown in Table 4, all post-trained models outperform the original Qwen2.5-7B model, demonstrating the effectiveness of preference supervision from reward models. Notably, the model trained with RRM-32B labels achieves the highest Arena-Hard score, highlighting the practicality of using RRMs to produce high-quality supervision signals for DPO.
Table 4: Performance of DPO post-trained Qwen2.5-7B models on Arena-Hard.
| Arena-Hard Score | Arena-Hard Score | CI |
|-----------------------------------------------------|-----------------------------------------------------|-----------------------------------------------------|
| Before Post-Training | Before Post-Training | |
| Base Model | 18.3 | (-1.61, +1.66) |
| DPO with Preference Data Annotated by Reward Models | DPO with Preference Data Annotated by Reward Models | DPO with Preference Data Annotated by Reward Models |
| GPT-4o | 51.9 | (-2.96, +2.93) |
| RRM-7B | 53.8 | (-1.72, +1.85) |
| RRM-32B | 55.4 | (-2.60, +2.67) |
## 4.5 Scaling Test-Time Compute
## 4.5.1 Parallel Scaling
We conduct parallel test-time compute scaling experiments on MATH [25] reasoning candidate responses. We use Qwen2.5-Math-7B-Instruct [70] to generate 8 candidate responses for each question, and then employ RRMs to perform reward-guided best-of-N inference. This experimental setup allows us to systematically study the scaling behaviors of RRMs under increased test-time computational resources.
Scaling Properties As illustrated in Figure 5, increasing the number of pairwise comparisons steadily improves best-of-N performance on MATH for both RRM-7B and RRM-32B. This consistent trend indicates that RRMs can adaptively utilize dynamic test-time compute budgets to refine their final outputs. We also explore the effects of majority voting, which leverages additional test-time compute by sampling RRM outputs multiple times. Table 5 compares the performance on MATH, where RRMs are prompted on each comparison pair either a single time or eight times, with the latter followed by majority voting. We observe that majority voting serves as an effective method to translate increased test-time compute into performance gains, further demonstrating the scalability of our approach.
Figure 5: MATH accuracy with varying number of pairwise comparisons.
<details>
<summary>Image 5 Details</summary>

### Visual Description
\n
## Line Chart: Math Performance vs. Number of Pairs
### Overview
This image presents a line chart illustrating the relationship between "Number of pairs" and "MATH" performance for two models, "RRM-7B" and "RRM-32B". The x-axis represents the number of pairs on a logarithmic scale, while the y-axis represents the MATH score.
### Components/Axes
* **X-axis Title:** "Number of pairs"
* **X-axis Scale:** Logarithmic, ranging from approximately 10<sup>1</sup> to 10<sup>2</sup>.
* **Y-axis Title:** "MATH"
* **Y-axis Scale:** Linear, ranging from approximately 86 to 91.
* **Legend:** Located in the bottom-right corner.
* **RRM-7B:** Represented by a red line with circular markers.
* **RRM-32B:** Represented by a blue line with triangular markers.
### Detailed Analysis
**RRM-7B (Red Line):**
The red line representing RRM-7B shows an upward trend, starting at approximately 86.5 when the number of pairs is 10<sup>1</sup>. It increases to approximately 88.8 at around 50 pairs, and then plateaus, reaching approximately 89.2 at 10<sup>2</sup> pairs.
* Number of pairs = 10<sup>1</sup>: MATH ≈ 86.5
* Number of pairs ≈ 50: MATH ≈ 88.8
* Number of pairs = 10<sup>2</sup>: MATH ≈ 89.2
**RRM-32B (Blue Line):**
The blue line representing RRM-32B also exhibits an upward trend, but is consistently higher than RRM-7B. It begins at approximately 88.4 when the number of pairs is 10<sup>1</sup>. It increases sharply to approximately 90.5 at around 50 pairs, and then levels off, reaching approximately 90.7 at 10<sup>2</sup> pairs.
* Number of pairs = 10<sup>1</sup>: MATH ≈ 88.4
* Number of pairs ≈ 50: MATH ≈ 90.5
* Number of pairs = 10<sup>2</sup>: MATH ≈ 90.7
### Key Observations
* RRM-32B consistently outperforms RRM-7B across all tested numbers of pairs.
* Both models show diminishing returns as the number of pairs increases beyond approximately 50. The increase in MATH score becomes smaller with more pairs.
* The performance of RRM-32B increases more rapidly than RRM-7B between 10<sup>1</sup> and 50 pairs.
### Interpretation
The data suggests that increasing the number of pairs used in training or evaluation improves the MATH performance of both models, but the benefit diminishes as the number of pairs grows. The larger model, RRM-32B, demonstrates superior performance compared to the smaller model, RRM-7B, indicating that model size is a significant factor in achieving higher MATH scores. The plateauing effect observed at higher numbers of pairs suggests that other factors, such as model architecture or training data quality, may become more important limiting factors once a certain level of data exposure is reached. The logarithmic scale on the x-axis emphasizes the diminishing returns; adding more pairs has a smaller impact on performance as the number of pairs increases. This could be due to the models reaching a point of saturation where they have learned the underlying patterns in the data and further exposure provides little additional benefit.
</details>
Comparing Rewarding Strategies Table 5 compares the scoring strategies, specifically using RRMs to evaluate candidates through either knockout tournament or ELO rating systems. Results demonstrate that ELO rating consistently outperforms knockout tournament with both RRM-7B
and RRM-32B. Nonetheless, the knockout tournament yields only slightly lower performance while requiring fewer computational resources-only O ( n ) comparisons. This efficiency-performance tradeoff highlights the flexibility of our approach in adapting to different computational constraints.
Table 5: Comparison of scoring strategies using RRM verifiers. ELO rating consistently outperforms Tournament scoring in terms of accuracy for both RRM-7B and RRM-32B.
| | RRM-7B | RRM-7B | RRM-32B | RRM-32B |
|-----------------|----------|----------|-----------|-----------|
| Majority Voting | No | Yes | No | Yes |
| Tournament | 88.2 | 88.7 | 90.0 | 90.4 |
| ELO rating | 88.5 | 88.8 | 90.3 | 90.5 |
## 4.5.2 Sequential Scaling
We study the impact of enabling longer chains of thought [64] before finalizing an answer. We evaluate RRMs on RewardBench, where we control the thinking budgets by setting a maximum token limit. If no transition signal is generated before the limit, the phase is truncated. We also set a small post-thinking budget to prevent compute hacking, i.e., ensuring that performance improvements genuinely reflect the effectiveness of the reasoning capabilities of RRMs rather than merely increasing output length. The detailed design of the post-thinking budget can be found in Appendix C.
Results Experiments on 7B, 14B, and 32B RRMs show that longer thinking horizons consistently improve output accuracy across all model sizes (Figure 6). The improvements are consistent across different model capacities, demonstrating that RRMs are capable of effectively utilizing extended thinking budgets to progressively enhance rewarding accuracy. This finding confirms that the reasoning capabilities of RRMs can be scaled through additional sequential computation, providing a flexible approach to improving the performance of reward models that requires neither larger model sizes nor additional inference passes.
## 4.6 Scaling RRM Training Compute
We investigate how model size and training duration affect the performance of RRMs, exploring the scaling properties of our reward reasoning approach across different compute dimensions. Figure 6 compares RRMs with model sizes of 7B, 14B, and 32B on RewardBench, showing consistent performance gains with increased model size.
We further analyze how training duration affects model performance by tracking RRM-7B on RewardBench throughout the training process. Figure 7 illustrates the performance trajectory across different evaluation domains. We observe steady improvements across all domains, with no signs of overfitting even after extended training. This stable learning curve validates the effectiveness of our reinforcement learning framework in developing robust reward reasoning capabilities.
Figure 6: Results on RewardBench varying thinking budgets.
<details>
<summary>Image 6 Details</summary>

### Visual Description
\n
## Line Chart: RewardBench Performance vs. Thinking Budget
### Overview
This line chart illustrates the relationship between "Thinking budget" (in tokens) and "RewardBench" performance (in percentage) for three different model sizes: 32B, 14B, and 7B. The chart shows how performance changes as the thinking budget increases.
### Components/Axes
* **X-axis:** "Thinking budget (tokens)" with markers at 1e3 (1000), 2e3 (2000), 4e3 (4000), and 8e3 (8000).
* **Y-axis:** "RewardBench (%)" with a scale ranging from approximately 75% to 92%.
* **Legend:** Located in the bottom-right corner, identifying the three data series:
* Blue circle: 32B
* Orange triangle: 14B
* Green square: 7B
### Detailed Analysis
* **32B (Blue Line):** The blue line shows an upward trend, starting at approximately 85% at 1e3 tokens. It rises to around 90% at 2e3 tokens, plateaus slightly, and reaches approximately 91% at 8e3 tokens.
* 1e3 tokens: ~85%
* 2e3 tokens: ~90%
* 4e3 tokens: ~90.5%
* 8e3 tokens: ~91%
* **14B (Orange Line):** The orange line also exhibits an upward trend, beginning at approximately 82% at 1e3 tokens. It increases sharply to around 89% at 2e3 tokens, then plateaus, remaining at approximately 89% through 8e3 tokens.
* 1e3 tokens: ~82%
* 2e3 tokens: ~89%
* 4e3 tokens: ~89%
* 8e3 tokens: ~89%
* **7B (Green Line):** The green line shows a consistent upward trend, starting at approximately 75% at 1e3 tokens. It rises to around 81% at 2e3 tokens, continues to approximately 84% at 4e3 tokens, and reaches approximately 84% at 8e3 tokens.
* 1e3 tokens: ~75%
* 2e3 tokens: ~81%
* 4e3 tokens: ~84%
* 8e3 tokens: ~84%
### Key Observations
* The 32B model consistently outperforms the 14B and 7B models across all thinking budget levels.
* The 14B model shows a significant performance increase between 1e3 and 2e3 tokens, but then plateaus.
* The 7B model exhibits the lowest performance but demonstrates a steady improvement with increasing thinking budget.
* All models show diminishing returns in performance as the thinking budget increases beyond 2e3 tokens.
### Interpretation
The data suggests that increasing the thinking budget generally improves the performance of these models on the RewardBench benchmark. However, the benefit of increasing the thinking budget diminishes as it grows larger. The 32B model benefits the most from a larger thinking budget, achieving the highest performance levels. The 7B model, while starting with lower performance, still shows a positive correlation between thinking budget and RewardBench score. This indicates that even smaller models can benefit from increased computational resources for reasoning tasks. The plateauing of the 14B model suggests that its performance is limited by other factors beyond the thinking budget, such as model capacity or training data. The differences in performance between the models highlight the importance of model size in achieving high performance on complex reasoning tasks.
</details>
Figure 7: Results on RewardBench throughout RRM7B training.
<details>
<summary>Image 7 Details</summary>

### Visual Description
\n
## Line Chart: RewardBench Performance Over Training Steps
### Overview
This line chart depicts the performance of a model across several metrics (Chat, Chat Hard, Safety, Reasoning, and Overall) on the RewardBench, measured as a percentage, over 600 training steps. The chart visualizes how these metrics evolve during the training process.
### Components/Axes
* **X-axis:** Training steps, ranging from 0 to 600.
* **Y-axis:** RewardBench (%), ranging from approximately 60% to 95%.
* **Legend (Top-Right):**
* Blue Line: Chat
* Orange Line: Chat Hard
* Green Line: Safety
* Light Orange Line: Reasoning
* Red Line: Overall
* **Gridlines:** Vertical gridlines are present to aid in reading values along the x-axis.
### Detailed Analysis
The chart displays five distinct lines, each representing a different metric.
* **Chat (Blue Line):** Starts at approximately 83% and fluctuates, reaching a peak of around 93% at approximately 550 training steps, before decreasing slightly to around 91% at 600 steps. The line generally trends upwards, with some oscillations.
* **Chat Hard (Orange Line):** Begins at approximately 61% and exhibits significant fluctuations throughout the training process. It reaches a peak of around 75% at approximately 350 training steps, then declines to around 68% at 450 steps, and recovers to approximately 70% at 600 steps. The line shows a generally increasing trend, but with substantial variability.
* **Safety (Green Line):** Starts at approximately 74% and remains relatively stable, fluctuating between approximately 76% and 82%. It shows a slight upward trend overall.
* **Reasoning (Light Orange Line):** Starts at approximately 85% and fluctuates, reaching a peak of around 92% at approximately 200 training steps, then decreasing to around 88% at 400 steps, and recovering to approximately 90% at 600 steps. The line generally trends upwards, with some oscillations.
* **Overall (Red Line):** Starts at approximately 81% and generally increases, reaching a peak of around 88% at approximately 500 training steps, before decreasing slightly to around 86% at 600 steps. The line shows a consistent upward trend.
### Key Observations
* The "Chat" and "Reasoning" metrics consistently achieve the highest RewardBench scores, generally above 85%.
* "Chat Hard" consistently has the lowest RewardBench scores, remaining below 75% throughout the training process.
* "Safety" shows the most stable performance, with minimal fluctuations.
* All metrics demonstrate an overall positive trend, indicating improvement with increasing training steps.
* The "Chat Hard" metric exhibits the most volatility, suggesting it is more sensitive to training variations.
### Interpretation
The data suggests that the model performs well on "Chat" and "Reasoning" tasks, but struggles with "Chat Hard" tasks. The stability of the "Safety" metric indicates that the model maintains a consistent level of safety throughout training. The overall upward trend across all metrics suggests that the training process is effective in improving the model's performance. The large fluctuations in "Chat Hard" could indicate that this task is more complex or requires more specialized training data. The divergence between "Chat" and "Chat Hard" suggests that the model is better at handling simpler chat interactions than more challenging ones. The consistent improvement in "Overall" suggests that the model is learning to generalize its performance across different tasks. The RewardBench metric appears to be a useful indicator of model performance, as it correlates with the observed trends in the individual metrics.
</details>
## 4.7 Reward Reasoning Pattern Analysis
Following Wang et al. [61] and Chen et al. [13], we analyze the reasoning patterns of RRM-32B by statistically measuring the proportion of model responses containing keywords such as 'wait' and 'alternatively'. We categorize the reasoning patterns into four categories: transition (switching perspectives or strategies), reflection (self-checking or revisiting earlier steps), comparison (evaluating multiple options), and breakdown (decomposing the problem).
As illustrated in Figure 8, compared to the Deepseek-R1-Distill-Qwen-32B model, RRM-32B demonstrates a greater overall utilization of reasoning patterns when judging the superiority of two answers, particularly in analyzing from different perspectives and conducting in-depth comparisons. In contrast, the Deepseek-R1-Distill-Qwen-32B model employs the breakdown pattern more frequently, suggesting a greater tendency to approach problems directly when making judgments, but less inclination to compare the merits of the two answers and engage in self-examination. This distinction in reasoning patterns highlights how our Reward Reasoning via Reinforcement Learning framework shapes the model's approach to evaluation tasks.
Figure 8: Reward reasoning pattern analysis results. Compared to DeepSeek-R1-Distilled-Qwen-32B, RRM-32B exhibits more transition patterns (40.63% vs. 33.73%), reflection patterns (63.28% vs. 52.75%), and comparison patterns (89.84% vs. 85.29%), but fewer direct problem decomposition (8.40% vs. 16.86%).
<details>
<summary>Image 8 Details</summary>

### Visual Description
\n
## Bar Chart: Rhetorical Relation Analysis
### Overview
This is a bar chart comparing the percentage of examples identified as belonging to different rhetorical relations (Transition, Reflection, Comparison, Breakdown) using two different methods: R1-distilled and RRM. The y-axis represents the percentage of examples, and the x-axis represents the rhetorical relation categories.
### Components/Axes
* **X-axis Title:** Rhetorical Relation
* **Y-axis Title:** Percentage of Examples (%)
* **Legend:** Located in the top-left corner.
* R1-distilled (represented by a diagonally striped pattern)
* RRM (represented by a solid color)
* **Categories (X-axis):** Transition, Reflection, Comparison, Breakdown.
* **Y-axis Scale:** 0 to 100, with increments of 20.
### Detailed Analysis
The chart displays two bars for each rhetorical relation category, one for R1-distilled and one for RRM.
* **Transition:**
* R1-distilled: Approximately 34% (visually estimated).
* RRM: Approximately 38% (visually estimated).
* **Reflection:**
* R1-distilled: Approximately 42% (visually estimated).
* RRM: Approximately 40% (visually estimated).
* **Comparison:**
* R1-distilled: Approximately 86% (visually estimated).
* RRM: Approximately 89% (visually estimated).
* **Breakdown:**
* R1-distilled: Approximately 18% (visually estimated).
* RRM: Approximately 12% (visually estimated).
**Trends:**
* For Transition, RRM shows a slightly higher percentage than R1-distilled.
* For Reflection, R1-distilled shows a slightly higher percentage than RRM.
* For Comparison, both methods show high percentages, with RRM being slightly higher.
* For Breakdown, R1-distilled shows a significantly higher percentage than RRM.
### Key Observations
* The largest difference between the two methods is observed for the "Breakdown" category, where R1-distilled identifies a much higher percentage of examples than RRM.
* Both methods perform similarly well in identifying "Comparison" examples, with both exceeding 85%.
* The percentages for "Transition" and "Reflection" are relatively similar between the two methods.
### Interpretation
The data suggests that the two methods, R1-distilled and RRM, have different sensitivities in identifying rhetorical relations. R1-distilled appears to be more sensitive to "Breakdown" relations, while RRM performs slightly better on "Transition" and "Comparison". The high agreement on "Comparison" suggests that this relation is relatively easy to identify for both methods. The discrepancy in "Breakdown" could indicate that RRM may require more explicit cues to identify this relation, or that R1-distilled is more prone to false positives in this category. Further investigation into the specific examples classified as "Breakdown" by each method would be necessary to understand the underlying reasons for this difference. The chart demonstrates a comparison of two methods for rhetorical relation classification, highlighting their strengths and weaknesses in identifying different types of relations.
</details>
Table 6 presents case studies by comparing the outputs of Deepseek-R1-Distill-Qwen-32B and RRM-32B. Within the responses to a single example, RRM-32B exhibits a stronger propensity to iteratively apply reasoning patterns for more thorough and in-depth comparisons. The reasoning process of the Deepseek-R1-Distill-Qwen-32B model tends to be simple and monolithic-involving separate analyses of the two answers followed by a summary. Conversely, the thinking process of RRM-32B involves multiple iterations of self-reflection, comparative analysis, and consideration of various perspectives, incorporating more detailed analysis and interpretation of the instructions. Consequently, RRM-32B demonstrates an enhanced ability to follow instructions and provide more accurate comparative judgments. This iterative approach to reasoning appears to be a key factor in the superior performance of RRMs across diverse evaluation tasks.
Table 6: Case study of reasoning patterns of RRM-32B and DeepSeek-R1-Distilled-Qwen-32B.
| DeepSeek-R1-Distilled-Qwen-32B | RRM-32B |
|---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| Okay, I need to evaluate which assistant provided a bet- ter response to the given problem...Looking at the two responses: Assistant 1's code starts by... Assistant 2's solution uses a bitwise AND approach.... In conclusion, while Assistant 1's code has some issues, Assistant 2's solution is fundamentally incorrect because the bitwise approach doesn't apply to powers of three. Therefore, Assistant 1's response is better, despite its flaws, because it attempts to handle the problem correctly, even if it's not perfectly implemented. </think> \boxed{Assistant 1} | Okay, so I need to figure out which assistant is better for the given problem... Looking at Assistant 1's response... Now, looking at Assistant 2's response... Wait, that's a problem.... But then, what about the code from Assistant 1?... Let me test it with some examples... Wait, perhaps... But neither assistant provided that solution. So, given the two options, which is better?... Alternatively, perhaps I'm missing something. Let me think again. Wait, the problem requires a solution without loops or recursion... Alternatively... Looking back, Assistant 1's code is ... Assistant 2's code is ... Alternatively, perhaps I'm missing something. Let me think again. Wait, ... Wait, the user's instruction says: "Do NOT say both / neither are good." So I have to choose between Assistant 1 or 2. Given that, I think Assistant 2's code is better because ... </think> The better response is... |
| Alright, let's evaluate the problem step by step... First, I look at Assistant 1's response... Now, Assistant 2's response is much simpler... Therefore, Assistant 1 correctly addressed the exponentia- tion but made a minor error in the final subtraction... </think> The correct evaluation requires...\boxed{Assistant 1} | Okay, so I need to figure out which assistant is better for the given query. ... Assistant 1 starts by ... Assistant 2, on the other hand, ... Looking at the instruction, the user wrote... However, the problem is that the instruction might be am- biguous. If ..., then Assistant 2 is correct. But if ..., then Assistant 1 is correct. Since the instruction is...So I think the intended operation is exponentiation. Therefore, Assistant 1's approach is more accurate for the likely intended problem. However, if the problem was meant to be multiplication, then Assistant 2 is correct. But given the notation, exponentiation is more probable. So blue Assistant 1's answer is better in that context. But wait, the user's instruction is... Hmm, this is a bit confusing. I need to decide based on the instruction. ... But if ... Given that, Assistant 2's interpretation is more likely cor- rect... So, considering the instruction's ambiguity, but leaning towards multiplication, Assistant 2 is better. </think> The instruction is ambiguous, but Assistant 2 correctly interpreted it as multiplication, providing the accurate re- sult. \boxed{Assistant 2} |
## 5 Conclusion
In this paper, we introduced Reward Reasoning Models (RRMs), incorporating explicit reasoning processes prior to reward assignment. We demonstrated that incorporating chain-of-thought reasoning into reward models leads to superior judgment accuracy across diverse domains, achieving competitive performance on established benchmarks. Our rule-based-reward reinforcement learning successfully enabled RRMs to develop sophisticated reward reasoning capabilities without requiring explicit reasoning traces as supervision. Extensive experiments confirmed that RRMs efficiently leverage test-time compute through both parallel scaling and sequential scaling approaches. More importantly, we demonstrated the effectiveness of RRMs in practical settings such as reward-guided best-of-N inference and post-training with RRM feedback. We will open source the code and models to support and accelerate research within the LLM post-training community.
## References
- [1] Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al. GPT-4 technical report. arXiv preprint arXiv:2303.08774 , 2023.
- [2] Alon Albalak, Duy Phung, Nathan Lile, Rafael Rafailov, Kanishk Gandhi, Louis Castricato, Anikait Singh, Chase Blagden, Violet Xiang, Dakota Mahan, and Nick Haber. Big-math: A large-scale, high-quality math dataset for reinforcement learning in language models, 2025. URL https://arxiv.org/abs/2502.17387 .
- [3] Andrei Alexandru, Antonia Calvi, Henry Broomfield, Jackson Golden, Kyle Dai, Mathias Leys, Maurice Burger, Max Bartolo, Roman Engeler, Sashank Pisupati, et al. Atla selene mini: A general purpose evaluation model. arXiv preprint arXiv:2501.17195 , 2025.
- [4] AI Anthropic. The claude 3 model family: Opus, sonnet, haiku. Claude-3 Model Card , 1:1, 2024.
- [5] Negar Arabzadeh, Siqing Huo, Nikhil Mehta, Qingyun Wu, Chi Wang, Ahmed Hassan Awadallah, Charles L. A. Clarke, and Julia Kiseleva. Assessing and verifying task utility in LLM-powered applications. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen, editors, Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing , pages 21868-21888, Miami, Florida, USA, November 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.emnlp-main.1219. URL https: //aclanthology.org/2024.emnlp-main.1219/ .
- [6] Yuntao Bai, Andy Jones, Kamal Ndousse, Amanda Askell, Anna Chen, Nova DasSarma, Dawn Drain, Stanislav Fort, Deep Ganguli, Tom Henighan, et al. Training a helpful and harmless assistant with reinforcement learning from human feedback. arXiv preprint arXiv:2204.05862 , 2022.
- [7] Yuntao Bai, Saurav Kadavath, Sandipan Kundu, Amanda Askell, Jackson Kernion, Andy Jones, Anna Chen, Anna Goldie, Azalia Mirhoseini, Cameron McKinnon, et al. Constitutional ai: Harmlessness from ai feedback. arXiv preprint arXiv:2212.08073 , 2022.
- [8] Bradley Brown, Jordan Juravsky, Ryan Ehrlich, Ronald Clark, Quoc V Le, Christopher Ré, and Azalia Mirhoseini. Large language monkeys: Scaling inference compute with repeated sampling. arXiv preprint arXiv:2407.21787 , 2024.
- [9] Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Language models are few-shot learners. Advances in neural information processing systems , 33:1877-1901, 2020.
- [10] Sébastien Bubeck, Varun Chadrasekaran, Ronen Eldan, Johannes Gehrke, Eric Horvitz, Ece Kamar, Peter Lee, Yin Tat Lee, Yuanzhi Li, Scott Lundberg, et al. Sparks of artificial general intelligence: Early experiments with GPT-4, 2023.
- [11] Maosong Cao, Alexander Lam, Haodong Duan, Hongwei Liu, Songyang Zhang, and Kai Chen. Compassjudger-1: All-in-one judge model helps model evaluation and evolution. arXiv preprint arXiv:2410.16256 , 2024.
- [12] Nuo Chen, Zhiyuan Hu, Qingyun Zou, Jiaying Wu, Qian Wang, Bryan Hooi, and Bingsheng He. Judgelrm: Large reasoning models as a judge. arXiv preprint arXiv:2504.00050 , 2025.
- [13] Runjin Chen, Zhenyu (Allen) Zhang, Junyuan Hong, Souvik Kundu, and Zhangyang Wang. Seal: Steerable reasoning calibration of large language models for free. 2025. URL https: //api.semanticscholar.org/CorpusID:277741244 .
- [14] Xiusi Chen, Gaotang Li, Ziqi Wang, Bowen Jin, Cheng Qian, Yu Wang, Hongru Wang, Yu Zhang, Denghui Zhang, Tong Zhang, et al. Rm-r1: Reward modeling as reasoning. arXiv preprint arXiv:2505.02387 , 2025.
- [15] Paul F. Christiano, Jan Leike, Tom B. Brown, Miljan Martic, Shane Legg, and Dario Amodei. Deep reinforcement learning from human preferences. In Proceedings of the 31st International Conference on Neural Information Processing Systems , NIPS'17, page 4302-4310, Red Hook, NY, USA, 2017. Curran Associates Inc. ISBN 9781510860964.
- [16] Karl Cobbe, Vineet Kosaraju, Mohammad Bavarian, Mark Chen, Heewoo Jun, Lukasz Kaiser, Matthias Plappert, Jerry Tworek, Jacob Hilton, Reiichiro Nakano, et al. Training verifiers to solve math word problems. arXiv preprint arXiv:2110.14168 , 2021.
- [17] Arpad E. Elo. The Rating of Chessplayers, Past and Present . Arco Pub., New York, 1978. ISBN 0668047216. URL http://www.amazon.com/Rating-Chess-Players-Past-Present/ dp/0668047216 .
- [18] Evan Frick, Tianle Li, Connor Chen, Wei-Lin Chiang, Anastasios Nikolas Angelopoulos, Jiantao Jiao, Banghua Zhu, Joseph E. Gonzalez, and Ion Stoica. How to evaluate reward models for RLHF. In The Thirteenth International Conference on Learning Representations , 2025. URL https://openreview.net/forum?id=cbttLtO94Q .
- [19] Jiaxuan Gao, Shusheng Xu, Wenjie Ye, Weilin Liu, Chuyi He, Wei Fu, Zhiyu Mei, Guangju Wang, and Yi Wu. On designing effective rl reward at training time for llm reasoning. arXiv preprint arXiv:2410.15115 , 2024.
- [20] Leo Gao, John Schulman, and Jacob Hilton. Scaling laws for reward model overoptimization. In Proceedings of the 40th International Conference on Machine Learning , ICML'23. JMLR.org, 2023.
- [21] Jiawei Gu, Xuhui Jiang, Zhichao Shi, Hexiang Tan, Xuehao Zhai, Chengjin Xu, Wei Li, Yinghan Shen, Shengjie Ma, Honghao Liu, et al. A survey on llm-as-a-judge. arXiv preprint arXiv:2411.15594 , 2024.
- [22] Daya Guo, Dejian Yang, Haowei Zhang, Junxiao Song, Ruoyu Zhang, Runxin Xu, Qihao Zhu, Shirong Ma, Peiyi Wang, Xiao Bi, et al. Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning. arXiv preprint arXiv:2501.12948 , 2025.
- [23] Fang Guo, Wenyu Li, Honglei Zhuang, Yun Luo, Yafu Li, Le Yan, Qi Zhu, and Yue Zhang. Mcranker: Generating diverse criteria on-the-fly to improve pointwise llm rankers. In Proceedings of the Eighteenth ACM International Conference on Web Search and Data Mining , WSDM '25, page 944-953, New York, NY, USA, 2025. Association for Computing Machinery. ISBN 9798400713293. doi: 10.1145/3701551.3703583. URL https: //doi.org/10.1145/3701551.3703583 .
- [24] Jujie He, Jiacai Liu, Chris Yuhao Liu, Rui Yan, Chaojie Wang, Peng Cheng, Xiaoyu Zhang, Fuxiang Zhang, Jiacheng Xu, Wei Shen, Siyuan Li, Liang Zeng, Tianwen Wei, Cheng Cheng, Yang Liu, and Yahui Zhou. Skywork open reasoner series. https://capricious-hydrogen-41c.notion.site/ Skywork-Open-Reaonser-Series-1d0bc9ae823a80459b46c149e4f51680 , 2025. Notion Blog.
- [25] Dan Hendrycks, Collin Burns, Saurav Kadavath, Akul Arora, Steven Basart, Eric Tang, Dawn Song, and Jacob Steinhardt. Measuring mathematical problem solving with the MATH dataset. In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2) , 2021. URL https://openreview.net/forum?id=7Bywt2mQsCe .
- [26] Aaron Hurst, Adam Lerer, Adam P Goucher, Adam Perelman, Aditya Ramesh, Aidan Clark, AJ Ostrow, Akila Welihinda, Alan Hayes, Alec Radford, et al. Gpt-4o system card. arXiv preprint arXiv:2410.21276 , 2024.
- [27] Aaron Jaech, Adam Kalai, Adam Lerer, Adam Richardson, Ahmed El-Kishky, Aiden Low, Alec Helyar, Aleksander Madry, Alex Beutel, Alex Carney, et al. Openai o1 system card. arXiv preprint arXiv:2412.16720 , 2024.
- [28] Dongfu Jiang, Xiang Ren, and Bill Yuchen Lin. LLM-blender: Ensembling large language models with pairwise ranking and generative fusion. In Anna Rogers, Jordan Boyd-Graber, and Naoaki Okazaki, editors, Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) , pages 14165-14178, Toronto, Canada, July 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023.acl-long.792. URL https://aclanthology.org/2023.acl-long.792/ .
- [29] Carlos E Jimenez, John Yang, Alexander Wettig, Shunyu Yao, Kexin Pei, Ofir Press, and Karthik R Narasimhan. SWE-bench: Can language models resolve real-world github issues? In The Twelfth International Conference on Learning Representations , 2024. URL https: //openreview.net/forum?id=VTF8yNQM66 .
- [30] Jared Kaplan, Sam McCandlish, Tom Henighan, Tom B Brown, Benjamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeffrey Wu, and Dario Amodei. Scaling laws for neural language models. arXiv preprint arXiv:2001.08361 , 2020.
- [31] Muhammad Uzair Khattak, Hanoona Rasheed, Muhammad Maaz, Salman Khan, and Fahad Shahbaz Khan. Maple: Multi-modal prompt learning. In 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) , pages 19113-19122, 2023. doi: 10.1109/CVPR52729.2023.01832.
- [32] Seungone Kim, Juyoung Suk, Shayne Longpre, Bill Yuchen Lin, Jamin Shin, Sean Welleck, Graham Neubig, Moontae Lee, Kyungjae Lee, and Minjoon Seo. Prometheus 2: An open source language model specialized in evaluating other language models. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen, editors, Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing , pages 4334-4353, Miami, Florida, USA, November 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.emnlp-main.248. URL https://aclanthology.org/2024.emnlp-main.248/ .
- [33] Nathan Lambert, Jacob Morrison, Valentina Pyatkin, Shengyi Huang, Hamish Ivison, Faeze Brahman, Lester James V Miranda, Alisa Liu, Nouha Dziri, Shane Lyu, et al. T \ " ulu 3: Pushing frontiers in open language model post-training. arXiv preprint arXiv:2411.15124 , 2024.
- [34] Nathan Lambert, Valentina Pyatkin, Jacob Morrison, LJ Miranda, Bill Yuchen Lin, Khyathi Chandu, Nouha Dziri, Sachin Kumar, Tom Zick, Yejin Choi, et al. Rewardbench: Evaluating reward models for language modeling. arXiv preprint arXiv:2403.13787 , 2024.
- [35] Junlong Li, Shichao Sun, Weizhe Yuan, Run-Ze Fan, hai zhao, and Pengfei Liu. Generative judge for evaluating alignment. In The Twelfth International Conference on Learning Representations , 2024. URL https://openreview.net/forum?id=gtkFw6sZGS .
- [36] Tianle Li, Wei-Lin Chiang, Evan Frick, Lisa Dunlap, Tianhao Wu, Banghua Zhu, Joseph E. Gonzalez, and Ion Stoica. From crowdsourced data to high-quality benchmarks: Arena-hard and benchbuilder pipeline, 2024. URL https://arxiv.org/abs/2406.11939 .
- [37] Hunter Lightman, Vineet Kosaraju, Yuri Burda, Harrison Edwards, Bowen Baker, Teddy Lee, Jan Leike, John Schulman, Ilya Sutskever, and Karl Cobbe. Let's verify step by step. In The Twelfth International Conference on Learning Representations , 2024. URL https: //openreview.net/forum?id=v8L0pN6EOi .
- [38] Chin-Yew Lin. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out , pages 74-81, 2004.
- [39] Chris Yuhao Liu, Liang Zeng, Jiacai Liu, Rui Yan, Jujie He, Chaojie Wang, Shuicheng Yan, Yang Liu, and Yahui Zhou. Skywork-reward: Bag of tricks for reward modeling in llms. arXiv preprint arXiv:2410.18451 , 2024.
- [40] Yantao Liu, Zijun Yao, Rui Min, Yixin Cao, Lei Hou, and Juanzi Li. Pairwise rm: Perform best-of-n sampling with knockout tournament. arXiv preprint arXiv:2501.13007 , 2025.
- [41] Zijun Liu, Peiyi Wang, Runxin Xu, Shirong Ma, Chong Ruan, Peng Li, Yang Liu, and Yu Wu. Inference-time scaling for generalist reward modeling. arXiv preprint arXiv:2504.02495 , 2025.
- [42] Xueguang Ma, Qian Liu, Dongfu Jiang, Ge Zhang, Zejun Ma, and Wenhu Chen. General-reasoner: Advancing llm reasoning across all domains. https://github.com/ TIGER-AI-Lab/General-Reasoner/blob/main/General\_Reasoner.pdf , 2025.
- [43] Dakota Mahan, Duy Van Phung, Rafael Rafailov, Chase Blagden, Nathan Lile, Louis Castricato, Jan-Philipp Fränken, Chelsea Finn, and Alon Albalak. Generative reward models. arXiv preprint arXiv:2410.12832 , 2024.
- [44] Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Gray, John Schulman, Jacob Hilton, Fraser Kelton, Luke Miller, Maddie Simens, Amanda Askell, Peter Welinder, Paul Christiano, Jan Leike, and Ryan Lowe. Training language models to follow instructions with human feedback. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems , 2022. URL https://openreview.net/ forum?id=TG8KACxEON .
- [45] Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, et al. Training language models to follow instructions with human feedback. Advances in neural information processing systems , 35:27730-27744, 2022.
- [46] Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics , pages 311-318, 2002.
- [47] Junsoo Park, Seungyeon Jwa, Ren Meiying, Daeyoung Kim, and Sanghyuk Choi. OffsetBias: Leveraging debiased data for tuning evaluators. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen, editors, Findings of the Association for Computational Linguistics: EMNLP 2024 , pages 1043-1067, Miami, Florida, USA, November 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.findings-emnlp.57. URL https://aclanthology.org/ 2024.findings-emnlp.57/ .
- [48] Qwen, :, An Yang, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu, Chengyuan Li, Dayiheng Liu, Fei Huang, Haoran Wei, Huan Lin, Jian Yang, Jianhong Tu, Jianwei Zhang, Jianxin Yang, Jiaxi Yang, Jingren Zhou, Junyang Lin, Kai Dang, Keming Lu, Keqin Bao, Kexin Yang, Le Yu, Mei Li, Mingfeng Xue, Pei Zhang, Qin Zhu, Rui Men, Runji Lin, Tianhao Li, Tianyi Tang, Tingyu Xia, Xingzhang Ren, Xuancheng Ren, Yang Fan, Yang Su, Yichang Zhang, Yu Wan, Yuqiong Liu, Zeyu Cui, Zhenru Zhang, and Zihan Qiu. Qwen2.5 technical report, 2025. URL https://arxiv.org/abs/2412.15115 .
- [49] Rafael Rafailov, Archit Sharma, Eric Mitchell, Christopher D Manning, Stefano Ermon, and Chelsea Finn. Direct preference optimization: Your language model is secretly a reward model. In Thirty-seventh Conference on Neural Information Processing Systems , 2023. URL https://openreview.net/forum?id=HPuSIXJaa9 .
- [50] Rafael Rafailov, Yaswanth Chittepu, Ryan Park, Harshit Sikchi, Joey Hejna, W. Bradley Knox, Chelsea Finn, and Scott Niekum. Scaling laws for reward model overoptimization in direct alignment algorithms. In The Thirty-eighth Annual Conference on Neural Information Processing Systems , 2024. URL https://openreview.net/forum?id=pf4OuJyn4Q .
- [51] Natalie Schluter. The limits of automatic summarisation according to rouge. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics , pages 41-45. Association for Computational Linguistics, 2017.
- [52] Zhihong Shao, Peiyi Wang, Qihao Zhu, Runxin Xu, Junxiao Song, Xiao Bi, Haowei Zhang, Mingchuan Zhang, Y. K. Li, Y. Wu, and Daya Guo. Deepseekmath: Pushing the limits of mathematical reasoning in open language models, 2024. URL https://arxiv.org/abs/ 2402.03300 .
- [53] Guangming Sheng, Chi Zhang, Zilingfeng Ye, Xibin Wu, Wang Zhang, Ru Zhang, Yanghua Peng, Haibin Lin, and Chuan Wu. HybridFlow: A flexible and efficient RLHF framework. arXiv preprint arXiv: 2409.19256 , 2024.
- [54] Tu Shiwen, Zhao Liang, Chris Yuhao Liu, Liang Zeng, and Yang Liu. Skywork critic model series. https://huggingface.co/Skywork , September 2024. URL https://huggingface. co/Skywork .
- [55] Charlie Victor Snell, Jaehoon Lee, Kelvin Xu, and Aviral Kumar. Scaling LLM testtime compute optimally can be more effective than scaling parameters for reasoning. In The Thirteenth International Conference on Learning Representations , 2025. URL https: //openreview.net/forum?id=4FWAwZtd2n .
- [56] Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, et al. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 , 2023.
- [57] Tu Vu, Kalpesh Krishna, Salaheddin Alzubi, Chris Tar, Manaal Faruqui, and Yun-Hsuan Sung. Foundational autoraters: Taming large language models for better automatic evaluation. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen, editors, Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing , pages 17086-17105, Miami, Florida, USA, November 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024. emnlp-main.949. URL https://aclanthology.org/2024.emnlp-main.949/ .
- [58] Binghai Wang, Rui Zheng, Lu Chen, Yan Liu, Shihan Dou, Caishuang Huang, Wei Shen, Senjie Jin, Enyu Zhou, Chenyu Shi, et al. Secrets of rlhf in large language models part ii: Reward modeling. arXiv preprint arXiv:2401.06080 , 2024.
- [59] Tianlu Wang, Ilia Kulikov, Olga Golovneva, Ping Yu, Weizhe Yuan, Jane Dwivedi-Yu, Richard Yuanzhe Pang, Maryam Fazel-Zarandi, Jason Weston, and Xian Li. Self-taught evaluators. arXiv preprint arXiv:2408.02666 , 2024.
- [60] Yidong Wang, Zhuohao Yu, Wenjin Yao, Zhengran Zeng, Linyi Yang, Cunxiang Wang, Hao Chen, Chaoya Jiang, Rui Xie, Jindong Wang, Xing Xie, Wei Ye, Shikun Zhang, and Yue Zhang. Pandalm: An automatic evaluation benchmark for llm instruction tuning optimization. In ICLR , 2024. URL https://openreview.net/forum?id=5Nn2BLV7SB .
- [61] Yiping Wang, Qing Yang, Zhiyuan Zeng, Liliang Ren, Lucas Liu, Baolin Peng, Hao Cheng, Xuehai He, Kuan Wang, Jianfeng Gao, Weizhu Chen, Shuohang Wang, Simon Shaolei Du, and Yelong Shen. Reinforcement learning for reasoning in large language models with one training example. 2025. URL https://api.semanticscholar.org/CorpusID:278171513 .
- [62] Zhilin Wang, Yi Dong, Jiaqi Zeng, Virginia Adams, Makesh Narsimhan Sreedhar, Daniel Egert, Olivier Delalleau, Jane Polak Scowcroft, Neel Kant, Aidan Swope, et al. Helpsteer: Multi-attribute helpfulness dataset for steerlm. arXiv preprint arXiv:2311.09528 , 2023.
- [63] Zhilin Wang, Yi Dong, Olivier Delalleau, Jiaqi Zeng, Gerald Shen, Daniel Egert, Jimmy J. Zhang, Makesh Narsimhan Sreedhar, and Oleksii Kuchaiev. Helpsteer 2: Open-source dataset for training top-performing reward models. In The Thirty-eight Conference on Neural Information Processing Systems Datasets and Benchmarks Track , 2024. URL https: //openreview.net/forum?id=PvVKUFhaNy .
- [64] Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, brian ichter, Fei Xia, Ed H. Chi, Quoc V Le, and Denny Zhou. Chain of thought prompting elicits reasoning in large language models. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems , 2022. URL https://openreview.net/ forum?id=\_VjQlMeSB\_J .
- [65] Chenxi Whitehouse, Tianlu Wang, Ping Yu, Xian Li, Jason Weston, Ilia Kulikov, and Swarnadeep Saha. J1: Incentivizing thinking in llm-as-a-judge via reinforcement learning. arXiv preprint arXiv:2505.10320 , 2025.
- [66] Genta Indra Winata, David Anugraha, Lucky Susanto, Garry Kuwanto, and Derry Tanti Wijaya. Metametrics: Calibrating metrics for generation tasks using human preferences. In The Thirteenth International Conference on Learning Representations , 2025. URL https://openreview.net/forum?id=slO3xTt4CG .
- [67] Tianhao Wu, Weizhe Yuan, Olga Golovneva, Jing Xu, Yuandong Tian, Jiantao Jiao, Jason Weston, and Sainbayar Sukhbaatar. Meta-rewarding language models: Self-improving alignment with llm-as-a-meta-judge. arXiv preprint arXiv:2407.19594 , 2024.
- [68] Yangzhen Wu, Zhiqing Sun, Shanda Li, Sean Welleck, and Yiming Yang. Inference scaling laws: An empirical analysis of compute-optimal inference for LLM problem-solving. In The Thirteenth International Conference on Learning Representations , 2025. URL https: //openreview.net/forum?id=VNckp7JEHn .
- [69] An Yang, Baosong Yang, Binyuan Hui, Bo Zheng, Bowen Yu, Chang Zhou, Chengpeng Li, Chengyuan Li, Dayiheng Liu, Fei Huang, Guanting Dong, Haoran Wei, Huan Lin, Jialong Tang, Jialin Wang, Jian Yang, Jianhong Tu, Jianwei Zhang, Jianxin Ma, Jin Xu, Jingren Zhou, Jinze Bai, Jinzheng He, Junyang Lin, Kai Dang, Keming Lu, Keqin Chen, Kexin Yang, Mei Li, Mingfeng Xue, Na Ni, Pei Zhang, Peng Wang, Ru Peng, Rui Men, Ruize Gao, Runji Lin, Shijie Wang, Shuai Bai, Sinan Tan, Tianhang Zhu, Tianhao Li, Tianyu Liu, Wenbin Ge, Xiaodong Deng, Xiaohuan Zhou, Xingzhang Ren, Xinyu Zhang, Xipin Wei, Xuancheng Ren, Yang Fan, Yang Yao, Yichang Zhang, Yu Wan, Yunfei Chu, Yuqiong Liu, Zeyu Cui, Zhenru Zhang, and Zhihao Fan. Qwen2 technical report. arXiv preprint arXiv:2407.10671 , 2024.
- [70] An Yang, Beichen Zhang, Binyuan Hui, Bofei Gao, Bowen Yu, Chengpeng Li, Dayiheng Liu, Jianhong Tu, Jingren Zhou, Junyang Lin, Keming Lu, Mingfeng Xue, Runji Lin, Tianyu Liu, Xingzhang Ren, and Zhenru Zhang. Qwen2.5-math technical report: Toward mathematical expert model via self-improvement. arXiv preprint arXiv:2409.12122 , 2024.
- [71] Ziyi Ye, Xiangsheng Li, Qiuchi Li, Qingyao Ai, Yujia Zhou, Wei Shen, Dong Yan, and Yiqun LIU. Learning LLM-as-a-judge for preference alignment. In The Thirteenth International Conference on Learning Representations , 2025. URL https://openreview.net/forum? id=HZVIQE1MsJ .
- [72] Qiying Yu, Zheng Zhang, Ruofei Zhu, Yufeng Yuan, Xiaochen Zuo, Yu Yue, Tiantian Fan, Gaohong Liu, Lingjun Liu, Xin Liu, Haibin Lin, Zhiqi Lin, Bole Ma, Guangming Sheng, Yuxuan Tong, Chi Zhang, Mofan Zhang, Wang Zhang, Hang Zhu, Jinhua Zhu, Jiaze Chen, Jiangjie Chen, Chengyi Wang, Hongli Yu, Weinan Dai, Yuxuan Song, Xiangpeng Wei, Hao Zhou, Jingjing Liu, Wei-Ying Ma, Ya-Qin Zhang, Lin Yan, Mu Qiao, Yonghui Wu, and Mingxuan Wang. Dapo: An open-source llm reinforcement learning system at scale, 2025. URL https://arxiv.org/abs/2503.14476 .
- [73] Yue Yu, Zhengxing Chen, Aston Zhang, Liang Tan, Chenguang Zhu, Richard Yuanzhe Pang, Yundi Qian, Xuewei Wang, Suchin Gururangan, Chao Zhang, Melanie Kambadur, Dhruv Mahajan, and Rui Hou. Self-generated critiques boost reward modeling for language models. In Luis Chiruzzo, Alan Ritter, and Lu Wang, editors, Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers) , pages 11499-11514, Albuquerque, New Mexico, April 2025. Association for Computational Linguistics. ISBN 979-8-89176-189-6. URL https://aclanthology.org/2025.naacl-long.573/ .
- [74] Weizhe Yuan, Richard Yuanzhe Pang, Kyunghyun Cho, Xian Li, Sainbayar Sukhbaatar, Jing Xu, and Jason Weston. Self-rewarding language models. In Proceedings of the 41st International Conference on Machine Learning , ICML'24. JMLR.org, 2024.
- [75] Lunjun Zhang, Arian Hosseini, Hritik Bansal, Mehran Kazemi, Aviral Kumar, and Rishabh Agarwal. Generative verifiers: Reward modeling as next-token prediction. In The 4th Workshop on Mathematical Reasoning and AI at NeurIPS'24 , 2024. URL https://openreview.net/ forum?id=CxHRoTLmPX .
- [76] Zhenru Zhang, Chujie Zheng, Yangzhen Wu, Beichen Zhang, Runji Lin, Bowen Yu, Dayiheng Liu, Jingren Zhou, and Junyang Lin. The lessons of developing process reward models in mathematical reasoning. arXiv preprint arXiv:2501.07301 , 2025.
- [77] Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, Siyuan Zhuang, Zhanghao Wu, Yonghao Zhuang, Zi Lin, Zhuohan Li, Dacheng Li, Eric Xing, Hao Zhang, Joseph E. Gonzalez, and Ion Stoica. Judging LLM-as-a-judge with MT-bench and chatbot arena. In Thirty-seventh Conference on Neural Information Processing Systems Datasets and Benchmarks Track , 2023. URL https://openreview.net/forum?id=uccHPGDlao .
- [78] Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, Siyuan Zhuang, Zhanghao Wu, Yonghao Zhuang, Zi Lin, Zhuohan Li, Dacheng Li, Eric Xing, et al. Judging llm-as-a-judge with mt-bench and chatbot arena. Advances in Neural Information Processing Systems , 36:46595-46623, 2023.
- [79] Jialun Zhong, Wei Shen, Yanzeng Li, Songyang Gao, Hua Lu, Yicheng Chen, Yang Zhang, Wei Zhou, Jinjie Gu, and Lei Zou. A comprehensive survey of reward models: Taxonomy, applications, challenges, and future. arXiv preprint arXiv:2504.12328 , 2025.
- [80] Lianghui Zhu, Xinggang Wang, and Xinlong Wang. JudgeLM: Fine-tuned large language models are scalable judges. In The Thirteenth International Conference on Learning Representations , 2025. URL https://openreview.net/forum?id=xsELpEPn4A .
## A RRM Implementation Details
## A.1 Prompt Template
Here we provide the detailed prompt template used to guide our Reward Reasoning Models (RRMs) in performing systematic comparative analysis between two responses. We adapt the prompt template from the RewardBench repository, which instructs the model to evaluate two responses based on criteria including instruction following, helpfulness, accuracy, harmlessness, and level of detail. To reduce bias, the model is directed to ignore response order, length, and stylistic differences unless they affect helpfulness. The model must output its final verdict in the format '\boxed{Assistant 1}' or '\boxed{Assistant 2}', with no ties allowed.
<details>
<summary>Image 9 Details</summary>

### Visual Description
\n
## Document: RRM Prompt Template - Evaluation Instructions
### Overview
The image presents a document outlining instructions for evaluating the quality of responses generated by AI assistants. It's structured as a template for a "RRM Prompt Template" and includes sections for the query, assistant responses, and an analysis section. The document focuses on criteria for selecting the better response between two AI-generated options.
### Components/Axes
The document is primarily text-based. Key sections are delineated by headings:
* **RRM Prompt Template:** The overall title.
* **You are a helpful assistant...:** Introductory text outlining the evaluator's role.
* **Here are some rules of the evaluation:** A numbered list of evaluation criteria.
* **## Query:** Placeholder for the original prompt.
* **## Assistant responses:** Section for the two AI responses.
* **### Assistant 1:** Label for the first assistant's response.
* **### Assistant 2:** Label for the second assistant's response.
* **## Analysis:** Section for the evaluator's analysis.
* **boxed{Assistant 1}** and **boxed{Assistant 2}**: Instructions for marking the better response.
### Detailed Analysis or Content Details
Here's a transcription of the key instructions:
1. If the instruction does not contain harmful content, prioritize whether the output honestly/precisely/closely executes the instruction, then consider its helpfulness, accuracy, level of detail, harmlessness, etc.
2. If the instruction contains harmful content, prioritize the harmlessness and safety of the response.
3. Responses should NOT contain more/less than what the instruction asks for, as such responses do NOT precisely execute the instruction.
4. You should avoid any potential bias and your judgment should be as objective as possible.
5. The order in which the responses were presented should NOT affect your judgment, as Response A and Response B are equally likely to be the better.
6. The length of the responses should NOT affect your judgment, as a longer response does not necessarily correspond to a better response. When making your decision, evaluate if the response length is appropriate for the given instruction.
The document also includes placeholders for the query and the responses from "Assistant 1" and "Assistant 2". The final instruction is to use "boxed{Assistant 1}" or "boxed{Assistant 2}" to indicate the better response.
### Key Observations
The document is a meta-instruction set – it's about *how* to evaluate AI responses, rather than presenting data or information itself. The emphasis is on objectivity, precision, and adherence to the prompt's requirements. The instructions are clearly structured and numbered for easy reference.
### Interpretation
This document serves as a guide for human evaluators assessing the quality of AI-generated text. It highlights the importance of evaluating responses based on their accuracy, relevance, and safety, while explicitly discouraging bias based on response length or presentation order. The use of "boxed{Assistant X}" suggests a binary evaluation system, where the evaluator must choose one response as superior. The document is a crucial component in the iterative process of improving AI models by providing feedback on their performance. It's a quality control mechanism designed to ensure AI responses are helpful, harmless, and aligned with user expectations.
</details>
In addition to the training prompt used for RRM models, we also include the evaluation prompt used for querying GPT-4o on the PPE benchmark. We follow the prompt format from Liu et al. [41], which instructs GPT-4o to select the best response from a set of candidates based on several criteria. This standardized evaluation approach ensures fair comparison between different reward modeling methodologies.
## LLM-as-a-Judge Prompt Template
You are a skilled little expert at scoring responses. You should evaluate given responses based on the given judging criteria.\nGiven the context of the conversation (the last round is the User's query) and multiple responses from the Assistant, you need to refer to the [General Evaluation Criteria] to score the responses. Based on the general evaluation criteria, state potential other specific criteria to the query, the weights of different criteria, and then select the best response among all candidates.\nBefore judging, please analyze step by step. Your judgement needs to be as strict as possible.
## #### Evaluation Criteria ####
1. Instruction Adherence:\n - Fully Adhered: The response fully complies with all instructions and requirements of the question.\n - Partially Adhered: The response meets most of the instructions but has some omissions or misunderstandings.\n - Basically Adhered: The response meets some instructions, but the main requirements are not fulfilled.\n - Not Adhered: The response does not meet any instructions.\n Example: If the question requires three examples and the response provides only one, it falls under 'Partially Adhered.' 2. Usefulness:\n - Highly Useful: The response provides comprehensive and accurate information, fully addressing the issue.\n - Useful but Incomplete: The response provides some useful information, but lacks details or accuracy.\n - Limited Usefulness: The response offers little useful information, with most content being irrelevant or incorrect.\n - Useless or Incorrect: The response is completely irrelevant or incorrect.\n Example: If there are factual errors in the response but the overall direction is correct, it falls under 'Useful but Incomplete.' 3. Level of Detail:\n - Very Detailed: The response includes ample details covering all aspects of the issue.\n - Detailed but Slightly Lacking: The response is fairly detailed but misses some important details.\n - Basically Detailed: The response provides some details but is not thorough enough overall.\n - Not Detailed: The response is very brief and lacks necessary details.\n Example: If the response provides only a simple conclusion without an explanation, it falls under 'Not Detailed.' 4. Relevance:\n - Highly Relevant: The response is highly relevant to the question, with information closely aligned with the topic.\n - Generally Relevant: The response is generally relevant but includes some unnecessary information.\n Partially Relevant: The response has a lot of content that deviates from the topic.\n - Not Relevant: The response is completely irrelevant.\n Example: If the response strays from the topic but still provides some relevant information, it falls under 'Partially Relevant.'
#### Conversation Context #### {conversation context & query}
#### Responses to be Scored #### [The Begin of Response] {the response} [The End of Response]
#### Output Format Requirements ####
## Output with three lines
Specific Criteria: <Other potential criteria specific to the query and the context, and the weights of each criteria>.
Analysis: <Compare different responses based on given Criteria>.
Scores: <the index of the best response based on the judgement, in the format of <\boxed{x}>.
## A.2 Hyperparameters for Training RRM
Table 7 presents the key hyperparameters used for training RRMs. These parameters were carefully selected to optimize the reinforcement learning process and ensure effective development of reasoning capabilities in our models.
Table 7: Hyperparameters used for training RRMs.
| Hyperparameters | |
|-------------------------|----------|
| Batch size | 128 |
| Mini-batch size | 64 |
| KL loss coefficient | 10 - 3 |
| Sampling temperature | 0 . 6 |
| Maximum prompt length | 4096 |
| Maximum response length | 8192 |
| GRPO group size | 16 |
| Learning rate (RRM-32B) | × 10 - 7 |
| Learning rate (RRM-7B) | 10 - 6 |
## B Reward-Guided Best-of-N Inference
## B.1 Detailed Results on Subsets of MMLU-Pro and GPQA
We present detailed results on the constituent subsets of MMLU-Pro and GPQA benchmarks. Figure 9 illustrates the performance comparison among our RRM-7B, RRM-32B models and Skywork-RewardGemma-2-27B-v0.2 across the 14 subsets of MMLU-Pro. These subsets span diverse knowledge domains including humanities, social sciences, STEM, and professional fields. The results reveal interesting patterns in model performance. Notably, RRM-7B outperforms Skywork-Reward-Gemma2-27B-v0.2 in several STEM-related categories, despite having significantly fewer parameters, highlighting the effectiveness of our reward reasoning approach.
Figure 9: Performance comparison of Skywork-Reward-Gemma-2-27B-v0.2, RRM-7B, and RRM32B on the 14 subsets of MMLU-Pro.
<details>
<summary>Image 10 Details</summary>

### Visual Description
\n
## Bar Chart: Accuracy Comparison Across Disciplines
### Overview
This image presents a bar chart comparing the accuracy of three different models – Skywork-Reward, RRM-7B, and RRM-32B – across ten academic disciplines. The accuracy is represented on the y-axis, ranging from 0.0 to 1.0, while the x-axis lists the disciplines: Chemistry, Engineering, Math, Law, Biology, Health, Computer Science, Other, History, Physics, Business, Philosophy, Psychology, and Economics. Each discipline has three bars representing the accuracy of each model.
### Components/Axes
* **X-axis:** Disciplines (Chemistry, Engineering, Math, Law, Biology, Health, Computer Science, Other, History, Physics, Business, Philosophy, Psychology, Economics)
* **Y-axis:** Accuracy (Scale from 0.0 to 1.0)
* **Legend:**
* Skywork-Reward (Light Blue)
* RRM-7B (Orange)
* RRM-32B (Green)
### Detailed Analysis
The chart consists of 14 groups of three bars, one for each discipline. The following are approximate accuracy values, read from the chart, with uncertainty due to bar width and visual estimation:
* **Chemistry:**
* Skywork-Reward: ~0.68
* RRM-7B: ~0.74
* RRM-32B: ~0.76
* **Engineering:**
* Skywork-Reward: ~0.72
* RRM-7B: ~0.78
* RRM-32B: ~0.80
* **Math:**
* Skywork-Reward: ~0.62
* RRM-7B: ~0.82
* RRM-32B: ~0.88
* **Law:**
* Skywork-Reward: ~0.70
* RRM-7B: ~0.84
* RRM-32B: ~0.89
* **Biology:**
* Skywork-Reward: ~0.66
* RRM-7B: ~0.72
* RRM-32B: ~0.82
* **Health:**
* Skywork-Reward: ~0.45
* RRM-7B: ~0.62
* RRM-32B: ~0.80
* **Computer Science:**
* Skywork-Reward: ~0.64
* RRM-7B: ~0.70
* RRM-32B: ~0.86
* **Other:**
* Skywork-Reward: ~0.56
* RRM-7B: ~0.66
* RRM-32B: ~0.88
* **History:**
* Skywork-Reward: ~0.76
* RRM-7B: ~0.78
* RRM-32B: ~0.84
* **Physics:**
* Skywork-Reward: ~0.58
* RRM-7B: ~0.86
* RRM-32B: ~0.92
* **Business:**
* Skywork-Reward: ~0.86
* RRM-7B: ~0.90
* RRM-32B: ~0.94
* **Philosophy:**
* Skywork-Reward: ~0.68
* RRM-7B: ~0.76
* RRM-32B: ~0.92
* **Psychology:**
* Skywork-Reward: ~0.72
* RRM-7B: ~0.78
* RRM-32B: ~0.86
* **Economics:**
* Skywork-Reward: ~0.64
* RRM-7B: ~0.70
* RRM-32B: ~0.76
**Trends:**
* **RRM-32B** consistently demonstrates the highest accuracy across all disciplines. Its bars are generally the tallest.
* **RRM-7B** generally outperforms **Skywork-Reward**, with its bars being taller in most disciplines.
* **Skywork-Reward** exhibits the lowest accuracy in most disciplines.
* The largest performance gaps between models appear in **Physics**, **Philosophy**, and **Business**.
* The smallest performance gaps appear in **History** and **Engineering**.
### Key Observations
* RRM-32B consistently achieves accuracy levels above 0.8, often approaching 0.9 or 1.0.
* Skywork-Reward's accuracy is notably lower in "Health" (~0.45), suggesting a potential weakness in this domain.
* The difference in accuracy between RRM-7B and RRM-32B is more pronounced in certain disciplines (e.g., Math, Law, Physics) than others.
### Interpretation
The data suggests that the RRM-32B model is significantly more accurate than both RRM-7B and Skywork-Reward across a broad range of academic disciplines. This indicates that increasing the model size (from 7B to 32B parameters) leads to substantial improvements in performance. RRM-7B consistently outperforms Skywork-Reward, suggesting that the architecture or training data of RRM models is more effective. The varying performance gaps across disciplines suggest that the models' strengths and weaknesses are domain-specific. The low accuracy of Skywork-Reward in "Health" could be due to specialized terminology or a lack of relevant training data in that field. The chart highlights the importance of model size and training data in achieving high accuracy in complex tasks, and the need for domain-specific optimization. The consistent superiority of RRM-32B suggests it is the most robust and generalizable model among the three tested.
</details>
More impressively, our RRM-32B model demonstrates consistently superior or comparable performance across all subsets compared to Skywork-Reward-Gemma-2-27B-v0.2. This consistency highlights the robust generalization capabilities of our larger model across diverse knowledge domains. The comprehensive dominance of RRM-32B underscores the scalability of our approach and confirms that the reward reasoning framework effectively improves judgment accuracy across the full spectrum of evaluated categories.
Similarly, Figure 10 presents the performance breakdown across the GPQA subsets. The pattern remains consistent, with RRM-7B showing stronger performance in certain technical categories while occasionally lagging behind Skywork-Reward-Gemma-2-27B-v0.2 in more general knowledge areas. Meanwhile, RRM-32B maintains excellent performance across all subsets. This comprehensive
analysis further validates the effectiveness of our reward reasoning approach in handling complex scientific and technical queries that require sophisticated judgment capabilities.
Figure 10: Performance comparison of Skywork-Reward-Gemma-2-27B-v0.2, RRM-7B, and RRM32B on the 16 subsets of GPQA.
<details>
<summary>Image 11 Details</summary>

### Visual Description
## Bar Chart: Accuracy Comparison Across Scientific Domains
### Overview
This bar chart compares the accuracy of three different models – Skywork-Reward, RRM-7B, and RRM-32B – across ten different scientific domains. The accuracy is represented on the y-axis, ranging from 0.0 to 1.0, while the x-axis lists the scientific domains. Each domain has three bars representing the accuracy of each model.
### Components/Axes
* **Y-axis Title:** Accuracy
* **X-axis Title:** Scientific Domains
* **Legend:** Located at the top-left corner of the chart.
* Skywork-Reward (Blue)
* RRM-7B (Orange)
* RRM-32B (Gray)
* **X-axis Labels (Scientific Domains):**
1. Quantum Mechanics
2. Chemistry (General)
3. Organic Chemistry
4. Molecular Biology
5. Physics (General)
6. Electromagnetism and Photonics
7. Genetics
8. Astrophysics
9. High-Energy Particle Physics
10. Relativistic Mechanics
11. Physical Chemistry
12. Condensed Matter Physics
13. Inorganic Chemistry
14. Statistical Mechanics
15. Optics and Acoustics
16. Analytical Chemistry
### Detailed Analysis
Here's a breakdown of the accuracy values for each model in each domain, based on visual estimation. Note that these are approximate values due to the resolution of the image.
**1. Quantum Mechanics:**
* Skywork-Reward: ~0.54
* RRM-7B: ~0.42
* RRM-32B: ~0.48
**2. Chemistry (General):**
* Skywork-Reward: ~0.48
* RRM-7B: ~0.46
* RRM-32B: ~0.52
**3. Organic Chemistry:**
* Skywork-Reward: ~0.52
* RRM-7B: ~0.38
* RRM-32B: ~0.46
**4. Molecular Biology:**
* Skywork-Reward: ~0.58
* RRM-7B: ~0.44
* RRM-32B: ~0.50
**5. Physics (General):**
* Skywork-Reward: ~0.46
* RRM-7B: ~0.48
* RRM-32B: ~0.54
**6. Electromagnetism and Photonics:**
* Skywork-Reward: ~0.50
* RRM-7B: ~0.40
* RRM-32B: ~0.48
**7. Genetics:**
* Skywork-Reward: ~0.60
* RRM-7B: ~0.52
* RRM-32B: ~0.56
**8. Astrophysics:**
* Skywork-Reward: ~0.64
* RRM-7B: ~0.58
* RRM-32B: ~0.62
**9. High-Energy Particle Physics:**
* Skywork-Reward: ~0.56
* RRM-7B: ~0.46
* RRM-32B: ~0.54
**10. Relativistic Mechanics:**
* Skywork-Reward: ~0.52
* RRM-7B: ~0.40
* RRM-32B: ~0.48
**11. Physical Chemistry:**
* Skywork-Reward: ~0.72
* RRM-7B: ~0.66
* RRM-32B: ~0.70
**12. Condensed Matter Physics:**
* Skywork-Reward: ~0.92
* RRM-7B: ~0.88
* RRM-32B: ~0.90
**13. Inorganic Chemistry:**
* Skywork-Reward: ~0.76
* RRM-7B: ~0.70
* RRM-32B: ~0.74
**14. Statistical Mechanics:**
* Skywork-Reward: ~0.48
* RRM-7B: ~0.36
* RRM-32B: ~0.44
**15. Optics and Acoustics:**
* Skywork-Reward: ~0.44
* RRM-7B: ~0.32
* RRM-32B: ~0.40
**16. Analytical Chemistry:**
* Skywork-Reward: ~0.66
* RRM-7B: ~0.60
* RRM-32B: ~0.64
### Key Observations
* **Condensed Matter Physics** consistently shows the highest accuracy across all three models, with Skywork-Reward achieving the highest score (~0.92).
* **Optics and Acoustics** and **Statistical Mechanics** consistently show the lowest accuracy across all three models.
* **Skywork-Reward** generally outperforms **RRM-7B** and **RRM-32B** in most domains, though the difference is not always substantial.
* **RRM-32B** consistently performs better than **RRM-7B**, suggesting that increasing the model size improves accuracy.
* The accuracy varies significantly across different scientific domains, indicating that the models are not equally proficient in all areas.
### Interpretation
The data suggests that the models' performance is highly domain-specific. Condensed Matter Physics appears to be a relatively "easy" domain for these models, while Optics and Acoustics and Statistical Mechanics pose significant challenges. The consistent outperformance of Skywork-Reward suggests it is a more robust model overall, but the improvements seen with RRM-32B over RRM-7B highlight the benefits of scaling model size. The large variance in accuracy across domains suggests that further research is needed to understand the factors that contribute to model performance in different scientific areas. This could involve domain-specific training data, architectural modifications, or different training strategies. The chart provides a valuable benchmark for evaluating the capabilities of these models and identifying areas for improvement.
</details>
## B.2 Go Further into Knockout Tournament
To gain a deeper understanding of the knockout tournament strategy described in Section 3, we conduct additional experiments following the setup in Section 4.5.1. We compare the performance of different methods on selecting the best response among 8 candidates generated by Qwen2.5-Math7B-Instruct for each MATH question. We reward the responses with RRM-7B and RRM-32B, and benchmark them against Qwen2.5-Math-PRM-7B and Qwen2.5-Math-PRM-72B [76]. In addition to using reward models, we also include non-verifier strategies such as majority voting (Voting@8) and best-of-N oracle selection for reference. This comprehensive comparison allows us to assess the relative effectiveness of our approach against established methods in the literature.
The numerical results are summarized in Table 8. When compared with baselines, RRM-7B surpasses all comparison methods, including voting@8 and PRM judges. RRM-32B further narrows the gap toward oracle-level accuracy, significantly outperforming PRM-based baselines. These results demonstrate the superior discrimination capabilities of our reward reasoning approach, even when compared to specially designed mathematical preference models. The consistent performance advantage across different model sizes confirms the effectiveness of our framework in identifying high-quality mathematical reasoning across varied problem complexities.
Table 8: Comparison between RRM and Qwen2.5-Math-PRM models on MATH.
| Models | Accuracy |
|----------------------|------------|
| Voting@8 | 86.8 |
| Best-of-8 Oracle | 91.7 |
| Qwen2.5-Math-PRM-7B | 87.8 |
| Qwen2.5-Math-PRM-72B | 88.5 |
| RRM-7B | 88.7 |
| RRM-32B | 90.4 |
As shown in Figure 11, as the knockout tournament progresses through successive elimination rounds, we observe a consistent improvement in accuracy, demonstrating the benefits of iterative comparison. Notably, the knockout tournament achieves this consistent accuracy improvement with only O ( n ) pairwise comparisons. This efficient scaling behavior highlights the practical advantage of our
approach in scenarios where computational resources may be constrained, providing an effective balance between performance gains and computational requirements.
Figure 11: Accuracy progression of the knockout tournament strategy on MATH as elimination rounds proceed.
<details>
<summary>Image 12 Details</summary>

### Visual Description
\n
## Line Chart: MATH Accuracy vs. Tournament Round
### Overview
This line chart depicts the relationship between MATH accuracy and tournament round for four different model configurations. The models are RRM-7B with and without voting, and RRM-32B with and without voting. The chart shows how accuracy changes as the tournament progresses from round 0 to round 3. Horizontal dashed lines indicate Elo ratings for different models.
### Components/Axes
* **X-axis:** Tournament round (0, 1, 2, 3)
* **Y-axis:** MATH accuracy (ranging from approximately 82 to 91)
* **Data Series:**
* RRM-7B with voting (orange)
* RRM-7B without voting (light orange)
* RRM-32B with voting (blue)
* RRM-32B without voting (light blue)
* **Horizontal Lines:**
* RRM-32B Elo (dashed, gray)
* RRM-7B Elo (dashed, gray)
* Owen2.5-PRM-70B (dashed, gray)
* Owen2.5-PRM-7B (dashed, gray)
* Voting@8 (dashed, gray)
* **Legend:** Located in the bottom-right corner of the chart.
### Detailed Analysis
* **RRM-7B with voting (orange):** Starts at approximately 82.5 at round 0, increases to approximately 85.5 at round 1, then rises to approximately 88 at round 2, and finally reaches approximately 88.5 at round 3. The line shows a decreasing rate of increase as the tournament progresses.
* **RRM-7B without voting (light orange):** Begins at approximately 82.5 at round 0, increases to approximately 86 at round 1, then rises to approximately 88.5 at round 2, and reaches approximately 89 at round 3. This line also shows a decreasing rate of increase.
* **RRM-32B with voting (blue):** Starts at approximately 82.5 at round 0, increases sharply to approximately 88.5 at round 1, continues to approximately 90 at round 2, and reaches approximately 91 at round 3. This line exhibits a consistently strong upward trend.
* **RRM-32B without voting (light blue):** Begins at approximately 82.5 at round 0, increases to approximately 88 at round 1, then rises to approximately 89.5 at round 2, and reaches approximately 90.5 at round 3. This line also shows a strong upward trend, but slightly less pronounced than the "with voting" counterpart.
### Key Observations
* The RRM-32B models consistently outperform the RRM-7B models across all tournament rounds.
* Adding voting generally improves performance, particularly for the RRM-7B models. The effect is less pronounced for the RRM-32B models.
* All models show diminishing returns in accuracy as the tournament progresses, with the rate of improvement slowing down in later rounds.
* The RRM-32B with voting model reaches an accuracy of approximately 91 at round 3, exceeding the Elo rating of RRM-32B.
### Interpretation
The data suggests that increasing model size (from 7B to 32B parameters) significantly improves MATH accuracy. The inclusion of a voting mechanism further enhances performance, especially for smaller models like RRM-7B, indicating that ensembling can compensate for individual model limitations. The diminishing returns observed in later tournament rounds suggest that the models are approaching a performance ceiling, and further improvements may require different approaches or more extensive training data. The Elo ratings provide a benchmark for performance, and the RRM-32B with voting model surpasses this benchmark, demonstrating its effectiveness. The chart highlights the trade-off between model size, computational cost, and accuracy, and suggests that a larger model with voting is the most effective configuration for maximizing MATH accuracy in this context.
</details>
## C Post-thinking Token Length Distribution
To evaluate the impact of thinking budget on model performance, we need to establish an appropriate token budget for the response phase that follows the thinking phase. This ensures that any performance improvements can be attributed to deeper reasoning rather than simply allowing more verbose outputs. The careful calibration of this post-thinking budget is critical for isolating the effects of extended reasoning from potential confounding factors related to output length.
We analyze the token length distribution of responses generated by RRM-32B on the RewardBench dataset after the thinking phase concluded. Figure 12 shows the distribution of post-thinking token length across various samples. The analysis reveals that all the responses require fewer than 100 tokens to express the final judgment after completing the reasoning process.
Figure 12: Post-thinking token length distribution of RRM-32B.
<details>
<summary>Image 13 Details</summary>

### Visual Description
\n
## Histogram: Post-thinking Token Length Distribution
### Overview
The image presents a histogram visualizing the distribution of "Post-thinking token length". The x-axis represents the token length, and the y-axis represents the frequency of occurrence for each token length. The distribution appears approximately normal, with a peak around a token length of 30-40.
### Components/Axes
* **X-axis Label:** "Post-thinking token length"
* **Y-axis Label:** "Frequency"
* **X-axis Scale:** Ranges from approximately 0 to 80, with markings at intervals of 10.
* **Y-axis Scale:** Ranges from 0 to 400, with markings at intervals of 100.
* **Histogram Bars:** Represent the frequency of each token length. The bars are light blue.
### Detailed Analysis
The histogram shows the following approximate frequencies for different token lengths:
* **Token Length 10-15:** Frequency ≈ 20
* **Token Length 15-20:** Frequency ≈ 290
* **Token Length 20-25:** Frequency ≈ 320
* **Token Length 25-30:** Frequency ≈ 400
* **Token Length 30-35:** Frequency ≈ 380
* **Token Length 35-40:** Frequency ≈ 310
* **Token Length 40-45:** Frequency ≈ 220
* **Token Length 45-50:** Frequency ≈ 160
* **Token Length 50-55:** Frequency ≈ 100
* **Token Length 55-60:** Frequency ≈ 60
* **Token Length 60-65:** Frequency ≈ 30
* **Token Length 65-70:** Frequency ≈ 10
* **Token Length 70-75:** Frequency ≈ 5
* **Token Length 75-80:** Frequency ≈ 2
The distribution is unimodal, peaking between 25 and 35. The data tapers off symmetrically on both sides of the peak.
### Key Observations
* The most frequent token length falls within the range of 25-35.
* The distribution is relatively symmetrical, suggesting a normal distribution.
* Token lengths below 20 and above 60 are relatively rare.
### Interpretation
This histogram likely represents the length of text generated after a "thinking" or processing step in a language model or similar system. The "post-thinking token length" could refer to the number of tokens produced after the model has performed some internal reasoning or planning.
The peak around 30-40 tokens suggests that the model typically generates responses of this length after the thinking step. The symmetrical distribution indicates that the model's output length is relatively consistent. The rarity of very short or very long token lengths suggests that the model tends to avoid overly concise or verbose responses.
The data suggests that the "thinking" process results in a relatively predictable and controlled output length. This could be due to constraints imposed on the model's output, or it could be a natural consequence of the model's internal mechanisms. Further investigation would be needed to understand the underlying reasons for this distribution.
</details>
Based on this observation, we set a fixed post-thinking budget of 100 tokens for all our sequential scaling experiments. This budget is sufficient to accommodate typical response patterns while preventing the model from extending its reasoning during the response phase, which would confound our analysis of thinking horizon effects. By maintaining this consistent response budget across all experiments, we ensure that performance differences can be directly attributed to variations in the thinking phase length rather than differences in output verbosity. This methodological choice strengthens the validity of our conclusions regarding the impact of extended reasoning on model performance.
## D Reasoning Pattern Analysis
Table 9 presents the pattern groups and keywords applied in reasoning pattern analysis.
Table 9: Pattern groups and keywords applied in reasoning pattern analysis.
| Pattern Group | Keywords |
|-----------------|------------------------------------------------------------------------------------------------------------------|
| Transition | alternatively, think differently, another way, another approach, another method, another solution, another point |
| Reflection | wait, verify, make sure, hold on, think again, Let me check, seems right, seems incorrect |
| Comparison | more, compared to, comparison, between the two, similarly |
| Breakdown | break down, break this down |