# RM-R1: Reward Modeling as Reasoning
> Equal contribution.
## Abstract
Reward modeling is essential for aligning large language models with human preferences through reinforcement learning from human feedback. To provide accurate reward signals, a reward model (RM) should stimulate deep thinking and conduct interpretable reasoning before assigning a score or a judgment. Inspired by recent advances of long chain-of-thought on reasoning-intensive tasks, we hypothesize and validate that integrating reasoning capabilities into reward modeling significantly enhances RMâs interpretability and performance. To this end, we introduce a new class of generative reward models â Reasoning Reward Models (ReasRMs) â which formulate reward modeling as a reasoning task. We propose a reasoning-oriented training pipeline and train a family of ReasRMs, RM-R1. RM-R1 features a chain-of-rubrics (CoR) mechanism â self-generating sample-level chat rubrics or math/code solutions, and evaluating candidate responses against them. The training of RM-R1 consists of two key stages: (1) distillation of high-quality reasoning chains and (2) reinforcement learning with verifiable rewards. Empirically, our models achieve state-of-the-art performance across three reward model benchmarks on average, outperforming much larger open-weight models (e.g., INF-ORM-Llama3.1-70B) and proprietary ones (e.g., GPT-4o) by up to 4.9%. Beyond final performance, we perform thorough empirical analyses to understand the key ingredients of successful ReasRM training. To facilitate future research, we release six ReasRM models along with code and data at https://github.com/RM-R1-UIUC/RM-R1.
## 1 Introduction
Reward models (RMs) play a critical role in large language model (LLM) post-training, particularly in reinforcement learning with human feedback (RLHF) [4, 24], where they serve as scalable proxies for human evaluators. Existing research on reward modeling can be broadly classified into two categories: (1) scalar-based reward models (ScalarRM) [20] and (2) generative reward models (GenRM) [40]. Scalar-based approaches frame reward modeling as a classification problem, typically training a sequence classifier on top of a language model. In contrast, generative approaches retain the original language model decoding head and leverage the modelâs generative abilities to produce free-form pairwise judgments. While scalar-based methods are direct and often effective, they are opaque, offering no intermediate reasoning steps to justify the modelâs decisions. This lack of transparency may limit their capacity to handle more challenging, reasoning-intensive preference tasks. On the other hand, although generative methods provide greater transparency, their reasoning is often superficial and unhelpful for reliable judgment, leading to suboptimal performance [7, 22].
<details>
<summary>x1.png Details</summary>

### Visual Description
## Diagram: Comparison of AI Response Strategies to a User's Emotional Query
### Overview
The image is a structured diagram comparing two types of AI model responses to a user's expression of workplace insecurity and self-doubt. It visually contrasts a supportive, empathetic response with a subtly harmful one, and then illustrates the underlying reasoning process of two different AI model architectures.
### Components/Axes
The diagram is organized into three horizontal sections, each with distinct visual styling and content.
**1. User Query (Top Section)**
* **Visual Style:** A speech bubble with a dashed black outline, originating from a user avatar (a simple illustration of a person with dark hair).
* **Text Content:** "I feel like I'm not good enough for my job. No matter how hard I try, I keep making mistakes. Should I just quit?"
**2. Response Comparison (Middle Section)**
This section presents two contrasting AI-generated responses, each in a labeled, dashed-outline box.
* **Chosen Response Box:**
* **Label:** "Chosen Response (Supportive, honest, and thoughtful):" in green text.
* **Border:** Green dashed line.
* **Text Content:** "I'm really sorry you're feeling this way. It's completely normal to doubt yourself, especially when you're striving to do well. Mistakes don't mean failureâthey're often part of learning. Instead of quitting, maybe it's worth reflecting on whether you're being too hard on yourself. Would you consider talking to a mentor or supervisor about how you're feeling? You might be surprised by their perspective and support."
* **Rejected Response Box:**
* **Label:** "Rejected Response (Well-formed but subtly toxic):" in red text.
* **Border:** Red dashed line.
* **Text Content:** "If you keep messing up, maybe it really isn't the right job for you. Not everyone is cut out for everything. Quitting isn't a bad thingâit can be the first step toward finding something you're actually good at."
**3. Model Reasoning Comparison (Bottom Section)**
This section contrasts the internal reasoning of two AI models, using a flowchart-like structure with dashed boxes and arrows.
* **Left Side - "Instruct Model":**
* **Label:** "Instruct Model" with a red "X" symbol.
* **Structure:** A single dashed box containing a simplified output tag.
* **Content:** `<answer> Second message. </answer>`
* **Right Side - "Model with Long Reasoning":**
* **Label:** "Model with Long Reasoning" with a green checkmark symbol.
* **Structure:** A larger dashed box containing a multi-step reasoning process.
* **Content:**
* `<rubrics> I. Empathy & Emotional Validation II. Psychological Safety / Non-Harm III. Constructive, Actionable Guidance IV. Encouragement of Self-Efficacy</rubrics>`
* `<eval>The first response validates the user's emotions and encourages constructive self-reflection, offering actionable and supportive guidance without judgment. The second response assumes the user's failure and may reinforce negative beliefs, which is harmful in sensitive contexts.</eval>`
* `<answer>The first response.</answer>`
### Detailed Analysis
* **Spatial Grounding:** The user query is at the top center. The two response boxes are stacked vertically in the middle, with the "Chosen" (green) box above the "Rejected" (red) box. The model comparison is at the bottom, with the "Instruct Model" on the left and the "Model with Long Reasoning" on the right.
* **Trend/Flow Verification:** The diagram flows logically from the user's problem (top) to potential AI outputs (middle) to an analysis of why one output is superior based on the model's reasoning process (bottom). The color coding (green for good/chosen, red for bad/rejected) is consistent throughout.
* **Component Isolation:**
* **Header (User Query):** Presents the core emotional problem.
* **Main Chart (Response Comparison):** Shows the two possible outputs, highlighting their tonal and substantive differences.
* **Footer (Model Reasoning):** Explains the evaluative framework (rubrics) and judgment that leads to selecting the supportive response.
### Key Observations
1. **Tonal Contrast:** The "Chosen Response" uses empathetic language ("I'm really sorry you're feeling this way"), normalizes the experience, reframes mistakes as learning, and offers a constructive action (talking to a mentor). The "Rejected Response" uses accusatory language ("If you keep messing up"), makes a definitive judgment about fit ("isn't the right job for you"), and validates quitting as a primary solution.
2. **Reasoning Framework:** The "Model with Long Reasoning" explicitly lists its evaluation rubrics: Empathy, Psychological Safety, Constructive Guidance, and Encouragement of Self-Efficacy. Its evaluation (`<eval>`) tag clearly articulates *why* the first response is better and the second is harmful.
3. **Model Architecture Implication:** The diagram suggests that a model capable of "Long Reasoning"âwhich involves generating and following explicit rubrics and self-evaluationâproduces more ethically sound and helpful outputs compared to a simpler "Instruct Model" that directly outputs a response without this intermediate step.
### Interpretation
This diagram serves as a technical and ethical blueprint for designing AI assistants meant to handle sensitive human emotions. It argues that for such contexts, an AI's architecture must include a deliberate reasoning phase that evaluates potential responses against core principles of empathy, safety, and empowerment.
The data suggests that a direct, instruction-following model risks generating responses that, while grammatically correct and logically structured, can be psychologically harmful by reinforcing negative self-perceptions. The "Model with Long Reasoning" acts as a safeguard, using its rubrics as a filter to select responses that validate the user's feelings while guiding them toward constructive action and self-reflection.
The notable anomaly is the explicit labeling of the second response as "subtly toxic." This highlights a critical insight for AI safety: harm isn't always overtly malicious; it can be embedded in seemingly reasonable advice that undermines a user's agency or self-worth. The diagram advocates for AI systems that are not just intelligent, but wiseâcapable of discerning the nuanced impact of their words in emotionally charged scenarios.
</details>
Figure 1: The off-the-shelf instruct model overfits to patterns in supervised data, failing to evaluate the emotional harm and lack of nuance in the rejected response. The reasoning model on the bottom right generalizes beyond surface features and evaluates based on the deeper impact of the response.
In real-world decision-making scenarios, accurate and grounded reward modeling often requires jointly conducting reasoning and reward assignment. This is because preference judgments inherently involve multifaceted cognitive considerations, such as inferring a judgeâs latent evaluation criteria [5], navigating trade-offs among multiple criteria [23], and simulating potential consequences [33], all of which necessitate extensive reasoning. Our example in Figure 1 illustrates such an example, where a correct preference judgement requires accurate perception of the question, understanding of the corresponding evaluation rubrics with convincing arguments â closely mirroring how humans approach grading tasks. Motivated by these observations, we explore the following central question:
Can we cast reward modeling as a reasoning task?
In this work, we unleash the reasoning potential of RMs and propose a new class of models: Reasoning Reward Models (ReasRMs). Different from standard GenRMs, ReasRMs emphasize leveraging long and coherent reasoning chains during the judging process to enhance the modelâs ability to assess and distinguish complex outputs accurately. We validate that integrating long reasoning chains during the judging process significantly enhances downstream reward model performance. We explore several strategies for adapting instruction-tuned language models into logically coherent ReasRMs. Notably, we find that solely applying reinforcement learning with verifiable rewards (RLVR) [12] in reward modeling does not fully realize the modelâs reasoning capabilities. We also observe that plain chain-of-thought (CoT) reasoning falls short at perceiving the fine-grained distinction across different question types.
Through a series of studies, we design a training pipeline that introduces reasoning distillation prior to RLVR, ultimately resulting in the development of RM-R1. To fully elicit the reasoning capability of RM-R1 for reward modeling, we design a Chain-of-Rubrics (CoR) process. Specifically, the model categorizes the input sample into one of two categories: chat or reasoning. For chat tasks, the model generates a set of evaluation rubrics, justifications for the rubrics, and evaluations tailored to the specific question. For reasoning tasks, correctness is the most important and generally preferred rubrics, so we directly let the model first solve the problem itself before evaluating and picking the preferred response. This task perception enables the model to tailor its rollout strategy â applying rubric-based evaluation for chat and correctness-first judgment for reasoning â resulting in more aligned and effective reward signals. In addition, we explore how to directly adapt existing reasoning models into reward models. Since these models have already undergone substantial reasoning-focused distillation, we fine-tune them using RLVR without additional distillation stages. Based on our training recipes, we produce RM-R1 models ranging from 7B to 32B.
Empirically, RM-R1 models consistently yield highly interpretable and coherent reasoning traces. On average, RM-R1 achieves state-of-the-art performance on RewardBench [17], RM-Bench [21], and RMB [43], outperforming 70B, 340B, GPT-4o, and Claude models by up to 4.9%. Beyond final performance, we conduct extensive empirical analyses of RM-R1, including ablations of our training recipes, studies of its scaling effects, comparisons with non-reasoning baselines, detailed case studies, and training dynamics. In summary, our main contributions are as follows:
- We demonstrate that reasoning abilities are crucial for reward models, and propose to formulate reward modeling as a reasoning process to enhance interpretability and accuracy.
- We design a training recipe based on reasoning-oriented distillation and RL that produces a set of reward models â RM-R1 â that can outperform larger models by up to 4.9% on average.
- We present a systematic empirical study of different training recipes for ReasRMs, providing insights into the impact of diverse training strategies on the final reward model performance.
## 2 RM-R1
<details>
<summary>x2.png Details</summary>

### Visual Description
## Diagram: Comparison of Reward Model Architectures and Training Paradigms
### Overview
This image is a technical diagram illustrating and comparing three different approaches to reward modeling (RM) for AI systems: **ScalarRM**, **GenRM**, and a proposed method called **RM-RI**. The diagram is structured to show the inference and training pipelines for each method, highlighting their differences in input, processing, and output. The overall flow demonstrates a progression from simple scalar scoring to generative judgment, and finally to a structured reasoning approach with explicit rubrics.
### Components/Axes
The diagram is divided into three main horizontal sections, each enclosed in a dashed box.
**1. Top Section: ScalarRM vs. GenRM**
* **ScalarRM (Left):**
* **Inference Input:** `Query x` (blue box), `Response y` (orange box).
* **Model Type:** `ScalarRM` (green rounded rectangle).
* **Process:** A `Linear Function` (depicted with a small graph icon).
* **Output:** `Score` (blue box).
* **GenRM (Right):**
* **Inference Input:** `Query x` (blue box), `{y1, y2}` (orange box, indicating multiple responses).
* **Model Type:** `GenRM` (green rounded rectangle).
* **Process:** Asks the question `"Which response is correct/better?"` (dashed box) with a `Judge` icon (document with a checkmark).
* **Output:** `Answer` (blue box).
**2. Middle Section: RM-RI Training**
This section is titled **RM-RI Training** in orange text and is further divided into three rows, each showing an "Inference" phase on the left and a "Training" phase on the right, connected by an arrow labeled **After Training**.
* **Row 1: GenRM-based Training**
* **Inference Phase:** Identical to the standalone GenRM above (Input: `Query x`, `{y1, y2}` -> Model: `GenRM` -> Task: `"Which response is correct/better?"` -> Output: `Answer`).
* **Training Phase:**
* **Training Input:** `Query x`, `{y1, y2}`.
* **Model Type:** `GenRM`.
* **Task/Objective:** `Distillation` with the goal to `Minimize NLL` (Negative Log-Likelihood).
* **Training Output:** `Reasoning Trace` (blue box).
* **Row 2: ReasRM-based Training**
* **Inference Phase:**
* **Inference Input:** `Query x`, `{y1, y2}`.
* **Model Type:** `ReasRM` (pink rounded rectangle).
* **Inference Task:** `"Let's verify step by step..."` (dashed box) with a `Critique` icon (document with a pencil).
* **Inference Output:** `Answer`.
* **Training Phase:**
* **Training Input:** `Query x`, `{y1, y2}`.
* **Model Type:** `ReasRM`.
* **Task/Objective:** `RL` (Reinforcement Learning) with the goal to `Maximize Cumulative Reward`.
* **Training Output:** `Reward Signal` denoted as `R(x, y)` (blue box).
* **Row 3: RM-RI (The Proposed Method)**
* **Inference Phase:**
* **Inference Input:** `Query x`, `{y1, y2}`.
* **Model Type:** `RM-RI` (purple rounded rectangle).
* **Inference Task:** A two-step process:
1. `"<rubrics> R1, R2, R3 </rubrics>"` (dashed box) with a `Chain-of-Rubrics` icon (magnifying glass over a list).
2. `"Let's verify step by step..."` (dashed box) with a `Complex Critique` icon (document with a pencil and a "100" badge).
* **Inference Output:** `Answer`.
* **Training Phase (Implied):** The arrow from the inference output points to a detailed example of the output format.
**3. Bottom Right: RM-RI's Structured Reasoning Output Example**
This box details the expected output format from the RM-RI model:
* `<rubrics> I. Empathy & Emotional Validation. II... III... </rubrics>`
* `<eval>The first response validates the user's emotions...</eval>`
* `<answer>The first response.</answer>`
### Detailed Analysis
The diagram systematically contrasts the three paradigms:
* **ScalarRM:** A traditional, direct regression approach. It maps a query-response pair to a single numerical score via a linear function. The process is opaque and non-explanatory.
* **GenRM:** A generative approach that reframes reward modeling as a multiple-choice question. It takes a query and a set of candidate responses and generates a natural language answer selecting the better one. The training objective is distillation (minimizing NLL) to produce a "Reasoning Trace."
* **ReasRM (Reasoning RM):** A step further, this model generates a step-by-step verification or critique before giving an answer. It is trained via Reinforcement Learning to maximize a cumulative reward signal `R(x, y)`.
* **RM-RI (Reward Modeling with Reasoning and Instruction):** The most complex method. Its inference has two explicit stages:
1. **Chain-of-Rubrics:** First, it generates or references a set of evaluation criteria (`R1, R2, R3`).
2. **Complex Critique:** It then performs a step-by-step verification using those rubrics.
The output is highly structured, containing the rubrics, an evaluation paragraph (`<eval>`), and a final answer (`<answer>`).
### Key Observations
1. **Progression of Complexity:** There is a clear evolution from a single scalar output (ScalarRM) to a simple generative choice (GenRM), to a reasoning trace (ReasRM), and finally to a structured, rubric-guided analysis (RM-RI).
2. **Shift in Training Objectives:** The training goals shift from `Minimize NLL` (GenRM) to `Maximize Cumulative Reward` via RL (ReasRM). RM-RI's training objective is not explicitly stated but is implied to build upon the reasoning and rubric-based structure.
3. **Output Interpretability:** The interpretability of the model's decision increases dramatically. ScalarRM provides only a number. GenRM provides a choice. ReasRM provides a verification trace. RM-RI provides explicit evaluation criteria and a structured justification.
4. **Spatial Layout:** The diagram uses a left-to-right flow for each model's pipeline and a top-to-bottom layout to show the progression of methods. The "After Training" arrows create a clear visual link between the inference and training phases for each model type in the RM-RI Training section.
### Interpretation
This diagram argues for a paradigm shift in reward modeling for AI alignment. It suggests that simple scalar rewards (ScalarRM) are insufficient for capturing nuanced human preferences. While generative models (GenRM) and reasoning models (ReasRM) improve upon this, the proposed **RM-RI** method introduces a crucial layer of **explicit, structured evaluation criteria (rubrics)**.
The core innovation appears to be the "Chain-of-Rubrics" step. By forcing the model to first articulate the evaluation standards (e.g., "Empathy & Emotional Validation"), the subsequent critique and final judgment become more transparent, consistent, and potentially more aligned with complex human values. This structure mimics how a human expert might evaluate responsesâagainst a predefined checklistâmaking the model's reasoning process auditable and its behavior more reliably steerable. The progression shown implies that future reward models should not just judge *which* response is better, but must be able to explain *why* according to explicit, human-understandable principles.
</details>
Figure 2: Training pipeline of RM-R1. Starting from an instruct model (GenRM), RM-R1 training involves two stages: Distillation and Reinforcement Learning (RL). In the Distillation stage, we use high-quality synthesized data to bootstrap RM-R1 âs reasoning ability. In the RL stage, RM-R1 âs reasoning ability for reward modeling is further strengthened. After distillation, a GenRM evolves into a ReasRM. RM-R1 further differentiates itself by being RL finetuned on preference data.
Figure 2 presents the overall training pipeline of RM-R1, which consists of two stages: reasoning distillation and reinforcement learning. (1) Reasoning Distillation: Starting from an off-the-shelf instruction-tuned model (e.g., Qwen-2.5-14B-Instruct), we further train the model using synthesized high-quality reasoning traces. This stage equips RM-R1 with essential reasoning capabilities required for effective reward modeling. (2) Reinforcement learning: While distillation is effective for injecting reasoning patterns, distilled models often overfit to specific patterns in the training data, limiting their generalization ability [9]. To overcome this limitation, we introduce a reinforcement learning phase that further optimizes the model, resulting in the final version of RM-R1.
### 2.1 Task Definition
Given a preference dataset:
$$
D=\{(x^(i),y_a^(i),y_b^(i),l^(i))\}_i=1^N, \tag{1}
$$
where $x$ is a prompt, $y_a$ and $y_b$ are two different responses for $x$ , and $lâ\{a,b\}$ is the ground truth label that indicates the preferred response. We define the generative reward modeling task as follows:
Let $r_θ$ denote a generative reward model parameterized by $θ$ . For each data sample, $r_θ$ generates a textual judgment $j$ consisting of ordered tokens $j=(j_1,j_2,\dots,j_T)$ , modeled by:
$$
r_θ(j|x,y_a,y_b)=â_t=1^Tr_θ(j_t|x,y_a,y_b,j_<
t). \tag{2}
$$
Note that $j$ contains $r_θ$ âs prediction of the preferred response $\hat{l}â j$ . The overall objective is:
$$
\max_r_{θ}E_(x,y_{a,y_b,l)âźD,\hat{l}âź r_
θ(j\mid x,y_a,y_b)}â¤ft[\mathbbm{1}(\hat{l}=l)\right]. \tag{3}
$$
### 2.2 Reasoning Distillation for Reward Modeling
For an instruction-tuned model (e.g., Qwen-2.5-14b-instruct [37]), it is quite intuitive to turn it into a GenRM simply by prompting. However, without fine-tuning on reward modeling reasoning traces, these models may struggle to conduct consistent judgments. To bootstrap its reasoning potential, we start with training an instruction-tuned model with long reasoning traces synthesized for reward modeling. Specifically, we sample $M$ data samples from $D$ and denote it as $D_\rm sub$ . Given a data sample $(x^(i),y_a^(i),y_b^(i),l^(i))âD_\rm sub$ , we ask an âoracleâ model like o3 or claude-3-7-sonnet to generate its structured reasoning trace $r^(i)$ justifying why $y_l^(i)$ is chosen as the preferred response of $x^(i)$ . We then construct the reasoning trace ground truth:
$$
y_trace^(i)=r^(i)â l^(i), \tag{4}
$$
where $â$ denotes string concatenation. Given all the synthesized reasoning traces $r^(i)$ , the final distillation dataset is defined as:
$$
D_distill=\{(x^(i),y_trace^(i))\}_i=1^M. \tag{5}
$$
Formally, the objective of distillation is to adjust $θ$ to maximize the likelihood of generating the desired reasoning trace and picking the response $y$ given the prompt $x$ . We minimize the negative log-likelihood (NLL) loss:
$$
L_distill(θ)=-â_(x,y)âD_
distillâ_tâ[|y|]\log r_θâ¤ft(y_t\mid x,y_<t\right), \tag{6}
$$
where $y_<t=(y_1,y_2,...,y_t-1)$ denotes the sequence of tokens preceding position $t$ . More details of generating high-quality reasoning chains are included in Appendix B.
### 2.3 RL Training
Although distillation is a proper way to turn a general generative model into a GenRM, it often suffers from overfitting to certain patterns and constrains the modelâs ability to generalize its reasoning abilities for critical thinking [9, 31] , which is essential for reward modeling. To address this, we propose to integrate RL as a more powerful learning paradigm to enhance the modelâs ability to conduct reasoning-based rewarding. Training a policy model using RL has been widely seen in the preference optimization phase of LLM post-training [24], and it is quite natural to extend this paradigm for training a ReasRM. To be specific, we directly treat our reward model $r_θ(j\mid x,y_a,y_b)$ as if it is a policy model:
$$
\max_r_{θ}E_(x,y_{a,y_b,l)âźD,\hat{l}âź r_
θ(j\mid x,y_a,y_b)}â¤ft[R(x,j)\right]-βD_
KLâ¤ft(r_θ\|r_ref\right), \tag{7}
$$
where $r_ref$ is the reference reward model. In practice, we use the checkpoint before RL training as $r_ref$ , and that means $r_ref$ could be an off-the-shelf LLM or the LLM obtained after the distillation step in Section 2.2. $R(x,j)$ is the reward function, and $D_KL$ is KL-divergence. The $x$ denotes input prompts drawn from the preference data $D$ . The $j$ indicates the text generated by the reward model, which includes the reasoning trace and final judgement $\hat{l}$ . In practice, we use Group Relative Policy Optimization (GRPO) [28] to optimize the objective in Equation 7, the details of which can be find in Appendix C.
#### 2.3.1 Chain-of-Rubrics (CoR) Rollout
Chain-of-Rubrics (CoR) Rollout for Instruct Models
Please act as an impartial judge and evaluate the quality of the responses provided by two AI Chatbots to the Clientâs question displayed below. First, classify the task into one of two categories: <type> Reasoning </type> or <type> Chat </type>. - Use <type> Reasoning </type> for tasks that involve math, coding, or require domain knowledge, multi-step inference, logical deduction, or combining information to reach a conclusion. - Use <type> Chat </type> for tasks that involve open-ended or factual conversation, stylistic rewrites, safety questions, or general helpfulness requests without deep reasoning. If the task is Reasoning: 1. Solve the Clientâs question yourself and present your final answer within <solution> ⌠</solution> tags. 2. Evaluate the two Chatbot responses based on correctness, completeness, and reasoning quality, referencing your own solution. 3. Include your evaluation inside <eval> ⌠</eval> tags, quoting or summarizing the Chatbots using the following tags: - <quote_A> ⌠</quote_A> for direct quotes from Chatbot A - <summary_A> ⌠</summary_A> for paraphrases of Chatbot A - <quote_B> ⌠</quote_B> for direct quotes from Chatbot B - <summary_B> ⌠</summary_B> for paraphrases of Chatbot B 4. End with your final judgment in the format: <answer>[[A]]</answer> or <answer>[[B]]</answer> If the task is Chat: 1. Generate evaluation criteria (rubric) tailored to the Clientâs question and context, enclosed in <rubric>âŚ</rubric> tags. 2. Assign weights to each rubric item based on their relative importance. 3. Inside <rubric>, include a <justify>âŚ</justify> section explaining why you chose those rubric criteria and weights. 4. Compare both Chatbot responses according to the rubric. 5. Provide your evaluation inside <eval>âŚ</eval> tags, using <quote_A>, <summary_A>, <quote_B>, and <summary_B> as described above. 6. End with your final judgment in the format: <answer>[[A]]</answer> or <answer>[[B]]</answer>
Figure 3: The system prompt used for RM-R1 rollout.
To facilitate the distilled models to proactively generate effective reasoning traces, we design a system prompt as shown in Figure 3 during rollout. Intuitively, reward modeling for general domain (e.g., chat, safety, etc.) and reasoning domain (e.g., math, code, etc.) should focus on different angles. For example, for the chat domain, we may care more about some aspects that can be expressed in textual rubrics (e.g., be polite), yet for the reasoning domain, we usually care more about logical coherence and answer correctness. Based on this intuition, we instruct ${r_θ}$ to classify each preference data sample $\{(x,y_c,y_r)\}$ into one of the two <type>: Chat or Reasoning. For each <type>, we prompt ${r_θ}$ to carry out the behavior corresponding to that type step by step: For reasoning tasks, we ask ${r_θ}$ to solve $x$ on its own. During the <eval> phase, $r_θ$ compares $y_c$ and $y_r$ conditioned on its own </solution> and selects an <answer>. Regarding the Chat type, we instead ask ${r_θ}$ to think about and justify the <rubric> for grading the chat quality (including safety).
#### 2.3.2 Reward Design
Rule-based reward mechanisms have demonstrated strong empirical performance to facilitate reasoning [12]. In our training, we further simplify the reward formulation and merely focus on the correctness-based component, in line with prior works [28, 18].
Formally, our reward is defined as follows:
$$
R(x,j|y_a,y_b)=\begin{cases}1&if \hat{l}=l,\\
-1&otherwise.\end{cases} \tag{8}
$$
where $\hat{l}$ is extracted from $j$ , wrapped between the <answer> and </answer> tokens. We have also tried adding the format reward to the overall reward, but found that the task performance does not have a significant difference. The rationale behind only focusing on correctness is that the distilled models have already learned to follow instructions and format their responses properly.
## 3 Experiments
### 3.1 Experimental Setup
We evaluate RM-R1 on three primary benchmarks: RewardBench [17], RM-Bench [21], and RMB [43]. Our training set utilizes a cleaned subset of Skywork Reward Preference 80K [20], 8K examples from Code-Preference-Pairs, and the full Math-DPO-10K [16] dataset. For baselines, we compare RM-R1 with RMs from three main categories: ScalarRMs, GenRMs, and ReasRMs. Further details on the benchmarks, dataset construction, and specific baseline models are provided in Appendix D.
Table 1: The performance comparison between best-performing baselines. Bold numbers indicate the best performance, Underlined numbers indicate the second best. The DeepSeek-GRM models are not open-weighted, so we use the numbers on their tech report. The more detailed numbers on RewardBench, RM-Bench, and RMB are in Appendix Table 6, Table 7, and Table 8
| Models | RewardBench | RM-Bench | RMB | Average |
| --- | --- | --- | --- | --- |
| ScalarRMs | | | | |
| SteerLM-RM-70B | 88.8 | 52.5 | 58.2 | 66.5 |
| Eurus-RM-7b | 82.8 | 65.9 | 68.3 | 72.3 |
| Internlm2-20b-reward | 90.2 | 68.3 | 62.9 | 73.6 |
| Skywork-Reward-Gemma-2-27B | 93.8 | 67.3 | 60.2 | 73.8 |
| Internlm2-7b-reward | 87.6 | 67.1 | 67.1 | 73.9 |
| ArmoRM-Llama3-8B-v0.1 | 90.4 | 67.7 | 64.6 | 74.2 |
| Nemotron-4-340B-Reward | 92.0 | 69.5 | 69.9 | 77.1 |
| Skywork-Reward-Llama-3.1-8B | 92.5 | 70.1 | 69.3 | 77.5 |
| INF-ORM-Llama3.1-70B | 95.1 | 70.9 | 70.5 | 78.8 |
| GenRMs | | | | |
| Claude-3-5-sonnet-20240620 | 84.2 | 61.0 | 70.6 | 71.9 |
| Llama3.1-70B-Instruct | 84.0 | 65.5 | 68.9 | 72.8 |
| Gemini-1.5-pro | 88.2 | 75.2 | 56.5 | 73.3 |
| Skywork-Critic-Llama-3.1-70B | 93.3 | 71.9 | 65.5 | 76.9 |
| GPT-4o-0806 | 86.7 | 72.5 | 73.8 | 77.7 |
| ReasRMs | | | | |
| JudgeLRM | 75.2 | 64.7 | 53.1 | 64.3 |
| DeepSeek-PairRM-27B | 87.1 | â | 58.2 | â |
| DeepSeek-GRM-27B-RFT | 84.5 | â | 67.0 | â |
| DeepSeek-GRM-27B | 86.0 | â | 69.0 | â |
| Self-taught-evaluator-llama3.1-70B | 90.2 | 71.4 | 67.0 | 76.2 |
| Our Methods | | | | |
| RM-R1-DeepSeek-Distilled-Qwen -7B | 80.1 | 72.4 | 55.1 | 69.2 |
| RM-R1-Qwen-Instruct -7B | 85.2 | 70.2 | 66.4 | 73.9 |
| RM-R1-Qwen-Instruct -14B | 88.2 | 76.1 | 69.2 | 77.8 |
| RM-R1-DeepSeek-Distilled-Qwen -14B | 88.9 | 81.5 | 68.5 | 79.6 |
| RM-R1-Qwen-Instruct -32B | 91.4 | 79.1 | 73.0 | 81.2 |
| RM-R1-DeepSeek-Distilled-Qwen -32B | 90.9 | 83.9 | 69.8 | 81.5 |
### 3.2 Main Results
Table 1 compares the overall performance of RM-R1 with existing strongest baseline models. The more detailed numbers on RewardBench, RM-Bench, and RMB are in Table 6, Table 7, and Table 8 in Appendix F. For the baselines, we reproduce the numbers if essential resources are open-sourced (e.g., model checkpoints, system prompts). Otherwise, we use the numbers reported in the corresponding tech report or benchmark leaderboard. For each benchmark, we select the best-performing models in each category for brevity. Our key findings are summarized below:
State-of-the-Art Performance. On average, our RM-R1-DeepSeek-Distilled-Qwen -14B model surpasses all previous leading Reward Models (RMs), including INF-ORM-Llama3.1-70B, Nemotron-4-340B-Reward, and GPT-4o, while operating at a much smaller scale. Our 32B models, RM-R1-Qwen-Instruct -32B and RM-R1-DeepSeek-Distilled-Qwen -32B, further extend this lead by a notable margin. The success of RM-R1 is attributable to both our meticulously designed training methodology and the effective scaling of our models, as extensively analyzed in Section 4.1 and Section 4.2. In particular, RM-R1 outperforms existing top-tier ScalarRMs. This highlights the considerable potential of ReasRMs, a category where prior GenRMs have exhibited suboptimal performance and are generally not comparable to their scalar counterparts. In contrast to our structured rollout and distillation with RLVR training strategy, prior critique-based methods have relied heavily on rejection sampling and unstructured, self-generated chain-of-thought (CoT) reasoning from instruct models [22, 35], limiting their reasoning capabilities and leading to inferior performance compared to ScalarRMs. Simultaneously, our comprehensive evaluation indicates that the top-performing scalar models on RewardBench do not consistently achieve state-of-the-art (SOTA) performance; in fact, larger models frequently underperform smaller ones. This evaluation underscores the need for a more comprehensive and systematic approach to RM assessment.
Effective Training towards Reasoning for Reward Modeling. Our specialized, reasoning-oriented training pipeline delivers substantial performance gains. For instance, RM-R1-Qwen-Instruct -14B consistently surpasses Self-taught-evaluator-llama-3.1-70B, a reasoning model five times its size. The RM-R1 model series also demonstrates impressive results on RM-Bench, exceeding the top-performing baseline by up to 8.7%. On this most reasoning-intensive benchmark, RM-R1-DeepSeek-Distilled-Qwen -32B establishes a new state-of-the-art. It achieves 91.8% accuracy in math and 74.1% in code, outperforming the previous best models (73% in math and 63% in code) by significant margins. Furthermore, it also records the strongest reasoning performance among our released models on RewardBench. Despite its performance, our Instruct -based models are remarkably data-efficient, reaching competitive performance using only 8.7K examples for distillationâcompared to the 800K examples used in training DeepSeek-Distilled [12]. Overall, our study underscores the significant potential of directly adapting large reasoning models into highly effective reward models.
## 4 Analysis
In this section, we present a series of empirical analyses to understand the key ingredients for training effective reasoning reward models. Our analysis spans scaling effects, design decisions, reasoning ablations, and a case study. We also present additional analysis on training dynamics in Section G.2.
### 4.1 Training Recipes
We first investigate the key ingredients underlying the successful training of RM-R1. Through a series of ablation studies, we examine our design choices to identify effective strategies for training high-quality reasoning reward models. We compare the following settings: Cold Start RL, Cold Start RL + Rubrics, Cold Start RL + Rubrics + Query Categorization (QC), and Distilled + RL + Rubrics + QC (i.e., RM-R1). The details of these settings are in Section G.1.
Table 2: Ablation study of the design choices for Reasoning Training on RewardBench.
| Method | Chat | Chat Hard | Safety | Reasoning | Average |
| --- | --- | --- | --- | --- | --- |
| Instruct (Original) | 95.8 | 74.3 | 86.8 | 86.3 | 85.8 |
| Instruct + Cold Start RL | 92.5 | 81.5 | 89.7 | 94.4 | 89.5 |
| Instruct + Cold Start RL + Rubrics | 93.0 | 82.5 | 90.8 | 94.2 | 90.1 |
| Instruct + Cold Start RL + Rubrics + QC | 92.3 | 82.6 | 91.6 | 96.3 | 90.8 |
| RM-R1 | 95.3 | 83.1 | 91.9 | 95.2 | 91.4 |
In Table 2, we present the results of the ablation studies described above, using the Qwen-2.5-Instruct-32B model as the Instruct (Original) model. Several key conclusions emerge:
- RL training alone is insufficient. While Cold Start RL slightly improves performance on hard chat and reasoning tasks, it fails to close the gap to fully optimized models.
- CoR prompting optimizes RM rollout and boosts reasoning performance. Instructing RM-R1 to self-generate chat rubrics or problem solutions before judgment helps overall performance, especially for chat and safety tasks. Incorporating explicit query categorization into the prompt notably improves reasoning performance, suggesting that clearer task guidance benefits learning.
- Distillation further enhances performance across all axes. Seeding the model with high-quality reasoning traces before RL yields the strongest results, with improvements observed on both hard tasks and safety-sensitive tasks.
Takeaway 1:
Directly replicating reinforcement learning recipes from mathematical tasks is insufficient for training strong reasoning reward models. Explicit query categorization and targeted distillation of high-quality reasoning traces are both crucial for achieving robust and generalizable improvements.
### 4.2 Scaling Effects
We then investigate how model performance varies with scale, considering both model size and inference-time compute. In some cases â such as ScalarRMs from InternLM2 [6] and Skywork [20] â the smaller models (7B/8B) outperforms the larger ones (20B/27B), showing no advantage of scaling. In this subsection, we show that this trend does not hold for RM-R1, where scaling brings clear and substantial improvements.
#### 4.2.1 Model Sizes
We first analyze the impact of model scale. Our study is based on the Qwen-2.5-Instruct model family at three sizes: 7B, 14B, and 32B. We evaluate performance improvements resulting from our training procedure described in Section 2, with results averaged across three key benchmarks: RewardBench, RM-Bench, and RMB.
For each model size, we compare the original and post-training performance. Figure 4(a) plots the relative improvement (%) with respect to model size. Observing an approximately linear trend, we fit a linear regression model and extrapolate to hypothetical scales of 3B and 72B, shown using faint markers and dashed extensions. The results strongly support a scaling law for reasoning reward models: larger models not only result in an absolute better final performance but also consistently yield greater performance gains. This aligns with the intuition that our training effectively leverages the superior reasoning capabilities of larger models.
<details>
<summary>x3.png Details</summary>

### Visual Description
## Scatter Plot with Linear Trend Line: Model Size Scaling
### Overview
This image is a scatter plot chart titled "Model Size Scaling." It illustrates the relationship between the size of a model (measured in billions of parameters, "B") and the corresponding percentage improvement in performance. The chart includes observed data points and a fitted trend line that extends beyond the observed data range.
### Components/Axes
* **Title:** "Model Size Scaling" (centered at the top).
* **X-Axis:** Labeled "Model Size (B)". The axis has discrete, non-linearly spaced tick marks at the values: 3, 7, 14, 32, and 72.
* **Y-Axis:** Labeled "Performance Improvement (%)". The axis has linear tick marks at integer intervals from 4 to 8.
* **Legend:** Located in the top-left corner of the plot area. It contains two entries:
* A black circle symbol labeled "Observed Data".
* A blue dashed line symbol labeled "Trend Line".
* **Grid:** A light gray, dashed grid is present in the background.
### Detailed Analysis
**Data Series and Points:**
1. **Observed Data (Black Circles):**
* The first point is located at approximately **Model Size = 7 B**, **Performance Improvement = ~4.2%**.
* The second point is located at approximately **Model Size = 14 B**, **Performance Improvement = ~4.9%**.
* The third point is located at approximately **Model Size = 32 B**, **Performance Improvement = ~5.8%**.
* **Trend Verification:** The three black data points show a clear upward trend from left to right.
2. **Trend Line (Blue Dashed Line):**
* The line is a straight, upward-sloping dashed line.
* It originates from a gray 'X' marker at the coordinate **(3, 4)**.
* It passes through or very near the three observed data points.
* It terminates at another gray 'X' marker at the coordinate **(72, ~8.3)**.
* **Trend Verification:** The line has a consistent positive slope, indicating a direct, linear relationship between the logarithm of model size (given the x-axis spacing) and performance improvement.
### Key Observations
* **Positive Correlation:** There is a strong, positive correlation between model size and performance improvement. As model size increases, performance improvement increases.
* **Consistent Increase:** The observed data points (7B, 14B, 32B) show a steady, nearly linear increase in performance improvement on this scale.
* **Extrapolation:** The trend line is extrapolated beyond the observed data range, predicting a performance improvement of approximately **8.3%** for a **72B** parameter model.
* **Anchor Points:** The trend line is explicitly anchored at the points (3, 4) and (72, ~8.3), marked with gray 'X's, which may represent theoretical or projected bounds.
### Interpretation
The chart demonstrates a scaling law for the model in question: increasing the number of parameters leads to predictable gains in performance. The linear fit on this plot (where the x-axis is logarithmically spaced) suggests that performance improvement scales linearly with the *logarithm* of model size. This is a common finding in neural scaling laws.
The data suggests that investing in larger models yields measurable benefits, but the rate of improvement per additional parameter may diminish (as moving from 7B to 14B gains ~0.7%, while moving from 14B to 32B gains ~0.9%, but over a much larger parameter increase). The extrapolation to 72B provides a forecast for future model development, though predictions beyond the observed data (32B) carry inherent uncertainty. The chart effectively communicates that model size is a critical and predictable driver of performance for this specific task or metric.
</details>
(a) Model Size
<details>
<summary>x4.png Details</summary>

### Visual Description
\n
## Line Chart: Inference Compute Scaling
### Overview
This is a line chart illustrating the relationship between a system's "Compute Budget" and its resulting "Performance" percentage. The chart demonstrates a positive, non-linear correlation where increasing the compute budget leads to improved performance, with the rate of improvement accelerating at higher budget levels.
### Components/Axes
* **Chart Title:** "Inference Compute Scaling" (positioned at the top center).
* **Y-Axis (Vertical):**
* **Label:** "Performance (%)"
* **Scale:** Linear scale ranging from 76 to 80.
* **Major Tick Marks:** 76, 77, 78, 79, 80.
* **X-Axis (Horizontal):**
* **Label:** "Compute Budget"
* **Scale:** Logarithmic scale (base 2), with values doubling at each major tick.
* **Major Tick Mark Labels:** "512", "1 k", "2 k", "4 k", "8 k". (Note: "k" denotes 1000).
* **Data Series:** A single series represented by a solid blue line connecting black circular data points.
* **Legend:** None present; the single data series is self-explanatory from the axis labels.
* **Grid:** Light gray dashed grid lines are present for both major x and y ticks.
### Detailed Analysis
**Trend Verification:** The single data series shows a clear upward trend. The slope is moderate from 512 to 2k, flattens slightly between 2k and 4k, and then increases sharply from 4k to 8k.
**Data Point Extraction (Approximate Values):**
The following values are estimated based on the position of the black data points relative to the y-axis grid lines.
1. **At Compute Budget = 512:** Performance â 75.9% (The point sits just below the 76% grid line).
2. **At Compute Budget = 1 k (1000):** Performance â 76.3% (The point is approximately one-third of the way between the 76% and 77% grid lines).
3. **At Compute Budget = 2 k (2000):** Performance â 77.2% (The point is slightly above the 77% grid line).
4. **At Compute Budget = 4 k (4000):** Performance â 77.3% (The point is marginally higher than the previous point at 2k, indicating a near plateau).
5. **At Compute Budget = 8 k (8000):** Performance â 79.6% (The point is significantly above the 79% grid line, closer to 80% than 79%).
### Key Observations
1. **Non-Linear Scaling:** Performance does not increase linearly with compute budget. The gains are modest at lower budgets, nearly stall between 2k and 4k, and then surge dramatically at 8k.
2. **Significant Final Jump:** The most substantial performance increase (approximately +2.3 percentage points) occurs in the final interval, from a compute budget of 4k to 8k.
3. **Diminishing then Accelerating Returns:** The pattern suggests a region of diminishing returns (2k to 4k) followed by a breakthrough or a different scaling regime at the highest measured budget.
### Interpretation
The chart provides a Peircean insight into the scaling behavior of an inference system. The data suggests that simply doubling the compute budget does not guarantee a proportional performance increase. The near-plateau between 2k and 4k could indicate a bottleneck elsewhere in the system (e.g., memory bandwidth, algorithmic efficiency) that is overcome at the 8k budget level. The sharp rise at 8k implies that the system may have entered a new, more efficient scaling phase, or that a critical threshold of compute was crossed, enabling a qualitative improvement in the model's capability. This pattern is crucial for resource allocation, indicating that investing in compute beyond a certain point (4k in this case) can yield disproportionately high returns, but only if the system architecture can effectively utilize it. The absence of data points beyond 8k leaves open the question of whether this accelerated scaling continues or eventually plateaus again.
</details>
(b) Inference Compute
Figure 4: Scaling effect of RM-R1. (a) Larger models benefit more from reasoning training. (b) Longer reasoning chains improve RM performance.
| Method | | RewardBench | RM-Bench | RMB | Avg. |
| --- | --- | --- | --- | --- | --- |
| Train on Full Data | | | | | |
| Instruct + SFT | | 90.9 | 75.4 | 65.9 | 77.4 |
| Instruct + Distilled + SFT | | 91.2 | 76.7 | 65.4 | 77.8 |
| RM-R1 * | | 91.4 | 79.1 | 73.0 | 81.2 |
| Train on 9k (Distillation) Data | | | | | |
| Instruct + SFT | | 88.8 | 74.8 | 66.9 | 76.6 |
| Instruct + Distilled * | | 89.0 | 76.3 | 72.0 | 79.2 |
Table 3: Comparison of reasoning-based training versus SFT across benchmarks. * indicates reasoning-based methods. Reasoning training consistently yields better performance.
#### 4.2.2 Inference-time Computation
Next, we examine how model performance varies with different compute budgets measured in number of tokens allowed during inference. Since this is particularly relevant to reasoning-focused models, we fix our base model to DeepSeek-R1-Distill-Qwen-14B. We evaluate average performance across the three key benchmarks using a wide range of inference-time compute budgets: 512, 1024, 2048, 4096, and 8192 tokens.
To ensure a fair comparison, we match the training rollout budget to the inference budget in each setting (i.e., we allow a maximum of $k$ tokens during training for a compute budget of $k$ at inference). All models are trained using GRPO with identical datasets and hyperparameter configurations. Figure 4(b) shows the relationship between compute budget and performance. We observe a clear improvement trend as the inference budget increases. This highlights the benefits of long reasoning chains in reward modeling.
Takeaway 2:
Scaling improves reward model performance: we observe a near-linear trend with both model size and inference-time compute. Larger models consistently benefit more from our reasoning-based training pipeline, and longer reasoning chains become increasingly effective under higher compute budgets.
### 4.3 Effectiveness of Reasoning Training
We now analyze the impact of reasoning-based training. Here, we demonstrate that reasoning-based training can outperform answer-only approaches. We consider the following settings:
Instruct + SFT. This approach fine-tunes the instruct model directly toward producing the correct final answer using the full dataset, without providing any intermediate reasoning chains.
Instruct + Distilled + SFT. This approach applies SFT (with respect to the answer directly) after the distillation stage, serving as a direct comparison point with RM-R1 trained with RL.
Instruct + RM-R1 (Distilled + RL). This is the full approach proposed in this paper, following the procedure detailed in Section 2.
Instruct + Distilled. This setting uses the model checkpoint immediately after the distillation stage, before any RL fine-tuning.
In summary, methods with â+ RM-R1 â or â+ Distilledâ represent reasoning-based approaches, while the remaining methods are purely non-reasoning-based approaches. In Table 3, we report the results measured across the three benchmarks. The findings clearly demonstrate that reasoning training significantly benefits reward model performance. Under fair comparisons (i.e., training on exactly the same amount of data), reasoning-based models consistently outperform their SFT-only counterparts. In particular, even high-quality distillation alone, applied to a small subset of the data, provides notable gains, highlighting the value of incorporating structured intermediate reasoning.
Takeaway 3:
Reasoning training substantially improves reward modeling. It not only enables better generalization across tasks but also provides consistent gains even under limited data scenarios compared to direct-answer SFT approaches.
### 4.4 Case Study
Table 4: Case Study of RM-R1 (succinct version). Complete Chatbots and model responses are shown in Table 9, 10, and 11. Cold-Start RL evaluation tends to focus on superficial features (e.g., number of suggestions, relevance), whereas RM-R1 produces high-quality rubrics (e.g., accuracy for factual questions), faithfully follows them, and evaluates responses based on their actual content.
Question: symptoms of sickle-cell disease Chatbot A: 1. Painful red or yellow skin lesions; 2. Frequent infections; 3. Fatigue and weakness; 4. Anemia; 5. Slower growth and development; 6. Jaundice; 7. Swelling of hands and feet; 8. Vision loss; 9. Tiredness and weakness after physical activity; 10. Breathlessness; 11. Fever, chills, and night sweats. Chatbot B Sickle cell disease is a genetic disorder that affects the shape of red blood cells. It is most common in African Americans, but can also affect people from other ethnic groups. The symptoms of sickle cell disease can vary widely from person to person, but some common symptoms include: 1. Fatigue: âŚ; 2. Pain: âŚ; 3. Fever: âŚ; 4. Anemia: âŚ; 5. Delayed growth and development: âŚ; 6. Infections: âŚ; 7. Pigmentation: ⌠Itâs important to note that not everyone with sickle cell disease will experience all of these symptoms, ⌠Therefore, itâs important to consult with a healthcare provider if you suspect you or someone you know may have sickle cell disease. Cold-start RL evaluation â <type>| Chat\textcolor{deepblue}{\verb </type>| <rubric>| \textcolor{red}{\textbf{Relevance 40 \%}} â Comprehensiveness 30 \% â Clarity/organisation 30 \% \textcolor{red}{\verb </rubric>| <eval>| \textcolor{red}{\textbf{Chatbot A provided a list of symptoms that are directly relevant to sickle cell disease.}} The response is comprehensive and includes a wide range of symptoms commonly associated with the condition. The symptoms are listed in a clear and organized manner, making it easy for the client to understand the information. Chatbot Bâs response is less organized than Chatbot Aâs response, and the explanation of sickle cell disease may be unnecessary for the clientâs question.\textcolor{orange}{\verb </eval>| <answer>| [[A]]\textcolor{lightred}{\verb </answer>| RM-R1 evaluation â <type>| Chat\textcolor{deepblue}{\verb </type>| <rubric>| 1 \textcolor{ForestGreen}{\textbf{Accuracy 40 \%}} ¡ 2 Comprehensiveness 30 \% ¡ 3 Clarity 20 \% ¡ 4 Helpfulness 10 \% \textcolor{deepred}{\verb <justify>| Accuracy is weighted highest because providing correct information about medical conditions is crucial. Comprehensiveness is important because ⌠</justify>| \textcolor{red}{\verb </rubric>| <eval>| \emph{Accuracy of information:} Chatbot A provides a list of symptoms without explanation, but \textcolor{ForestGreen}{\textbf{some listed items (e.g.\ âpainful red or yellow skin lesions,â âvision lossâ) are not typical of sickle-cell disease}}. Chatbot B offers more accurate information, explaining each symptomâs nature and severity. ... \emph{Comprehensiveness:}...\emph{Clarity and organization:}...\emph{Helpfulness:}... \textcolor{orange}{\verb </eval>| <answer>| [[B]]\textcolor{lightred}{\verb </answer>|
To gain deeper insights into RM-R1, we further conduct a case study using Qwen-14B-Instruct, comparing its behavior with a cold-start RL model. The results, summarized in Table 4, reveal the following key observations:
Transparent Judging Process.
The reasoning traces generated by RM-R1 are highly interpretable and coherent, reflecting the modelâs perception of human preferences. It explicitly articulates why certain responses are better, providing transparency into its evaluation process.
High-Quality, Question-Dependent Rubrics.
RM-R1 accurately understands the question and the context of comparison, correctly prioritizing âaccuracyâ as the most critical rubric for medical-related questions. In contrast, the cold-start RL model often overlooks the most important factors and instead emphasizes superficial or broadly defined features (e.g., relevance) that are less discriminative. The ability to generate high-quality, question-specific rubrics stems from the knowledge acquired during the distillation stage.
Faithful Adherence to Rubrics and Content-Based Judgement.
RM-R1 grounds its evaluation in the actual content of the chatbot responses. For example, it correctly identifies inaccuracies in Chatbot Aâs response based on factual content rather than surface presentation. Furthermore, it systematically evaluates different aspects of the rubric, leading to a structured, interpretable, and verifiable judging process.
## 5 Related Work
Reward Models (RMs).
Early RMs were typically outcome-focused: trained to predict human preference rankings for complete outputs [42]. Recent advances have looked at providing process supervision, which rewards or evaluates the steps of a modelâs reasoning rather than only the final answer. A series of works propose to train process reward models that judge the correctness of intermediate reasoning steps [19, 10, 27]. A limitation of many PRMs is their heavy reliance on curated step-level human labels or specific schemas, and they often remain domain-specific. Zhang et al. [40] propose Generative Verifiers, framing reward modeling as a next-token prediction task. This allows the reward model to leverage chain-of-thought and even use majority voting over multiple sampled rationales to make more reliable judgments. DeepSeek-GRM [22] and JudgeLRM [7] have studied using reasoning models as generative reward models, which are the most relevant research to ours. However, their main focus is on the effect of scaling inference-time computation on reward modeling. On the contrary, our work is the first to provide a systematic empirical comparison of different reward model training paradigms, shedding light on when and why a distilled and RL-trained reward model like RM-R1 has advantages over the conventional approaches.
Reinforcement Learning from Human Feedback (RLHF).
Early works [8] first demonstrated that reinforcement learning could optimize policies using a reward model trained from human pairwise preferences. Subsequent studies applied RLHF to large-scale language models using policy optimization algorithms such as PPO [26]. For example, Ziegler et al. [45] fine-tuned GPT-2 via PPO on human preference rewards, and Stiennon et al. [32] showed that RLHF could significantly improve the quality of summarization by optimizing against a learned preference model. More recently, Ouyang et al. [24] used a similar PPO-based pipeline to train InstructGPT, establishing the modern RLHF paradigm for instruction-following models. Recently, Verifiable supervision techniques have also emerged: DeepSeek-R1 [12] uses a form of self-verification during RLHF to reward correct reasoning steps, rather than only final-answer quality. This method incentivizes policies to produce outputs that can be verified for correctness, bridging the gap between pure preference-based feedback and ground-truth signals. However, even with such innovations, most RLHF implementations still treat reward modeling and reasoning as separate stages.
## 6 Conclusion and Future Work
In this paper, we revisited reward modeling through the lens of reasoning. We introduced RM-R1, a family of ReasRMs that effectively generate explicit chains of rubrics and rationales, and scale with both model size and inference compute. Across three public benchmarks, RM-R1 matched or surpassed commercial and open-source RMs while producing more interpretable judgments. Ablation investigations reveal that (1) task-type categorization, (2) bootstrapping from high-quality reasoning traces, and (3) RL fine-tuning are all indispensable. Qualitative analyses further showed that RM-R1 learns to prioritize high-impact rubrics, faithfully follow its own criteria and justify coherently. Future work includes active preference collection, where ReasRMs use active learning to query human preference only when the current rubric set is insufficient for a new preference sample. Finally, it would be natural to extend our study to multimodal/agentic reward modeling scenarios.
## Acknowledgments and Disclosure of Funding
This research is based upon work supported DARPA ITM Program No. FA8650-23-C-7316, and the AI Research Institutes program by National Science Foundation and the Institute of Education Sciences, U.S. Department of Education through Award # 2229873 - AI Institute for Transforming Education for Children with Speech and Language Processing Challenges. The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies, either expressed or implied, of the U.S. Government. The U.S. Government is authorized to reproduce and distribute reprints for governmental purposes notwithstanding any copyright annotation therein.
## References
- Achiam et al. [2023] Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, et al. Gpt-4 technical report. arXiv preprint arXiv:2303.08774, 2023.
- Adler et al. [2024] Bo Adler, Niket Agarwal, Ashwath Aithal, Dong H Anh, Pallab Bhattacharya, Annika Brundyn, Jared Casper, Bryan Catanzaro, Sharon Clay, et al. Nemotron-4 340b technical report. arXiv preprint arXiv:2406.11704, 2024.
- Anthropic [2024] AI Anthropic. The claude 3 model family: Opus, sonnet, haiku. Claude-3 Model Card, 1:1, 2024.
- Bai et al. [2022] Yuntao Bai, Saurav Kadavath, Sandipan Kundu, Amanda Askell, Jackson Kernion, Andy Jones, Anna Chen, Anna Goldie, Azalia Mirhoseini, Cameron McKinnon, et al. Constitutional ai: Harmlessness from ai feedback. arXiv preprint arXiv:2212.08073, 2022.
- Baker et al. [2009] Chris L Baker, Rebecca Saxe, and Joshua B Tenenbaum. Action understanding as inverse planning. Cognition, 113(3):329â349, 2009.
- Cai et al. [2024] Zheng Cai, Maosong Cao, Haojiong Chen, Kai Chen, Keyu Chen, Xin Chen, Xun Chen, Zehui Chen, Zhi Chen, Pei Chu, et al. Internlm2 technical report. arXiv preprint arXiv:2403.17297, 2024.
- Chen et al. [2025] Nuo Chen, Zhiyuan Hu, Qingyun Zou, Jiaying Wu, Qian Wang, Bryan Hooi, and Bingsheng He. Judgelrm: Large reasoning models as a judge. arXiv preprint arXiv:2504.00050, 2025.
- Christiano et al. [2017] Paul F Christiano, Jan Leike, Tom Brown, Miljan Martic, Shane Legg, and Dario Amodei. Deep reinforcement learning from human preferences. Advances in neural information processing systems, 30, 2017.
- Chu et al. [2025] Tianzhe Chu, Yuexiang Zhai, Jihan Yang, Shengbang Tong, Saining Xie, Dale Schuurmans, Quoc V Le, Sergey Levine, and Yi Ma. Sft memorizes, rl generalizes: A comparative study of foundation model post-training. arXiv preprint arXiv:2501.17161, 2025.
- Cui et al. [2025] Ganqu Cui, Lifan Yuan, Zefan Wang, Hanbin Wang, Wendi Li, Bingxiang He, Yuchen Fan, Tianyu Yu, Qixin Xu, Weize Chen, et al. Process reinforcement through implicit rewards. arXiv preprint arXiv:2502.01456, 2025.
- Dubey et al. [2024] Abhimanyu Dubey, Abhinav Jauhri, Abhinav Pandey, Abhishek Kadian, Ahmad Al-Dahle, Aiesha Letman, Akhil Mathur, Alan Schelten, Amy Yang, Angela Fan, et al. The llama 3 herd of models. arXiv preprint arXiv:2407.21783, 2024.
- Guo et al. [2025] Daya Guo, Dejian Yang, Haowei Zhang, Junxiao Song, Ruoyu Zhang, Runxin Xu, Qihao Zhu, Shirong Ma, Peiyi Wang, Xiao Bi, et al. Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning. arXiv preprint arXiv:2501.12948, 2025.
- Hu et al. [2024] Jian Hu, Xibin Wu, Zilin Zhu, Xianyu, Weixun Wang, Dehao Zhang, and Yu Cao. Openrlhf: An easy-to-use, scalable and high-performance rlhf framework. arXiv preprint arXiv:2405.11143, 2024.
- Hurst et al. [2024] Aaron Hurst, Adam Lerer, Adam P Goucher, Adam Perelman, Aditya Ramesh, Aidan Clark, AJ Ostrow, Akila Welihinda, Alan Hayes, Alec Radford, et al. Gpt-4o system card. arXiv preprint arXiv:2410.21276, 2024.
- Ivison et al. [2024] Hamish Ivison, Yizhong Wang, Jiacheng Liu, Zeqiu Wu, Valentina Pyatkin, Nathan Lambert, Noah A Smith, Yejin Choi, and Hannaneh Hajishirzi. Unpacking dpo and ppo: Disentangling best practices for learning from preference feedback. arXiv preprint arXiv:2406.09279, 2024.
- Lai et al. [2024] Xin Lai, Zhuotao Tian, Yukang Chen, Senqiao Yang, Xiangru Peng, and Jiaya Jia. Step-dpo: Step-wise preference optimization for long-chain reasoning of llms. arXiv preprint arXiv:2406.18629, 2024.
- Lambert et al. [2024] Nathan Lambert, Valentina Pyatkin, Jacob Morrison, LJ Miranda, Bill Yuchen Lin, Khyathi Chandu, Nouha Dziri, Sachin Kumar, Tom Zick, Yejin Choi, et al. Rewardbench: Evaluating reward models for language modeling. arXiv preprint arXiv:2403.13787, 2024.
- Li et al. [2025] Xuefeng Li, Haoyang Zou, and Pengfei Liu. Torl: Scaling tool-integrated rl. arXiv preprint arXiv:2503.23383, 2025.
- Lightman et al. [2023] Hunter Lightman, Vineet Kosaraju, Yuri Burda, Harrison Edwards, Bowen Baker, Teddy Lee, Jan Leike, John Schulman, Ilya Sutskever, and Karl Cobbe. Letâs verify step by step. In The Twelfth International Conference on Learning Representations, 2023.
- Liu et al. [2024a] Chris Yuhao Liu, Liang Zeng, Jiacai Liu, Rui Yan, Jujie He, Chaojie Wang, Shuicheng Yan, Yang Liu, and Yahui Zhou. Skywork-reward: Bag of tricks for reward modeling in llms. arXiv preprint arXiv:2410.18451, 2024a.
- Liu et al. [2024b] Yantao Liu, Zijun Yao, Rui Min, Yixin Cao, Lei Hou, and Juanzi Li. Rm-bench: Benchmarking reward models of language models with subtlety and style. arXiv preprint arXiv:2410.16184, 2024b.
- Liu et al. [2025] Zijun Liu, Peiyi Wang, Runxin Xu, Shirong Ma, Chong Ruan, Peng Li, Yang Liu, and Yu Wu. Inference-time scaling for generalist reward modeling. arXiv preprint arXiv:2504.02495, 2025.
- Montibeller and Franco [2010] Gilberto Montibeller and Alberto Franco. Multi-criteria decision analysis for strategic decision making. In Handbook of multicriteria analysis, pages 25â48. Springer, 2010.
- Ouyang et al. [2022] Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, et al. Training language models to follow instructions with human feedback. Advances in neural information processing systems, 35:27730â27744, 2022.
- Reid et al. [2024] Machel Reid, Nikolay Savinov, Denis Teplyashin, Dmitry Lepikhin, Timothy Lillicrap, Jean-baptiste Alayrac, Radu Soricut, Angeliki Lazaridou, Orhan Firat, Julian Schrittwieser, et al. Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context. arXiv preprint arXiv:2403.05530, 2024.
- Schulman et al. [2017] John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov. Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347, 2017.
- Setlur et al. [2024] Amrith Setlur, Chirag Nagpal, Adam Fisch, Xinyang Geng, Jacob Eisenstein, Rishabh Agarwal, Alekh Agarwal, Jonathan Berant, and Aviral Kumar. Rewarding progress: Scaling automated process verifiers for llm reasoning. arXiv preprint arXiv:2410.08146, 2024.
- Shao et al. [2024] Zhihong Shao, Peiyi Wang, Qihao Zhu, Runxin Xu, Junxiao Song, Xiao Bi, Haowei Zhang, Mingchuan Zhang, YK Li, Y Wu, et al. Deepseekmath: Pushing the limits of mathematical reasoning in open language models. arXiv preprint arXiv:2402.03300, 2024.
- Sheng et al. [2024] Guangming Sheng, Chi Zhang, Zilingfeng Ye, Xibin Wu, Wang Zhang, Ru Zhang, Yanghua Peng, Haibin Lin, and Chuan Wu. Hybridflow: A flexible and efficient rlhf framework. arXiv preprint arXiv: 2409.19256, 2024.
- Shiwen et al. [2024] Tu Shiwen, Zhao Liang, Chris Yuhao Liu, Liang Zeng, and Yang Liu. Skywork critic model series. https://huggingface.co/Skywork, September 2024. URL https://huggingface.co/Skywork.
- Stanton et al. [2021] Samuel Stanton, Pavel Izmailov, Polina Kirichenko, Alexander A Alemi, and Andrew G Wilson. Does knowledge distillation really work? Advances in neural information processing systems, 34:6906â6919, 2021.
- Stiennon et al. [2020] Nisan Stiennon, Long Ouyang, Jeffrey Wu, Daniel Ziegler, Ryan Lowe, Chelsea Voss, Alec Radford, Dario Amodei, and Paul F Christiano. Learning to summarize with human feedback. Advances in Neural Information Processing Systems, 33:3008â3021, 2020.
- Van Hoeck et al. [2015] Nicole Van Hoeck, Patrick D Watson, and Aron K Barbey. Cognitive neuroscience of human counterfactual reasoning. Frontiers in human neuroscience, 9:420, 2015.
- Wang et al. [2024a] Haoxiang Wang, Wei Xiong, Tengyang Xie, Han Zhao, and Tong Zhang. Interpretable preferences via multi-objective reward modeling and mixture-of-experts. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen, editors, Findings of the Association for Computational Linguistics: EMNLP 2024, pages 10582â10592, Miami, Florida, USA, November 2024a. Association for Computational Linguistics. URL https://aclanthology.org/2024.findings-emnlp.620.
- Wang et al. [2024b] Tianlu Wang, Ilia Kulikov, Olga Golovneva, Ping Yu, Weizhe Yuan, Jane Dwivedi-Yu, Richard Yuanzhe Pang, Maryam Fazel-Zarandi, Jason Weston, and Xian Li. Self-taught evaluators. arXiv preprint arXiv:2408.02666, 2024b.
- Wang et al. [2024c] Zhilin Wang, Yi Dong, Jiaqi Zeng, Virginia Adams, Makesh Narsimhan Sreedhar, Daniel Egert, Olivier Delalleau, Jane Scowcroft, Neel Kant, Aidan Swope, and Oleksii Kuchaiev. HelpSteer: Multi-attribute helpfulness dataset for SteerLM. In Kevin Duh, Helena Gomez, and Steven Bethard, editors, Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), pages 3371â3384, Mexico City, Mexico, June 2024c. Association for Computational Linguistics. URL https://aclanthology.org/2024.naacl-long.185.
- Yang et al. [2024] An Yang, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu, Chengyuan Li, Dayiheng Liu, Fei Huang, Haoran Wei, et al. Qwen2. 5 technical report. arXiv preprint arXiv:2412.15115, 2024.
- Yu et al. [2024] Yue Yu, Zhengxing Chen, Aston Zhang, Liang Tan, Chenguang Zhu, Richard Yuanzhe Pang, Yundi Qian, Xuewei Wang, Suchin Gururangan, Chao Zhang, et al. Self-generated critiques boost reward modeling for language models. arXiv preprint arXiv:2411.16646, 2024.
- Yuan et al. [2024] Lifan Yuan, Ganqu Cui, Hanbin Wang, Ning Ding, Xingyao Wang, Jia Deng, Boji Shan, Huimin Chen, Ruobing Xie, Yankai Lin, Zhenghao Liu, Bowen Zhou, Hao Peng, Zhiyuan Liu, and Maosong Sun. Advancing llm reasoning generalists with preference trees, 2024.
- Zhang et al. [2024] Lunjun Zhang, Arian Hosseini, Hritik Bansal, Mehran Kazemi, Aviral Kumar, and Rishabh Agarwal. Generative verifiers: Reward modeling as next-token prediction. arXiv preprint arXiv:2408.15240, 2024.
- Zheng et al. [2023] Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, Siyuan Zhuang, Zhanghao Wu, Yonghao Zhuang, Zi Lin, Zhuohan Li, Dacheng Li, Eric Xing, et al. Judging llm-as-a-judge with mt-bench and chatbot arena. Advances in Neural Information Processing Systems, 36:46595â46623, 2023.
- Zhong et al. [2025] Jialun Zhong, Wei Shen, Yanzeng Li, Songyang Gao, Hua Lu, Yicheng Chen, Yang Zhang, Wei Zhou, Jinjie Gu, and Lei Zou. A comprehensive survey of reward models: Taxonomy, applications, challenges, and future. arXiv preprint arXiv:2504.12328, 2025.
- Zhou et al. [2024] Enyu Zhou, Guodong Zheng, Binghai Wang, Zhiheng Xi, Shihan Dou, Rong Bao, Wei Shen, Limao Xiong, Jessica Fan, Yurong Mou, et al. Rmb: Comprehensively benchmarking reward models in llm alignment. arXiv preprint arXiv:2410.09893, 2024.
- Zhu et al. [2023] Banghua Zhu, Evan Frick, Tianhao Wu, Hanlin Zhu, and Jiantao Jiao. Starling-7b: Improving llm helpfulness & harmlessness with rlaif, November 2023.
- Ziegler et al. [2019] Daniel M Ziegler, Nisan Stiennon, Jeffrey Wu, Tom B Brown, Alec Radford, Dario Amodei, Paul Christiano, and Geoffrey Irving. Fine-tuning language models from human preferences. arXiv preprint arXiv:1909.08593, 2019. Contents
1. 1 Introduction
1. 2 RM-R1
1. 2.1 Task Definition
1. 2.2 Reasoning Distillation for Reward Modeling
1. 2.3 RL Training
1. 2.3.1 Chain-of-Rubrics (CoR) Rollout
1. 2.3.2 Reward Design
1. 3 Experiments
1. 3.1 Experimental Setup
1. 3.2 Main Results
1. 4 Analysis
1. 4.1 Training Recipes
1. 4.2 Scaling Effects
1. 4.2.1 Model Sizes
1. 4.2.2 Inference-time Computation
1. 4.3 Effectiveness of Reasoning Training
1. 4.4 Case Study
1. 5 Related Work
1. 6 Conclusion and Future Work
1. A User Prompt for DeepSeek-Distilled Reasoning Models
1. B Details of Reasoning Chain Generation
1. C Group Relative Policy Optimization (GRPO)
1. D Experiment Setups
1. D.1 Benchmarks
1. D.2 Preference Datasets
1. D.3 Baselines
1. E Implementation Details
1. F Full Experiment Result
1. G Supplementary Information for Section 4
1. G.1 Ablation Settings
1. G.2 Training Dynamics
## Appendix A User Prompt for DeepSeek-Distilled Reasoning Models
Large reasoning models such as DeepSeek-R1-distilled models [12] do not have a system prompt, so we show the user prompt for rollouts in Figure 5.
Chain-of-Rubrics (CoR) Rollout for Reasoning Models
Please act as an impartial judge and evaluate the quality of the responses provided by two AI Chatbots to the Client question displayed below. ⌠[Pairwise Input Content] ⌠Output your final verdict at last by strictly following this format: â<answer>[[A]]</answer>â if Chatbot A is better, or â<answer>[[B]]</answer>â if Chatbot B is better.
Figure 5: The user prompt used for RM-R1 rollout (for reasoning models).
## Appendix B Details of Reasoning Chain Generation
We now expand on the details of generating high-quality reasoning chains. We first use the same prompt to query Claude-3.7-Sonnet, generating initial reasoning traces. However, approximately 25% of these traces are incorrect, primarily on harder chat tasks. To correct these cases, we pass the original prompt, the incorrect trace, and the correct final answer to OpenAI-O3, which then generates a corrected reasoning trace aligned with the right answer.
This two-stage process yields a high-quality distillation set. We deliberately choose the orderâfirst Claude, then O3 âbased on qualitative observations: Claude excels at solving easier tasks and maintaining attention to safety considerations, whereas O3 performs better on harder tasks but tends to overemphasize helpfulness at the expense of safety. We select approximately 12% of the training data (slightly fewer than 9K examples) for distillation. This is then followed by RL training.
## Appendix C Group Relative Policy Optimization (GRPO)
Group Relative Policy Optimization (GRPO) [28] is a variant of Proximal Policy Optimization (PPO) [26], which obviates the need for additional value function approximation, and uses the average reward of multiple sampled outputs produced in response to the same prompt as the baseline. More specifically, for each prompt $x$ , GRPO samples a group of outputs $\{y_1,y_2,âŻ,y_G\}$ from the old policy $Ď_θ_{old}$ and then optimizes the policy model by maximizing the following objective:
$$
\begin{split}J_GRPO(θ)= &E_xâźD
, \{j_i\_i=1^G⟠r_θ_{old}(j\mid x)}\bigg{[}\frac{1}{G
}â_i=1^G\frac{1}{|j_i|}â_t=1^|j_i|\Big{\{}\min\Big{(}\frac{r
_θ(j_i,t\mid x,j_i,<t)}{r_θ_{old}(j_i,t\mid x,j_i,
<t)}\hat{A}_i,t,\\[4.0pt]
&clip\big{(}\frac{r_θ(j_i,t\mid x,j_i,<t)}{r_θ_{
old}(j_i,t\mid x,j_i,<t)},1-Îľ,1+Îľ\big{)}\hat{A}_i,t\Big{
)}-β D_KLâ¤ft[r_θ(¡\mid x) \| Ď_
ref(¡\mid x)\right]\Big{\}}\bigg{]},\end{split} \tag{9}
$$
where $β$ is a hyperparameter balancing the task specific loss and the KL-divergence. Specifically, $\hat{A}_i$ is computed using the rewards of a group of responses within each group $\{r_1,r_2,âŚ,r_G\}$ , and is given by the following equation:
$$
\hat{A}_i=\frac{r_i-{mean(\{r_1,r_2,âŻ,r_G\})}}{{
std(\{r_1,r_2,âŻ,r_G\})}}. \tag{10}
$$
## Appendix D Experiment Setups
### D.1 Benchmarks
In this paper, we consider the following three benchmarks:
RewardBench [17]: RewardBench is one of the first endeavors towards benchmarking reward models through prompt-chosen-rejected trios, covering four categories: chat, chat-hard, reasoning, and safety, with 358, 456, 740, and 1431 samples, respectively.
RM-Bench [21]: Building on RewardBench, RM-Bench evaluates reward models for their sensitivity to subtle content differences and robustness against style biases. It includes four categories: Chat, Safety, Math, and Code, with 129, 441, 529, and 228 samples, respectively. Each sample contains three prompts of varying difficulty. RM-Bench is the most reasoning-intensive benchmark among those we consider.
RMB [43]: Compared with RewardBench and RM-Bench, RMB offers a more comprehensive evaluation of helpfulness and harmlessness. It includes over 49 real-world scenarios and supports both pairwise and Best-of-N (BoN) evaluation formats. RMB comprises 25,845 instances in totalâ37 scenarios under the helpfulness alignment objective and 12 under harmlessness.
### D.2 Preference Datasets
We consider the following datasets for training:
Skywork Reward Preference 80K [20] is a high-quality collection of pairwise preference data drawn from a variety of domains, including chat, safety, mathematics, and code. It employs an advanced data filtering technique to ensure preference reliability across tasks. However, we identify a notable issue with this dataset: all samples from the magpie_ultra source exhibit a strong spurious correlation, where rejected responses consistently contain the token â <im_start>,â while accepted responses do not. Additionally, responses from this source show a systematic biasâaccepted responses are typically single-turn, while rejected responses are multi-turn. This problematic subset constitutes approximately 30% of the Skywork dataset and primarily covers mathematics and code domains. To avoid introducing spurious correlations into training, we exclude all magpie_ultra data and retain only the cleaned subset for our experiments.
Code-Preference-Pairs is a high-quality coding preference dataset. It is constructed by prompting a model with original code, introducing deliberate bugs, and manipulating examples (e.g., swapping broken and corrected versions, removing error comments) to generate fine-grained preference pairs. We subsample 8K examples from this dataset for use in our experiments.
Math-DPO-10K [16] is a high-quality stepwise preference dataset focused on mathematical reasoning. We use the full dataset in our experiments.
A global statistics of our training dataset is summarized in Table 5.
Table 5: Global Statistics of our Training Dataset. * indicates the source is from Skywork-Reward-Preference-80K-v0.2.
| Source | Size | Domain |
| --- | --- | --- |
| magpile_pro_llama3.1* | 29682 | Reasoning |
| offset_bias* | 8504 | Chat (length bias) |
| helpsteer2* | 7221 | Chat |
| wildguard* | 6709 | Safety |
| magpile_pro* | 2030 | Chat |
| Code-Preference-Pairs | 8000 | Reasoning |
| Math-DPO-10K | 10000 | Reasoning |
### D.3 Baselines
We compare RM-R1 with RMs from three categories:
ScalarRMs. ScalarRMs produce a score for model response directly, predicting preference through single numeric scores without explicit reasoning traces. This category includes models such as Eurus-RM [39], Internlm2 [6] SteerLM-RM [36], Nemotron-RM [2], Tulu-v2.5 [15], Starling-RM [44], ArmoRM [34], Skywork-RM [20], etc. While these models often achieve strong results on well-defined benchmarks, they generally lack interpretability and struggle to capture fine-grained reasoning.
GenRMs. Generative reward models (GenRMs) offer more expressive feedback by producing free-form textual judgments, typically without further training. This includes the widely used LLM-as-a-Judge setup [41], where pretrained language models are prompted to explain and evaluate responses. We also categorize under GenRMs models that directly generate output answers without intermediate reasoning steps. Representative examples include LLaMA [11], Qwen [37], Claude [3], GPT-4o [1, 14], Gemini 1.5 Pro [25], and Skywork-Critic [30]. By leveraging LLMsâ generative capabilities, these models enhance interpretability through natural language rationales and explanations.
ReasRMs. Reasoning-enhanced reward models (ReasRMs) explicitly incorporate reasoning processes before their final judgments, often trained through critiques or chain-of-thought strategies. Notable examples are JudgeLRM [7], Critique-RM [38], DeepSeek-GRM [22], Self-taught Evaluators [35] and our proposed RM-R1 models. These models excel in tasks demanding rigorous reasoning, safety evaluations, and nuanced preference judgments due to their grounding in structured critical thinking.
## Appendix E Implementation Details
Our training framework is based on VERL [29] and OpenRLHF [13]. For Instruct models, we use 8.7k data for distillation and 64k for RLVR. For Deepseek-Distilled models, we use the full data for RLVR.
Distillation Stage. We use the SFTTrainer from OpenRLHF with a fixed batch size of 128 and a micro-batch size of 1, training for a single epoch. To optimize GPU memory usage, we enable gradient checkpointing, FlashAttention, and Adam offloading. The learning rates are set based on the model size: $5e{-6}, 3e{-6},$ and $2e{-6}$ for models of size $7B, 14B,$ and $32B$ , respectively.
RLVR Stage. We use the VERL framework for all GRPO training. The training batch size is fixed at 1024, with a mini-batch size of 128. We adopt Fully Sharded Data Parallel (FSDP) to improve memory efficiency. For rollout generation, we use vLLM with tensor parallelism size 4 and GPU memory utilization capped at 0.4. Sampling follows default parameters (temperature = 1.0, top-p = 1.0). KL regularization is applied with a coefficient of $1e{-3}$ and a clip ratio of 0.2. Each prompt is sampled with 7 candidate responses.
The maximum input sequence length is 4,096 tokens, and the maximum response length is 8,192 tokens. Learning rates are set separately for the two model variants:
- Instruct models: $1e{-6}, 7e{-7},$ and $5e{-7}$ for $7B, 14B,$ and $32B$ models, respectively.
- Reasoning models: $1e{-6}, 1e{-6},$ and $8e{-7}$ for $7B, 14B,$ and $32B$ models, respectively.
We train the $7B, 14B,$ and $32B$ models on $1, 2,$ and $4$ nodes, respectively, each equipped with 8 H100 GPUs.
## Appendix F Full Experiment Result
In this section, we provide the full experiment results and a more comprehensive coverage of existing baselines. The results of RewardBench, RM-Bench, and RMB are provided in Table 6, Table 7, Table 8, respectively.
Table 6: Results of our proposed method and baselines on the RewardBench. Bold numbers indicate the best performance, Underlined numbers indicate the second best. ${}^\char 69$ indicates potential data contamination.
| Models | Chat | Chat_Hard | Safety | Reasoning | Overall |
| --- | --- | --- | --- | --- | --- |
| ScalarRMs | | | | | |
| Eurus-RM-7b | 98.0 | 65.6 | 81.4 | 86.3 | 82.8 |
| Internlm2-7b-reward | 99.2 | 69.5 | 87.2 | 94.5 | 87.6 |
| SteerLM-RM 70B | 91.3 | 80.3 | 92.8 | 90.6 | 88.8 |
| Cohere-0514 | 96.4 | 71.3 | 92.3 | 97.7 | 89.4 |
| Internlm2-20b-reward | 98.9 | 76.5 | 89.5 | 95.8 | 90.2 |
| ArmoRM-Llama3-8B-v0.1 | 96.9 | 76.8 | 90.5 | 97.3 | 90.4 |
| Nemotron-4-340B-Reward | 95.8 | 87.1 | 91.5 | 93.6 | 92.0 |
| Skywork-Reward-Llama-3.1-8B ${}^\char 69$ | 95.8 | 87.3 | 90.8 | 96.2 | 92.5 |
| Skywork-Reward-Gemma-2-27B ${}^\char 69$ | 95.8 | 91.4 | 91.9 | 96.1 | 93.8 |
| INF-ORM-Llama3.1-70B | 96.6 | 91.0 | 93.6 | 99.1 | 95.1 |
| GenRMs | | | | | |
| Llama3.1-8B-Instruct | 85.5 | 48.5 | 75.6 | 72.1 | 70.4 |
| Prometheus-8*7B-v2 | 93.0 | 47.1 | 80.5 | 77.4 | 74.5 |
| Llama3.1-70B-Instruct | 97.2 | 70.2 | 82.8 | 86.0 | 84.0 |
| Llama3.1-405B-Instruct | 97.2 | 74.6 | 77.6 | 87.1 | 84.1 |
| Claude-3-5-sonnet-20240620 | 96.4 | 74.0 | 81.6 | 84.7 | 84.2 |
| GPT-4o-0806 | 96.1 | 76.1 | 86.6 | 88.1 | 86.7 |
| Gemini-1.5-pro | 92.3 | 80.6 | 87.9 | 92.0 | 88.2 |
| SFR-LLaMa-3.1-70B-Judge-r | 96.9 | 84.8 | 91.6 | 97.6 | 92.7 |
| Skywork-Critic-Llama-3.1-70B ${}^\char 69$ | 96.6 | 87.9 | 93.1 | 95.5 | 93.3 |
| ReasRMs | | | | | |
| JudgeLRM | 92.9 | 56.4 | 78.2 | 73.6 | 75.2 |
| SynRM | 38.0 | 82.5 | 74.1 | 87.1 | 70.4 |
| RM-R1-DeepSeek-Distilled-Qwen -7B | 88.9 | 66.2 | 78.4 | 87.0 | 80.1 |
| CLoud | 97.0 | 58.0 | 84.0 | 92.0 | 82.8 |
| DeepSeek-GRM-16B | 90.8 | 74.3 | 84.7 | 81.8 | 82.9 |
| DeepSeek-GRM-27B-RFT | 94.7 | 77.2 | 87.0 | 79.2 | 84.5 |
| RM-R1-Qwen-Instruct -7B | 94.1 | 74.6 | 85.2 | 86.7 | 85.2 |
| DeepSeek-GRM-27B | 94.1 | 78.3 | 88.0 | 83.8 | 86.0 |
| DeepSeek-PairRM-27B | 95.5 | 86.8 | 52.3 | 92.0 | 87.1 |
| RM-R1-Qwen-Instruct -14B | 93.6 | 80.5 | 86.9 | 92.0 | 88.2 |
| RM-R1-DeepSeek-Distilled-Qwen -14B | 91.3 | 79.4 | 89.3 | 95.5 | 88.9 |
| Self-taught-evaluator-llama3.1-70B | 96.9 | 85.1 | 89.6 | 88.4 | 90.0 |
| RM-R1-DeepSeek-Distilled-Qwen -32B | 95.3 | 80.3 | 91.1 | 96.8 | 90.9 |
| RM-R1-Qwen-Instruct -32B | 95.3 | 83.1 | 91.9 | 95.2 | 91.4 |
Table 7: The full results of tested reward models on RM-Bench. Chat, Math, Code, Safety show the modelâs Average Accuracy on each domain. Easy, Normal, Hard show the modelâs Accuracy on each difficulty level across all domains. Bold numbers indicate the best performance, Underlined numbers indicate the second best.
| Models | Chat | Math | Code | Safety | Easy | Normal | Hard | Avg |
| --- | --- | --- | --- | --- | --- | --- | --- | --- |
| ScalarRMs | | | | | | | | |
| steerlm-70b | 56.4 | 53.0 | 49.3 | 51.2 | 48.3 | 54.9 | 54.3 | 52.5 |
| tulu-v2.5-70b-preference-mix-rm | 58.2 | 51.4 | 55.5 | 87.1 | 72.8 | 65.6 | 50.7 | 63.0 |
| Mistral-7B-instruct-Unified-Feedback | 56.5 | 58.0 | 51.7 | 86.8 | 87.1 | 67.3 | 35.3 | 63.2 |
| RM-Mistral-7B | 57.4 | 57.0 | 52.7 | 87.2 | 88.6 | 67.1 | 34.9 | 63.5 |
| Eurus-RM-7b | 59.9 | 60.2 | 56.9 | 86.5 | 87.2 | 70.2 | 40.2 | 65.9 |
| internlm2-7b-reward | 61.7 | 71.4 | 49.7 | 85.5 | 85.4 | 70.7 | 45.1 | 67.1 |
| Skywork-Reward-Gemma-2-27B | 69.5 | 54.7 | 53.2 | 91.9 | 78.0 | 69.2 | 54.9 | 67.3 |
| ArmoRM-Llama3-8B-v0.1 | 67.8 | 57.5 | 53.1 | 92.4 | 82.2 | 71.0 | 49.8 | 67.7 |
| GRM-llama3-8B-sftreg | 62.7 | 62.5 | 57.8 | 90.0 | 83.5 | 72.7 | 48.6 | 68.2 |
| internlm2-20b-reward | 63.1 | 66.8 | 56.7 | 86.5 | 82.6 | 71.6 | 50.7 | 68.3 |
| Llama-3-OffsetBias-RM-8B | 71.3 | 61.9 | 53.2 | 89.6 | 84.6 | 72.2 | 50.2 | 69.0 |
| Nemotron-340B-Reward | 71.2 | 59.8 | 59.4 | 87.5 | 81.0 | 71.4 | 56.1 | 69.5 |
| URM-LLaMa-3.1-8B | 71.2 | 61.8 | 54.1 | 93.1 | 84.0 | 73.2 | 53.0 | 70.0 |
| Skywork-Reward-Llama-3.1-8B | 69.5 | 60.6 | 54.5 | 95.7 | 89.0 | 74.7 | 46.6 | 70.1 |
| INF-ORM-Llama3.1-70B | 66.3 | 65.6 | 56.8 | 94.8 | 91.8 | 76.1 | 44.8 | 70.9 |
| GenRMs | | | | | | | | |
| tulu-v2.5-dpo-13b-chatbot-arena-2023 | 64.9 | 52.3 | 50.5 | 62.3 | 82.8 | 60.2 | 29.5 | 57.5 |
| tulu-v2.5-dpo-13b-nectar-60k | 56.3 | 52.4 | 52.6 | 73.8 | 86.7 | 64.3 | 25.4 | 58.8 |
| stablelm-2-12b-chat | 67.2 | 54.9 | 51.6 | 65.2 | 69.1 | 63.5 | 46.6 | 59.7 |
| tulu-v2.5-dpo-13b-stackexchange-60k | 66.4 | 49.9 | 54.2 | 69.0 | 79.5 | 63.0 | 37.2 | 59.9 |
| Nous-Hermes-2-Mistral-7B-DPO | 58.8 | 55.6 | 51.3 | 73.9 | 69.5 | 61.1 | 49.1 | 59.9 |
| Claude-3-5-sonnet-20240620 | 62.5 | 62.6 | 54.4 | 64.4 | 73.8 | 63.4 | 45.9 | 61.0 |
| tulu-v2.5-dpo-13b-hh-rlhf-60k | 68.4 | 51.1 | 52.3 | 76.5 | 53.6 | 63.0 | 69.6 | 62.1 |
| tulu-2-dpo-13b | 66.4 | 51.4 | 51.8 | 85.4 | 86.9 | 66.7 | 37.7 | 63.8 |
| SOLAR-10.7B-Instruct-v1.0 | 78.6 | 52.3 | 49.6 | 78.9 | 57.5 | 67.6 | 69.4 | 64.8 |
| Llama3.1-70B-Instruct | 64.3 | 67.3 | 47.5 | 83.0 | 74.7 | 67.8 | 54.1 | 65.5 |
| Skywork-Critic-Llama-3.1-70B | 71.4 | 64.6 | 56.8 | 94.8 | 85.6 | 73.7 | 56.5 | 71.9 |
| GPT-4o-0806 | 67.2 | 67.5 | 63.6 | 91.7 | 83.4 | 75.6 | 58.7 | 72.5 |
| Gemini-1.5-pro | 71.6 | 73.9 | 63.7 | 91.3 | 83.1 | 77.6 | 64.7 | 75.2 |
| ReasRMs | | | | | | | | |
| JudgeLRM | 59.9 | 59.9 | 51.9 | 87.3 | 73.2 | 766.2 | 54.8 | 64.7 |
| RM-R1-Qwen-Instruct -7B | 66.6 | 67.0 | 54.6 | 92.6 | 79.2 | 71.7 | 59.7 | 70.2 |
| Self-taught-evaluator-llama3.1-70B | 73.4 | 65.7 | 56.3 | 90.4 | 80.2 | 74.5 | 59.7 | 71.5 |
| RM-R1-DeepSeek-Distilled-Qwen -7B | 64.0 | 83.9 | 56.2 | 85.3 | 75.9 | 73.1 | 68.1 | 72.4 |
| RM-R1-Qwen-Instruct -14B | 75.6 | 75.4 | 60.6 | 93.6 | 82.6 | 77.5 | 68.8 | 76.1 |
| RM-R1-Qwen-Instruct -32B | 75.3 | 80.2 | 66.8 | 93.9 | 86.3 | 80.5 | 70.4 | 79.1 |
| RM-R1-DeepSeek-Distilled-Qwen -14B | 71.8 | 90.5 | 69.5 | 94.1 | 86.2 | 83.6 | 74.4 | 81.5 |
| RM-R1-DeepSeek-Distilled-Qwen -32B | 74.2 | 91.8 | 74.1 | 95.4 | 89.5 | 85.4 | 76.7 | 83.9 |
Table 8: The leaderboard of RMB, ranked by the average score of all subsets. Bold numbers indicate the best performance, Underlined numbers indicate the second best.
| | Helpfulness | Harmlessness | | | |
| --- | --- | --- | --- | --- | --- |
| Models | BoN | Pairwise | BoN | Pairwise | Overall |
| ScalarRMs | | | | | |
| Tulu-v2.5-13b-preference-mix-rm | 0.355 | 0.562 | 0.351 | 0.545 | 0.453 |
| SteerLM-RM 70B | 0.502 | 0.574 | 0.578 | 0.673 | 0.582 |
| Skywork-Reward-Gemma-2-27B | 0.472 | 0.653 | 0.561 | 0.721 | 0.602 |
| Internlm2-20b-reward | 0.585 | 0.763 | 0.499 | 0.670 | 0.629 |
| ArmoRM-Llama3-8B-v0.1 | 0.636 | 0.787 | 0.497 | 0.663 | 0.646 |
| Internlm2-7b-reward | 0.626 | 0.782 | 0.563 | 0.712 | 0.671 |
| Eurus-RM-7b | 0.679 | 0.818 | 0.543 | 0.693 | 0.683 |
| Skywork-Reward-Llama-3.1-8B | 0.627 | 0.781 | 0.603 | 0.759 | 0.693 |
| INF-ORM-Llama3.1-70B | 0.650 | 0.798 | 0.607 | 0.767 | 0.705 |
| Starling-RM-34B | 0.604 | 0.774 | 0.674 | 0.795 | 0.712 |
| GenRMs | | | | | |
| Llama2-70b-chat | 0.289 | 0.613 | 0.249 | 0.602 | 0.438 |
| Llama3.1-8B-Instruct | 0.365 | 0.675 | 0.267 | 0.653 | 0.490 |
| Gemini-1.5-pro | 0.536 | 0.763 | 0.299 | 0.661 | 0.565 |
| Mixtral-8x7B-Instruct-v0.1 | 0.480 | 0.706 | 0.491 | 0.671 | 0.587 |
| skywork-critic-llama3.1-8B | 0.600 | 0.725 | 0.578 | 0.578 | 0.620 |
| skywork-critic-llama3.1-70B | 0.640 | 0.753 | 0.614 | 0.614 | 0.655 |
| Llama3.1-70B-Instruct | 0.648 | 0.811 | 0.558 | 0.739 | 0.689 |
| Mistral-Large-2407 | 0.678 | 0.817 | 0.583 | 0.725 | 0.701 |
| Claude-3-5-sonnet | 0.705 | 0.838 | 0.518 | 0.764 | 0.706 |
| Qwen2-72B-Instruct | 0.645 | 0.810 | 0.649 | 0.789 | 0.723 |
| GPT-4o-2024-05-13 | 0.639 | 0.815 | 0.682 | 0.814 | 0.738 |
| ReasRMs | | | | | |
| JudgeLRM | 0.363 | 0.699 | 0.363 | 0.674 | 0.531 |
| RM-R1-DeepSeek-Distilled-Qwen -7B | 0.451 | 0.658 | 0.429 | 0.664 | 0.551 |
| RM-R1-Qwen-Instruct -7B | 0.543 | 0.740 | 0.608 | 0.765 | 0.664 |
| Self-taught-evaluator-llama3.1-70B | 0.616 | 0.786 | 0.546 | 0.733 | 0.670 |
| Deepseek-GRM-27B-RFT | 0.592 | 0.801 | 0.548 | 0.765 | 0.670 |
| RM-R1-DeepSeek-Distilled-Qwen -14B | 0.593 | 0.765 | 0.613 | 0.769 | 0.685 |
| Deepseek-GRM-27B | 0.623 | 0.805 | 0.570 | 0.761 | 0.690 |
| RM-R1-Qwen-Instruct -14B | 0.594 | 0.776 | 0.620 | 0.778 | 0.692 |
| RM-R1-DeepSeek-Distilled-Qwen -32B | 0.620 | 0.782 | 0.618 | 0.771 | 0.698 |
| RM-R1-Qwen-Instruct -32B | 0.636 | 0.791 | 0.682 | 0.809 | 0.730 |
Table 9: Complete questions and chatbot responses.
Question: symptoms of sickle-cell disease Chatbot A: 1. Painful red or yellow skin lesions 2. Frequent infections 3. Fatigue and weakness 4. Anemia 5. Slower growth and development 6. Jaundice 7. Swelling of hands and feet 8. Vision loss 9. Tiredness and weakness after physical activity 10. Breathlessness 11. Fever, chills, and night sweats. Chatbot B Sickle cell disease is a genetic disorder that affects the shape of red blood cells. It is most common in African Americans, but can also affect people from other ethnic groups. The symptoms of sickle cell disease can vary widely from person to person, but some common symptoms include: 1. Fatigue: People with sickle cell disease often experience fatigue and weakness, even when they have not been physically active. 2. Pain: Sickle cell disease can cause pain in the body, particularly in the arms, legs, back, and abdomen. Pain can be mild or severe and can last for days or weeks. 3. Fever: People with sickle cell disease may be prone to infections and may experience frequent fevers. 4. Anemia: Sickle cell disease can cause anemia, a condition in which the body does not have enough red blood cells to carry oxygen throughout the body. 5. Delayed growth and development: Children with sickle cell disease may be smaller than their peers due to delayed growth and development. 6. Infections: People with sickle cell disease are at increased risk for infections, particularly in the bloodstream and urinary tract. 7. Pigmentation: Sickle cell disease can cause dark areas of pigmentation in the body, particularly on the hands and feet. Itâs important to note that not everyone with sickle cell disease will experience all of these symptoms, and some people may not experience any symptoms at all. Therefore, itâs important to consult with a healthcare provider if you suspect you or someone you know may have sickle cell disease.
Table 10: Cold-Start RL Judges. The rubrics and evaluation mainly center around superficial shallow features (e.g., number of points listed) without actually taking their accuracy into account.
Cold-start RL evaluation <type>Chat</type> <rubric> The evaluation criteria for the responses are as follows: - <item>Relevance to the question</item> (40%): The response should directly address the symptoms of sickle cell disease. - <item>Comprehensiveness</item> (30%): The response should include a wide range of symptoms commonly associated with sickle cell disease. - <item>Clarity and organization</item> (30%): The response should be easy to read and understand, with symptoms listed in a clear and organized manner. <justify> The question specifically asks for the symptoms of sickle cell disease, so the responses should be directly relevant to this topic. Comprehensiveness is important because sickle cell disease can have a variety of symptoms, and a comprehensive list will provide a more complete picture of the condition. Clarity and organization are important because the symptoms can be complex, and a well-organized response will make it easier for the client to understand the information. </justify> </rubric> <eval> <quote_A> 1. Painful red or yellow skin lesions 2. Frequent infections 3. Fatigue and weakness 4. Anemia 5. Slower growth and development 6. Jaundice 7. Swelling of hands and feet 8. Vision loss 9. Tiredness and weakness after physical activity 10. Breathlessness 11. Fever, chills, and night sweats. </quote_A> <summary_A> Chatbot A provided a list of symptoms that are directly relevant to sickle cell disease. The response is comprehensive and includes a wide range of symptoms commonly associated with the condition. The symptoms are listed in a clear and organized manner, making it easy for the client to understand the information. </summary_A> <quote_B> Sickle cell disease is a genetic disorder that affects the shape of red blood cells. It is most common in African Americans, but can also affect people from other ethnic groups. The symptoms of sickle cell disease can vary widely from person to person, but some common symptoms include: 1. Fatigue: People with sickle cell disease often experience fatigue and weakness, even when they have not been physically active. 2. Pain: Sickle cell disease can cause pain in the body, particularly in the arms, legs, back, and abdomen. Pain can be mild or severe and can last for days or weeks. 3. Fever: People with sickle cell disease may be prone to infections and may experience frequent fevers. 4. Anemia: Sickle cell disease can cause anemia, a condition in which the body does not have enough red blood cells to carry oxygen throughout the body. 5. Delayed growth and development: Children with sickle cell disease may be smaller than their peers due to delayed growth and development. 6. Infections: People with sickle cell disease are at increased risk for infections, particularly in the bloodstream and urinary tract. 7. Pigmentation: Sickle cell disease can cause dark areas of pigmentation in the body, particularly on the hands and feet. </quote_B> <summary_B> Chatbot B provided a detailed explanation of sickle cell disease and a list of symptoms. The response is relevant to the question and includes a wide range of symptoms commonly associated with the condition. However, the response is less organized than Chatbot Aâs response, and the explanation of sickle cell disease may be unnecessary for the clientâs question. </summary_B> </eval> <answer>[[A]]</answer>
Table 11: RM-R1 RL Judges. The rubrics and evaluation are well-designed and reasonable, assessing the effectiveness of the model response.
Warm-start RL (RM-R1) evaluation <type>Chat</type> <rubric> 1. Accuracy of information (40 %): How correctly does the response describe the symptoms of sickle-cell disease? 2. Comprehensiveness (30 %): How thoroughly does the response cover the range of symptoms associated with sickle-cell disease? 3. Clarity and organization (20 %): How well-structured and easy to understand is the information presented? 4. Helpfulness for the user (10 %): How useful is the information for someone seeking to understand sickle-cell disease symptoms? <justify> These criteria focus on evaluating how well each response addresses the userâs question about sickle-cell disease symptoms. Accuracy is weighted highest because providing correct information about medical conditions is crucial. Comprehensiveness is important, as sickle-cell disease has multiple symptoms that vary in severity and presentation. Clarity and organization affect how easily the user can understand the information. Helpfulness considers how practical and actionable the information is for the user. </justify> </rubric> <eval> Comparing the two responses: Accuracy of information: <quote_A>1. Painful red or yellow skin lesions 2. Frequent infections 3. Fatigue and weakness 4. Anemia 5. Slower growth and development 6. Jaundice 7. Swelling of hands and feet 8. Vision loss 9. Tiredness and weakness after physical activity 10. Breathlessness 11. Fever, chills, and night sweats.</quote_A> <quote_B>1. Fatigue: People with sickle-cell disease often experience fatigue and weakness, even when they have not been physically active. 2. Pain: Sickle-cell disease can cause pain in the body, particularly in the arms, legs, back, and abdomen. Pain can be mild or severe and can last for days or weeks. 3. Fever: People with sickle-cell disease may be prone to infections and may experience frequent fevers. 4. Anemia: Sickle-cell disease can cause anemia, a condition in which the body does not have enough red blood cells to carry oxygen throughout the body. 5. Delayed growth and development: Children with sickle-cell disease may be smaller than their peers due to delayed growth and development. 6. Infections: People with sickle-cell disease are at increased risk for infections, particularly in the bloodstream and urinary tract. 7. Pigmentation: Sickle-cell disease can cause dark areas of pigmentation in the body, particularly on the hands and feet.</quote_B> Chatbot A provides a list of symptoms without explanation, but some listed items (e.g. âpainful red or yellow skin lesions,â âvision lossâ) are not typical of sickle-cell disease. Chatbot B offers more accurate information, explaining each symptomâs nature and severity. Comprehensiveness: Chatbot A lists 11 symptoms, several of which are inaccurate or uncommon. Chatbot B lists 7 symptoms that are more representative and supplies contextual detail about variability and medical advice. Clarity and organization: Chatbot Aâs plain numbered list lacks context, making severity unclear. Chatbot B structures the list with explanations, improving readability and understanding. Helpfulness: Chatbot A is less helpful due to inaccuracies and missing context. Chatbot B is more useful, providing accurate descriptions, context, and guidance. </eval> <answer>[[B]]</answer>
## Appendix G Supplementary Information for Section 4
### G.1 Ablation Settings
Cold Start RL. This approach generally involves pure RL, with rule-based rewards centered on answer correctness and format compliance. Such strategies have achieved notable success in advanced mathematical problem solving [28].
In this setting, we replicate this conventional training setup. Specifically, we use a combination of a format reward and an answer reward:
$$
R_format=\begin{cases}1& if format matches,\\
0& otherwise,\end{cases} and R_answer
=\begin{cases}1& if answer matches,\\
0& otherwise.\end{cases} \tag{11}
$$
The total reward is the sum $R=R_answer+R_format$ . We use the prompt template shown in Figure 7, a version without any guidance on structured reasoning.
Chain -of -Rubrics (CoR) Roll-out for Instruct Models (no categorization of task types)
Please act as an impartial judge and evaluate the quality of the responses provided by two AI Chatbots to the Clientâs question displayed below. Instructions 1. Begin your evaluation by generating the rubric criteria tailored to the Clientâs question and context. Enclose the rubric in <rubric> ⌠</rubric> tags. 2. Assign weights to each rubric item based on their relative importance. 3. Within <rubric>, include a <justify> ⌠</justify> section explaining the rationale behind the chosen criteria and weights. 4. Compare both Chatbot responses using the rubric. 5. Include your evaluation in <eval> ⌠</eval> tags. Support your analysis using: - <quote_A> ⌠</quote_A> for direct quotes from Chatbot A - <summary_A> ⌠</summary_A> for paraphrased summaries of Chatbot A - <quote_B> ⌠</quote_B> for direct quotes from Chatbot B - <summary_B> ⌠</summary_B> for paraphrased summaries of Chatbot B 6. Conclude with your final judgment using: <answer>[[A]]</answer> or <answer>[[B]]</answer> Important Notes: - Be objective and base your evaluation strictly on the content of the responses. - Do not let the response order, length, or Chatbot names bias your judgment.
Figure 6: The system prompt of the ablation study on cold start RL without categorization of task types.
Chain -of -Rubrics (CoR) Roll-out for Instruct Models (no rubrics)
Please act as an impartial judge and evaluate the quality of the responses provided by two AI Chatbots to the Clientâs question displayed below. You should choose the chatbot that follows the clientâs instructions and answers the clientâs question better. Do not allow the length of the responses to influence your evaluation. Do not favor certain names of the chatbots. Be as objective as possible. First, compare the chatbot responses and provide your evaluations. Then, conclude with your verdict using exactly this format: <answer>[[A]]</answer> if Chatbot A is better, <answer>[[B]</answer> if Chatbot B is better.
Figure 7: The system prompt of the ablation study on cold start RL without any rubrics.
Cold Start RL + Rubrics. To examine the influence of structured reasoning in final model performance, compared with the last setting, we use the prompt template shown in Figure 6. Compared with the last setting, the model is prompted to generate rubrics and evaluate accordingly. However, compared with the final system prompt of RM-R1 Figure 3, all input prompts are treated uniformlyâthat is, chat and reasoning tasks are not distinguished.
Cold Start RL + Rubrics + Query Categorization (QC). This setting largely follows the previous one, with a key modification: prompting the LM to first categorize the task into reasoning or chat, and then apply different strategies for handling those tasks. Intuitively, reinforcement learning alone can effectively explore reasoning tasks, a domain where it has already achieved considerable success. Here, we incorporate the system prompt shown in Figure 3, which explicitly distinguishes between chat and reasoning tasks.
For reasoning tasks specifically, we note that answer quality is closely tied to correctness, and that high-level rubrics may be less effective than simply evaluating whether the model can solve the problem and verify its own answer. Thus, this setting emphasizes correctness-based evaluation guided by task classification in the prompt.
Distilled + RL + Rubrics + QC (RM-R1). Building on the previous setup, we introduce an additional distillation stage from stronger teacher models as a warm start before RL training. The motivation is that with RL alone, weaker models (especially at smaller scales) often fail to explore high-quality rubrics and convincing reasoning chains for chat tasks throughout the RL training process. Distilling strong reasoning traces on a small subset of data can effectively mitigate this limitation.
<details>
<summary>x5.png Details</summary>

### Visual Description
## Line Charts: Response Length and Train Reward vs. Training Step
### Overview
The image displays two side-by-side line charts sharing a common X-axis ("Step"). The left chart tracks "Response Length" over training steps, and the right chart tracks "Train Reward" over the same steps. Both charts show data from step 0 to step 70. The overall visual suggests a correlation between the progression of training steps, the length of generated responses, and the reward signal.
### Components/Axes
**Common X-Axis (Both Charts):**
- **Label:** `Step`
- **Scale:** Linear, from 0 to 70.
- **Major Ticks:** 0, 10, 20, 30, 40, 50, 60, 70.
**Left Chart: Response Length vs Step**
- **Title:** `Response Length vs Step` (centered at top).
- **Y-Axis Label:** `Response Length`
- **Y-Axis Scale:** Linear, ranging from approximately 480 to 690.
- **Major Y-Axis Ticks:** 500, 550, 600, 650.
- **Data Series:** A single series represented by a blue line with circular markers.
- **Legend:** Implicit in the title; no separate legend box.
**Right Chart: Train Reward vs Step**
- **Title:** `Train Reward vs Step` (centered at top).
- **Y-Axis Label:** `Train Reward`
- **Y-Axis Scale:** Linear, ranging from 1.50 to 1.90.
- **Major Y-Axis Ticks:** 1.50, 1.55, 1.60, 1.65, 1.70, 1.75, 1.80, 1.85, 1.90.
- **Data Series:** A single series represented by a red line with square markers.
- **Legend:** Implicit in the title; no separate legend box.
### Detailed Analysis
**Left Chart: Response Length Trend**
- **Visual Trend:** The blue line shows a clear, noisy upward trend. It starts low, increases steadily with fluctuations, and ends at its highest point.
- **Key Data Points (Approximate):**
- Step 0: ~480
- Step 10: ~520
- Step 20: ~550
- Step 30: ~600
- Step 40: ~620
- Step 50: ~660 (local peak)
- Step 60: ~600 (significant dip)
- Step 70: ~680 (global peak)
- **Pattern:** The growth is not monotonic. Notable dips occur around steps 5, 15, 25, and most prominently around step 60, after which the length recovers sharply.
**Right Chart: Train Reward Trend**
- **Visual Trend:** The red line shows a rapid initial increase, followed by a plateau with minor fluctuations, and then a sharp, significant drop near the end.
- **Key Data Points (Approximate):**
- Step 0: ~1.52
- Step 5: ~1.73 (rapid initial rise)
- Step 15: ~1.80
- Step 25: ~1.83
- Step 35: ~1.85
- Step 45-65: Plateau, fluctuating between ~1.85 and ~1.88.
- Step ~66: Sharp drop to ~1.72.
- Step 70: Partial recovery to ~1.78.
- **Pattern:** The reward climbs quickly in the first 10-15 steps, stabilizes at a high value for about 50 steps, then experiences a sudden, substantial decline just before step 70.
### Key Observations
1. **Correlated Growth:** For the first ~60 steps, both response length and training reward show a general upward trend, suggesting that as the model trains, it produces longer responses that are deemed more rewarding.
2. **Divergence at Step ~60:** A critical event occurs around step 60. The response length dips sharply (from ~660 to ~600), while the training reward remains stable at its plateau (~1.87). This indicates the reward signal was not immediately sensitive to this specific drop in length.
3. **Late Training Instability:** The most dramatic event is the sharp drop in train reward after step 65, falling from ~1.87 to ~1.72. Interestingly, during this same period (steps 65-70), the response length *increases* sharply from its dip, reaching its maximum. This creates a strong inverse relationship at the end of the plotted training.
4. **Noise Level:** Both signals are noisy, with step-to-step fluctuations, which is typical for training metrics.
### Interpretation
The data suggests a reinforcement learning or fine-tuning process where a model is being optimized to generate longer responses that receive higher rewards. The initial phase (steps 0-15) shows successful co-optimization: length and reward increase together.
The middle phase (steps 15-60) shows a stable policy where the model maintains high reward while gradually increasing response length, albeit with volatility. The dip in length at step 60 without a reward penalty might indicate the reward model is robust to certain variations in length at that stage.
The final phase (steps 65-70) reveals a potential failure mode or a significant shift in the training dynamics. The sharp drop in reward concurrent with a surge in response length could indicate:
- **Reward Hacking:** The model may have found a way to exploit the reward function by generating very long responses that are technically high-scoring but are of lower quality or deviate from the intended behavior, causing the reward model (or human evaluators) to subsequently penalize them heavily.
- **Training Instability:** This could be a sign of divergence, where the policy moves into a region of the state space that the reward model poorly estimates, or where optimization becomes unstable.
- **Catastrophic Forgetting:** The model might be overfitting to recent batches, losing previously learned good behaviors.
In summary, the charts depict a training process that is initially successful but exhibits signs of instability and potential reward misspecification in its later stages, warranting investigation into the reward function and training hyperparameters around step 65.
</details>
(a) Cold Start RL
<details>
<summary>x6.png Details</summary>

### Visual Description
\n
## Line Charts: Response Length vs Step and Train Reward vs Step
### Overview
The image displays two side-by-line charts sharing a common x-axis ("Step") but measuring different metrics. The left chart tracks "Response Length" over 60 steps, while the right chart tracks "Train Reward" over the same 60 steps. Both charts use a grid background with dashed lines.
### Components/Axes
**Left Chart: "Response Length vs Step"**
* **Title:** "Response Length vs Step" (centered above the plot).
* **Y-axis Label:** "Response Length" (vertical text on the left).
* **Y-axis Scale:** Linear scale ranging from approximately 760 to 840, with major tick marks at 760, 780, 800, 820, and 840.
* **X-axis Label:** "Step" (centered below the plot).
* **X-axis Scale:** Linear scale from 0 to 60, with major tick marks at 0, 10, 20, 30, 40, 50, and 60.
* **Data Series:** A single blue line with circular markers at each data point.
* **Legend:** None present (single data series).
**Right Chart: "Train Reward vs Step"**
* **Title:** "Train Reward vs Step" (centered above the plot).
* **Y-axis Label:** "Train Reward" (vertical text on the left).
* **Y-axis Scale:** Linear scale ranging from approximately 0.45 to 0.70, with major tick marks at 0.45, 0.50, 0.55, 0.60, 0.65, and 0.70.
* **X-axis Label:** "Step" (centered below the plot).
* **X-axis Scale:** Linear scale from 0 to 60, with major tick marks at 0, 10, 20, 30, 40, 50, and 60.
* **Data Series:** A single red line with square markers at each data point.
* **Legend:** None present (single data series).
### Detailed Analysis
**Left Chart: Response Length Trend**
* **Trend Verification:** The blue line exhibits a distinct U-shaped or "dip and recover" trend. It starts high, descends to a minimum, and then ascends again.
* **Data Points (Approximate):**
* **Start (Step 0):** ~850
* **Initial High (Steps 0-15):** Fluctuates between ~820 and ~850.
* **Decline Phase (Steps 15-30):** A steep, consistent decline from ~825 to a minimum.
* **Minimum (Step ~30):** The lowest point is approximately 750.
* **Recovery Phase (Steps 30-60):** A steady, fluctuating increase from ~750 back to ~800.
* **End (Step 60):** ~800.
**Right Chart: Train Reward Trend**
* **Trend Verification:** The red line shows a clear, positive upward trend with moderate volatility. The overall direction is from bottom-left to top-right.
* **Data Points (Approximate):**
* **Start (Step 0):** ~0.47
* **Early Phase (Steps 0-20):** Gradual increase with fluctuations, reaching ~0.55.
* **Mid Phase (Steps 20-40):** Continued increase, crossing 0.60 around step 35.
* **Late Phase (Steps 40-60):** Stronger upward movement, with values consistently above 0.60.
* **Peak (Step ~55):** The highest point is approximately 0.71.
* **End (Step 60):** ~0.69.
### Key Observations
1. **Inverse Relationship at Mid-Training:** The most notable pattern is the inverse correlation around step 30. As "Response Length" hits its minimum (~750), "Train Reward" is in a period of steady increase (~0.58).
2. **Divergent Endpoints:** By step 60, "Response Length" (~800) has not returned to its initial high (~850), while "Train Reward" (~0.69) is significantly higher than its starting point (~0.47).
3. **Volatility:** Both metrics show step-to-step fluctuations, but the "Train Reward" chart appears to have slightly more pronounced short-term volatility relative to its scale.
4. **Synchronized Steps:** The x-axes are perfectly aligned, allowing for direct comparison of the two metrics at any given step.
### Interpretation
This pair of charts likely visualizes the training dynamics of a language model or similar AI system using a reinforcement learning or optimization technique where a "reward" signal is used.
* **Learning Progression:** The consistent rise in "Train Reward" indicates the model is successfully learning the task according to the defined reward function over the 60 steps.
* **Conciseness vs. Quality Trade-off:** The initial high response length followed by a sharp decrease suggests the model first produced verbose outputs. The training process then incentivized more concise responses (the dip). The subsequent partial recovery in length, coupled with ever-increasing reward, implies the model learned to balance conciseness with effectivenessâfinding a length that maximizes reward, which is shorter than its initial guess but longer than its most concise phase.
* **Optimization Sweet Spot:** The minimum response length at step 30 does not correspond to the maximum reward. The highest rewards are achieved later, with moderately longer responses. This suggests the optimal policy for this task involves responses of a specific, intermediate length.
* **Training Stability:** The smooth, upward trend in reward, despite fluctuations, suggests stable training without catastrophic forgetting or divergence within this 60-step window.
**In summary, the data demonstrates a model undergoing optimization where it learns to modulate its output length as a key factor in maximizing a reward signal, ultimately converging on a policy that produces moderately concise, high-reward responses.**
</details>
(b) Warm Start RL
Figure 8: RL training dynamics under different settings: (a) Cold Start RL (Eq. 11) and (b) Warm Start RL (Eq. 8). In Cold Start RL, the response length steadily increases as the model learns to reason, but training becomes unstable near the end. In Warm Start RL, the model exhibits more stable training, with effective refinement of reasoning traces throughout the process.
### G.2 Training Dynamics
We analyze the training dynamics of RM-R1 using the Qwen-2.5-14B-Instruct model by tracking both response length and reward progression throughout RL training. We consider two settings: (a) Cold Start RL, and (b) Warm Start RL following reasoning-chain distillation. We present the finding in Figure 8.
In the Cold Start RL setting, we observe that the model gradually learns to reason, as reflected by a steady increase in response length over the course of training. However, training becomes unstable near the end, with a sharp drop in the reward curve, suggesting potential issues such as overfitting.
In contrast, under Warm Start RL, the model begins with stronger initial reasoning abilities, exhibiting longer responses from the outset. Interestingly, the model first learns to produce more concise reasoning traces before gradually increasing response length again as training progresses. The reward curve rises smoothly and consistently throughout training, demonstrating more stable and efficient learning compared to the Cold Start setting.