# Training-Free Group Relative Policy Optimization
tristanli@tencent.com https://github.com/TencentCloudADP/youtu-agent/tree/training_free_GRPO
(October 9, 2025)
## Abstract
Recent advances in Large Language Model (LLM) agents have demonstrated their promising general capabilities. However, their performance in specialized real-world domains often degrades due to challenges in effectively integrating external tools and specific prompting strategies. While methods like agentic reinforcement learning have been proposed to address this, they typically rely on costly parameter updates, for example, through a process that uses Supervised Fine-Tuning (SFT) followed by a Reinforcement Learning (RL) phase with Group Relative Policy Optimization (GRPO) to alter the output distribution. However, we argue that LLMs can achieve a similar effect on the output distribution by learning experiential knowledge as a token prior, which is a far more lightweight approach that not only addresses practical data scarcity but also avoids the common issue of overfitting. To this end, we propose Training-Free Group Relative Policy Optimization (Training-Free GRPO), a cost-effective solution that enhances LLM agent performance without any parameter updates. Our method leverages the group relative semantic advantage instead of numerical ones within each group of rollouts, iteratively distilling high-quality experiential knowledge during multi-epoch learning on a minimal ground-truth data. Such knowledge serves as the learned token prior, which is seamlessly integrated during LLM API calls to guide model behavior. Experiments on mathematical reasoning and web searching tasks demonstrate that Training-Free GRPO, when applied to DeepSeek-V3.1-Terminus, significantly improves out-of-domain performance. With just a few dozen training samples, Training-Free GRPO outperforms fine-tuned small LLMs with marginal training data and cost.
<details>
<summary>figures/spotlight.png Details</summary>

### Visual Description
## Bar Charts: AIME Benchmark Performance, Training Cost, and Training Data Comparison
### Overview
The image contains three distinct bar charts arranged horizontally. The leftmost and largest chart compares model performance on AIME benchmarks for two years (AIME24 and AIME25) under four different conditions. The middle chart compares the training cost in US dollars between two models. The rightmost chart compares the number of training data samples required by the same two models. The overall visual narrative demonstrates significant performance improvements and dramatic reductions in cost and data requirements for a method labeled "Ours."
### Components/Axes
**1. AIME Benchmarks Chart (Left)**
* **Title:** AIME Benchmarks
* **Y-axis:** Label: `Mean@32 (%)`. Scale: 40 to 90, with increments of 10.
* **X-axis:** Two main groups: `AIME24` and `AIME25`. Each group contains four bars.
* **Bar Categories (X-axis labels, left to right within each year group):**
1. `W/o tool` (Gray bar)
2. `W/o tool +Ours` (Light blue bar)
3. `With tool` (Gray bar)
4. `With tool +Ours` (Light blue bar)
* **Legend/Key:** Implicit from bar color and label. Gray bars represent baseline conditions ("W/o tool" or "With tool"). Light blue bars represent the same condition with the addition of the proposed method ("+Ours").
* **Visual Elements:** Light blue upward-curving arrows connect each gray baseline bar to its corresponding light blue "+Ours" bar, indicating improvement.
**2. Training Cost Chart (Middle)**
* **Title:** Training Cost
* **Y-axis:** Label: `Cost ($)`. Scale: 0 to 10000, with increments of 2000.
* **X-axis:** Two bars labeled:
1. `RL Train (32B)` (Light orange/peach bar)
2. `Ours (671B)` (Very short, light blue bar)
* **Visual Elements:** A large, light blue downward-curving arrow points from the top of the "RL Train" bar to the "Ours" bar. A magnifying glass graphic circles the value `~18` above the "Ours" bar, emphasizing its small size.
**3. Training Data Chart (Right)**
* **Title:** Training Data
* **Y-axis:** Label: `# Samples`. Scale: 0 to 17500, with increments of 2500.
* **X-axis:** Two bars labeled:
1. `RL Train (32B)` (Light orange/peach bar)
2. `Ours (671B)` (Very short, light blue bar)
* **Visual Elements:** A large, light blue downward-curving arrow points from the top of the "RL Train" bar to the "Ours" bar. A magnifying glass graphic circles the value `100` above the "Ours" bar, emphasizing its small size.
### Detailed Analysis
**AIME Benchmarks Data Points:**
* **AIME24:**
* `W/o tool`: 68.6%
* `W/o tool +Ours`: 72.6% (Increase of ~4.0 percentage points)
* `With tool`: 80.0%
* `With tool +Ours`: 82.7% (Increase of ~2.7 percentage points)
* **AIME25:**
* `W/o tool`: 52.9%
* `W/o tool +Ours`: 54.0% (Increase of ~1.1 percentage points)
* `With tool`: 67.9%
* `With tool +Ours`: 73.3% (Increase of ~5.4 percentage points)
**Training Cost Data Points:**
* `RL Train (32B)`: ~10000 (The bar reaches the 10000 line, and the label uses a tilde `~`).
* `Ours (671B)`: ~18 (Explicitly labeled with a tilde `~` inside a magnifying glass).
**Training Data Data Points:**
* `RL Train (32B)`: 17000 (The bar aligns with the 17000 grid line).
* `Ours (671B)`: 100 (Explicitly labeled inside a magnifying glass).
### Key Observations
1. **Consistent Performance Gain:** The "+Ours" method improves the Mean@32 score in all four tested scenarios (across both years and both tool-use conditions). The gains range from +1.1 to +5.4 percentage points.
2. **Higher Baseline with Tools:** Using tools (`With tool`) yields a substantially higher baseline performance than not using tools (`W/o tool`) in both years (e.g., 80.0% vs 68.6% in AIME24).
3. **Drastic Cost Reduction:** The training cost for "Ours (671B)" is approximately **$18**, compared to approximately **$10,000** for "RL Train (32B)". This represents a reduction of over 99.8%.
4. **Massive Data Efficiency:** The "Ours (671B)" model was trained on only **100 samples**, compared to **17,000 samples** for "RL Train (32B)". This is a 170x reduction in data volume.
5. **Model Size Paradox:** The "Ours" model is significantly larger (671B parameters) than the "RL Train" model (32B parameters), yet it achieves its results with vastly less data and cost.
### Interpretation
The charts collectively argue for the extreme efficiency and effectiveness of the proposed method ("Ours"). The data suggests that this method, when applied to a very large model (671B), can:
* **Boost Performance:** Consistently improve accuracy on challenging mathematical benchmarks (AIME), both with and without external tools.
* **Achieve Unprecedented Efficiency:** Do so while requiring **four orders of magnitude less financial cost** and **over two orders of magnitude less training data** than a standard reinforcement learning training approach on a smaller (32B) model.
The visual emphasis (magnifying glasses, dramatic downward arrows) highlights the most striking claim: that scaling up model size (to 671B) in conjunction with this specific method decouples performance from the traditional, resource-intensive requirements of large-scale training. The outlier is not a data point but the method itself—it presents a scenario where a much larger model is trained with trivial resources, challenging the common assumption that larger models always require proportionally more data and cost. The charts are designed to showcase a breakthrough in training efficiency.
</details>
Figure 1: Applying Training-Free GRPO on both prompting (without tools) and ReAct [1] (with tools) achieve improved Mean@32 on AIME benchmarks [2] with DeepSeek-V3.1-Terminus [3]. It consumes significantly fewer training data and lower costs on the 671B LLM than fine-tuning a 32B model [4], serving as a cost-effective alternative to RL methods. footnotetext: Full author list in contributions.
## 1 Introduction
Large Language Models (LLMs) are emerging as powerful general-purpose agents capable of interacting with complex, real-world environments. They have shown remarkable capabilities across a wide range of tasks, including complex problem-solving [4, 5, 6], advanced web research [7, 8, 9, 10], code generation and debugging [11, 12], and proficient computer use [13, 14, 15]. Despite their impressive capabilities, LLM agents often underperform in specialized, real-world domains. These scenarios typically demand the integration of external tools (e.g., calculators, APIs, databases), along with domain-specific task definitions and prompting strategies. Deploying a general-purpose agent out-of-the-box in such settings often results in suboptimal performance due to limited familiarity with domain-specific requirements or insufficient exposure to necessary tools.
To bridge this gap, agentic training has emerged as a promising strategy to facilitate the adaptation of LLM agents to specific domains and their associated tools [4, 7, 8, 16]. Recent advancements in agentic reinforcement learning (Agentic RL) approaches have employed Group Relative Policy Optimization (GRPO) [17] and its variants [18, 19, 20] to align model behaviors in the parameter space. Although these methods effectively enhance task-specific capabilities, their reliance on tuning LLM parameters poses several practical challenges:
- Computational Cost: Even for smaller models, fine-tuning demands substantial computational resources, making it both costly and environmentally unsustainable. For larger models, the costs become prohibitive. Furthermore, fine-tuned models require dedicated deployment and are often limited to specific applications, rendering them inefficient for low-frequency use cases compared to more versatile general-purpose models.
- Poor Generalization: Models optimized via parameter tuning often suffer from unsatisfactory cross-domain generalization, limiting their applicability to narrow tasks. Consequently, multiple specialized models must be deployed to handle a comprehensive set of tasks, significantly increasing system complexity and maintenance overhead.
- Data Scarcity: Fine-tuning LLMs typically necessitates large volumes of high-quality, carefully annotated data, which are often scarce and prohibitively expensive to obtain in specialized domains. Additionally, with limited samples, models are highly susceptible to overfitting, leading to poor generalization.
- Diminishing Returns: The prohibitive training costs usually compel existing approaches to fine-tune smaller LLMs with fewer than 32 billion parameters, due to resource constraints rather than optimal design choices. While larger models would be preferred, the computational expense of fine-tuning necessitates this compromise. Paradoxically, API-based or open-source larger LLMs often deliver better cost-performance ratios through scalability and continuous model updates. However, these general-purpose models underperform in specialized domains where fine-tuning is necessary, creating a cost-performance dilemma.
Such limitations inherent in parameter tuning motivate a fundamental research question: Is applying RL in parametric space the only viable approach? Can we enhance LLM agent performance in a non-parametric way with lower data and computational costs?
We answer this question affirmatively by proposing Training-Free Group Relative Policy Optimization (Training-Free GRPO), a novel and efficient method that improves LLM agent behavior in a manner similar to vanilla GRPO, while preserving the original model parameters unchanged. Our approach is motivated by the insight that LLMs already possess the fundamental capability to adapt to new scenarios, requiring only minimal practice through limited samples to achieve strong performance. Thus, instead of adapting their output distribution through parameter tuning, in-context learning [21] that leverages a lightweight token prior can also encapsulate experiential knowledge learned from a minimal training dataset.
Training-Free GRPO retains the multi-epoch learning mechanism of vanilla GRPO. In each epoch, multiple outputs are generated to deliver a group of rollouts for every query, which helps to explore the policy space and evaluate potential strategies. While vanilla GRPO relies on gradient-based parameter updates to iteratively improve policy performance, Training-Free GRPO eliminates this requirement by employing inference-only operations using LLMs. At each optimization step, rather than calculating a numerical advantage for gradient ascent within each group of rollouts, our method leverages LLMs to introspect on each group and distill a semantic advantage. Such advantage refines external experiential knowledge and guide policy outputs based on evolving contextual priors, thereby achieving policy optimization effects without modifying any model parameters.
By evaluating challenging mathematical reasoning and interactive web searching tasks, we demonstrate that our method significantly enhances the performance of frozen policy models such as DeepSeek-V3.1-Terminus [3] with only dozens of training samples. It surpasses fine-tuned 32B models in performance while requiring only a fraction of the computational resources, offering a simple and much more efficient alternative to traditional fine-tuning techniques.
Our principal contributions are threefold:
- A New Training-Free RL Paradigm: We introduce Training-Free GRPO, which shifts policy optimization from the parameter space to the context space by leveraging evolving experiential knowledge as token priors without gradient updates.
- Semantic Group Advantage: We replace numerical group advantage in vanilla GRPO with semantic group advantage, enabling LLMs to introspect their own rollouts and continuously updating experiential knowledge at multiple optimization steps.
- Data and Computational Efficiency: Experiments confirm that Training-Free GRPO effectively enhances the performance of a frozen policy with minimal training samples, offering a practical and cost-effective alternative across different domains.
- Superior Generalization: By leaving model parameters frozen and plugging in different token priors, our approach fully preserves the generalization power, eliminating the cost and complexity of deploying multiple fine-tuned specialists.
## 2 Training-Free GRPO
<details>
<summary>figures/training-free_GRPO.png Details</summary>

### Visual Description
## Diagram: Comparison of Vanilla GRPO and Training-free GRPO Architectures
### Overview
The image is a technical diagram comparing two reinforcement learning architectures for Large Language Models (LLMs): "(a) Vanilla GRPO" and "(b) Training-free GRPO". It illustrates the flow of data, model components, and update mechanisms for each method. The diagram uses icons (brain, robot, database, snowflake, fire) and labeled boxes connected by arrows to depict the process.
### Components/Axes
The diagram is divided into two horizontal panels, separated by a dashed line.
**Panel (a): Vanilla GRPO**
* **Title:** "(a) Vanilla GRPO" (top-left corner).
* **Input:** A box labeled `q` (query).
* **Core Models:**
* `policy model`: A box with a brain icon and a fire emoji (indicating it is updated/trained).
* `reference model`: A box with a brain icon and a snowflake emoji (indicating it is frozen).
* `reward model`: A box with a robot icon and a snowflake emoji.
* **Data Flow:**
* Outputs from the policy model are labeled `O₁`, `O₂`, ..., `O_G` (a group of outputs).
* Outputs from the reward model are labeled `r₁`, `r₂`, ..., `r_G` (a group of rewards).
* A box labeled `Group Computation` processes the rewards.
* Final outputs are labeled `A₁`, `A₂`, ..., `A_G` (likely advantages or actions).
* **Update Loop:** An orange arrow labeled `Update` flows from the `Group Computation` output back to the `policy model`.
**Panel (b): Training-free GRPO**
* **Title:** "(b) Training-free GRPO" (top-left corner of the lower panel).
* **Input:** A box labeled `q` (query) and a box labeled `experiences` with a database icon and a fire emoji.
* **Core Models:**
* `policy model`: A box with a brain icon and a snowflake emoji (frozen).
* `reward model`: A box with a robot icon and a snowflake emoji.
* `controller`: A box with a brain icon and a snowflake emoji, containing four buttons: `+ ADD`, `× DELETE`, `- KEEP`, `✎ MODIFY`.
* `LLM` (two instances): Boxes with brain icons and snowflake emojis, labeled `summarize` and `extract experience` respectively.
* **Data Flow:**
* Outputs from the policy model are labeled `O₁`, `O₂`, ..., `O_G`.
* Outputs from the reward model are labeled with symbols and `r` values: `✗ r₁`, `✓ r₂`, ..., `✗ r_G` (indicating rejected and accepted rewards).
* A large box labeled `Group Computation` contains the two `LLM` instances.
* Intermediate summaries are labeled `s₁`, `s₂`, ..., `s_G`.
* Final outputs are labeled `A_text₁`, `A_text₂`, ..., `A_text_G` (text-based advantages or actions).
* **Update Loop:** An orange arrow labeled `Update` flows from the `Group Computation` output back to the `experiences` box. Another orange arrow connects the `controller` to the `experiences` box.
### Detailed Analysis
**Vanilla GRPO Flow:**
1. A query `q` is fed into a trainable `policy model`.
2. The policy model generates a group of outputs `O₁` to `O_G`.
3. These outputs are evaluated by two frozen models: a `reference model` and a `reward model`.
4. The reward model produces scalar rewards `r₁` to `r_G`.
5. These rewards undergo `Group Computation` to produce advantages `A₁` to `A_G`.
6. The computed advantages are used to `Update` the policy model, completing the training loop.
**Training-free GRPO Flow:**
1. A query `q` and a set of `experiences` (from a database) are fed into a **frozen** `policy model`.
2. The policy model generates outputs `O₁` to `O_G`.
3. These outputs are evaluated by a frozen `reward model`, which produces rewards marked as accepted (`✓`) or rejected (`✗`).
4. The rewards enter a `Group Computation` module.
5. Inside this module, a frozen `LLM` first `summarize`s the group, producing summaries `s₁` to `s_G`.
6. A second frozen `LLM` then performs `extract experience` on these summaries to produce final text-based outputs `A_text₁` to `A_text_G`.
7. These text outputs are used to `Update` the `experiences` database.
8. A separate frozen `controller` (with ADD, DELETE, KEEP, MODIFY functions) also interacts with the `experiences` database.
### Key Observations
1. **Model State:** In Vanilla GRPO, only the policy model is trained (fire emoji). In Training-free GRPO, all models (policy, reward, controller, LLMs) are frozen (snowflake emoji); learning happens by updating the external `experiences` database.
2. **Output Nature:** Vanilla GRPO produces numerical advantages (`A_G`). Training-free GRPO produces textual advantages or actions (`A_text_G`).
3. **Reward Processing:** Vanilla GRPO uses raw rewards directly. Training-free GRPO filters rewards (✓/✗) and processes them through summarization and extraction steps using LLMs.
4. **Update Target:** The update in (a) targets the model's parameters. The update in (b) targets an external memory (`experiences`).
5. **Controller:** The `controller` is a unique component in (b), suggesting a mechanism for curating the experience database through discrete operations (add, delete, keep, modify).
### Interpretation
This diagram contrasts two paradigms for improving LLM-based agents via reinforcement learning.
* **Vanilla GRPO** represents a traditional **parametric learning** approach. The agent (policy model) improves by directly adjusting its internal weights based on computed advantages from a reward model. This is analogous to standard policy gradient methods.
* **Training-free GRPO** represents a **non-parametric or retrieval-augmented learning** approach. The agent's core model is frozen. Instead of changing weights, it improves by maintaining and updating an external database of "experiences." The system uses frozen LLMs to reason about groups of outcomes, summarize them, and extract textual lessons, which are then stored. The `controller` acts as a manager for this experience bank.
The key implication is that Training-free GRPO aims to achieve adaptation and learning without the computational cost and potential instability of backpropagation-based training. It shifts the learning burden from model parameters to a structured memory system, potentially offering more interpretable and controllable updates (via the controller's discrete actions). The use of LLMs for summarization and extraction suggests the system leverages the model's existing reasoning capabilities to create high-level, textual knowledge from raw experiences.
</details>
Figure 2: Comparison of vanilla GRPO and Training-Free GRPO.
<details>
<summary>figures/case.png Details</summary>

### Visual Description
## Diagram: AI Agent Problem-Solving Process for a Geometry Task
### Overview
The image is a flowchart illustrating the process of an AI agent solving a geometry problem. It compares two different solution trajectories (one incorrect, one correct) and shows how they are summarized and evaluated to derive a learning advantage. The diagram is divided into four main vertical sections: **Question**, **Rollout × G**, **Summarization**, and **Advantage**.
### Components/Axes
The diagram is structured as a process flow from left to right.
1. **Question Section (Far Left):**
* **Content:** A text box containing the geometry problem statement.
* **Text:** "Triangle ABC has vertices A(0, 8), B(2, 0), C(8, 0). A line through B cuts the area of triangle ABC in half; find the sum of the slope and y-intercept of this line."
* **Visual:** A small diagram of triangle ABC with vertices labeled A, B, C. A point "D?" is shown on the extension of side AC, indicating the unknown intersection point. A pixel-art user icon is adjacent to the text box.
2. **Rollout × G Section (Center-Left):**
* This section shows multiple solution attempts ("trajectories"). Two are detailed, with a third ("trajectory G") implied.
* **Trajectory 1 (Top, marked with a red X):** An incorrect solution path.
* **Steps:** "Let me find the intersection point D on AC." -> "executing code ..." -> "When m = -2.0, the area of triangle ABD is half the total area." -> "verifying the area of triangle ABD ..." -> "I found the solution! ..."
* **Visual:** A small diagram shows point D placed on the *extension* of line segment AC, outside the triangle, marked with a red X.
* **Trajectory 2 (Bottom, marked with a green checkmark):** A correct solution path.
* **Steps:** "For m = -2, intersection point D: (-4, 12). Let me verify this by checking if the line indeed divides the triangle into two equal areas." -> "executing code ..." -> "... The point D(-4,12) is not actually on the line segment AC, it's an extension of it. ..." -> "executing code ..." -> "The line y = 2x - 4 through point B indeed divides triangle ABC into two equal areas of 12 each."
* **Visual:** A small diagram shows the correct line through B intersecting side AC at a valid point D, marked with a green checkmark.
3. **Summarization Section (Center-Right):**
* This section provides a step-by-step breakdown of the reasoning in each trajectory.
* **Summarization 1 (Top, corresponding to Trajectory 1):**
* **Steps:** "Step3: The agent set up equations to find the intersection point D." -> "Step4: The agent tested various slopes numerically and found that when m = -2, the area of triangle ABD equals exactly half the total" -> "Step5: The agent verified the solution divides the area in half"
* **Summarization 2 (Bottom, corresponding to Trajectory 2):**
* **Steps:** "Step3: The agent tests a specific slope value m = -2 to find the intersection point D." -> "Step4: The agent attempts to verify the solution by calculating both areas (ABD and BCD)." -> "Step5: The agent calculates the area of triangle BCD and discovers it's 36. It realizes that point on the extension of segment AC, making the geometry invalid." -> "Step6: The agent properly sets up the area equation constraint and find the intersection point."
4. **Advantage Section (Far Right):**
* This section, labeled "Group Computation," analyzes the difference between the attempts to extract a learning point.
* **Sub-box: "advantage computation"**
* **Text (Analysis of Attempt 1):** "In attempt 1, the agent failed to consider the physical constraints of the triangle, the found intersection lies outside segment AC" (The phrase "the found intersection lies outside segment AC" is highlighted in red).
* **Text (Analysis of Attempt 2):** "In attempt 2, the agent recognized the error in step 3 and completely changed approach, properly setting up both the area and geometric constraints simultaneously." (The phrase "properly setting up both the area and geometric constraints simultaneously" is highlighted in green).
* **Conclusion (Bottom, with a lightbulb icon):** "When solving geometry problems involving intersections with bounded regions, always validate that mathematical solutions satisfy geometric constraints." (The phrase "always validate that mathematical solutions satisfy geometric constraints" is underlined in orange).
### Detailed Analysis
* **Problem Statement:** The core task is to find the equation of a line through vertex B(2,0) of triangle ABC that bisects its area. The vertices are A(0,8), B(2,0), C(8,0). The required output is the sum of the slope and y-intercept of this line.
* **Incorrect Trajectory (1):** The agent finds a mathematical solution (slope m = -2) that yields the correct area ratio but fails to check if the resulting intersection point D lies on the *segment* AC. The diagram shows D placed on the extension of AC, which is geometrically invalid for the problem's constraint ("a line through B cuts the area of triangle ABC").
* **Correct Trajectory (2):** The agent initially makes the same error (finding D at (-4,12), which is on the extension of AC). However, it then verifies the solution by checking the areas of both resulting triangles (ABD and BCD). Upon calculating the area of BCD as 36 (which is incorrect for a valid bisection), it recognizes the geometric constraint violation. It then correctly sets up the area equation with the constraint that D must lie on segment AC, leading to the valid solution: the line y = 2x - 4.
* **Summarization:** The summarization boxes distill the key reasoning steps. Summarization 1 shows a linear, uncorrected process. Summarization 2 shows a process with error detection, validation, and correction.
* **Advantage/Learning:** The final analysis explicitly contrasts the two approaches. The key differentiator is the validation of geometric constraints. The incorrect attempt focused only on the algebraic area condition. The correct attempt incorporated a check that the solution was physically possible within the triangle's boundaries.
### Key Observations
1. **Error Pattern:** The primary error in the first trajectory is a classic "algebraic solution without geometric validation." The agent solved for the area condition but ignored the domain constraint (point D must be between A and C).
2. **Correction Mechanism:** The successful trajectory includes a verification step (calculating the area of triangle BCD) that acts as a sanity check, revealing the inconsistency and prompting a reevaluation of the approach.
3. **Visual Coding:** The diagram uses clear visual cues: a red X for failure, a green checkmark for success, red text to highlight the specific error, and green text to highlight the corrective action. The final lesson is underlined in orange for emphasis.
4. **Process Flow:** The arrows clearly show the flow from problem, to multiple solution attempts, to summarization of each, and finally to a comparative analysis that yields a generalizable insight.
### Interpretation
This diagram is a pedagogical or technical illustration of a **meta-cognitive process in AI problem-solving**. It demonstrates not just solving a math problem, but the importance of **constraint satisfaction** and **self-verification**.
* **What it demonstrates:** It shows that for problems defined within bounded physical or geometric spaces, a mathematically correct solution to one part of the problem (area ratio) can be invalid if it violates other implicit constraints (point location). The system learns that robust problem-solving requires simultaneous consideration of all constraints.
* **Relationship between elements:** The "Advantage" section is the culmination. It uses the contrast between the two "Rollout" trajectories, as broken down in the "Summarization" sections, to extract a higher-order rule. This mimics a learning process where experience (trying and failing) is distilled into a principle.
* **Underlying message:** The diagram argues for AI agents that don't just execute algorithms but engage in **reflective reasoning**—checking their own intermediate results against the problem's full context. The final highlighted lesson is a direct instruction for future problem-solving behavior, aiming to prevent a entire class of errors. This is likely part of a framework for training or evaluating AI reasoning capabilities.
</details>
Figure 3: Example of a Training-Free GRPO learning step.
In this section, we introduce our Training-Free GRPO, a method designed to replicate the alignment benefits of the GRPO algorithm without performing any gradient-based updates to the policy model’s parameters.
Vanilla GRPO. As shown in Figure 2, the vanilla GRPO procedure operates by first generating a group of $G$ outputs $\{o_1,o_2,\dots,o_G\}$ for a given query $q$ using the current policy LLM $π_θ$ , i.e., $π_θ(o_i\mid q)$ . Each output $o_i$ is then independently scored with a reward model $R$ . Subsequently, with rewards $r=\{r_1,\dots,r_G\}$ , it calculates a group-relative advantage $\hat{A}_i=\frac{r_i-mean(r)}{std(r)}$ for each output $o_i$ . By combining a KL-divergence penalty against a reference model, it constructs a PPO-clipped objective function $J_GRPO(θ)$ , which is then maximized to update the LLM parameters $θ$ .
Training-Free GRPO repurposes the core logic of this group-based relative evaluation but translates it into a non-parametric, inference-time process. Instead of updating the parameters $θ$ , we leave $θ$ permanently frozen and maintain an external experiential knowledge $E$ , which is initialized to $∅$ .
Rollout and Reward. As shown in Figure 2, our rollout and reward process mirrors that of GRPO exactly. Given a query $q$ , we perform a parallel rollout to generate a group of $G$ outputs $\{o_1,o_2,\dots,o_G\}$ using the LLM. Notably, while GRPO uses the current trainable policy $π_θ$ , our policy conditions on the experiential knowledge, $π_θ(o_i|q,E)$ . Identical to the standard GRPO setup, we score each output $o_i$ by the reward model $R$ to obtain a scalar reward $r_i=R(q,o_i)$ .
Group Advantage Computation. To provide an optimization direction for policy parameters, vanilla GRPO computes a numerical advantage $\hat{A}_i$ that quantifies each output $o_i$ ’s relative quality within its group. Similarly, Training-Free GRPO performs an analogous comparison within each group but produces a group relative semantic advantage in the form of natural language experience, as shown in Figure 3. Since $\hat{A}_i=0$ when all $G$ outputs receive identical rewards (i.e., $std(r)=0$ ) in vanilla GRPO, we generate such semantic advantages only for groups with both clear winners and losers. Specifically, for each output $o_i$ , we first ask the same LLM $M$ to provide a corresponding summary $s_i=M(p_summary,q,o_i)$ separately, where $p_summary$ is a prompt template that incorporates the query $q$ and output $o_i$ to form a structured summarization request. Given the summaries $\{s_1,s_2,\dots,s_G\}$ and the current experiential knowledge $E$ , the LLM $M$ articulates the reasons for the relative success or failure of the outputs, followed by extracting a concise natural language experience $A_text=M({p_extract},q,s_i,E)$ , where $p_extract$ is another prompt template for experience extraction. This natural language experience $A_text$ serves as our semantic advantage, functionally equivalent to vanilla GRPO’s $\hat{A}_i$ , encoding the critical experiential knowledge of what actions lead to high rewards.
Optimization. Whereas vanilla GRPO updates its model parameters $θ$ via gradient ascent on $J_GRPO(θ)$ computed by all advantages in a single batch, we update our experience library $E$ using all semantic advantages $A_text$ from the current batch. Specifically, given the existing experiences library $E$ , we prompt the LLM to generate a list of operations based on all these $A_text$ , where each operation could be:
- Add: Directly append the experience described in $A_text$ to the experience library $E$ .
- Delete: Based on $A_text$ , remove a low-quality experience from the experience library $E$ .
- Modify: Refine or improve an existing experience in the experience library $E$ based on insights from $A_text$ .
- Keep: The experience library $E$ remains unchanged.
After updating the experience library $E$ , the conditioned policy $π_θ(y|q,E)$ produces a shifted output distribution in subsequent batches or epochs. This mirrors the effect of a GRPO policy update by steering the model towards higher-reward outputs, but achieves this by altering the context rather than the model’s fundamental parameters. The frozen base model $π_θ$ acts as a strong prior, ensuring output coherence and providing a built-in stability analogous to the KL-divergence constraint in GRPO that prevents the policy from deviating excessively from $π_ref$ .
## 3 Evalution
To compare Training-Free GRPO with competitive baselines, we conduct comprehensive experiments on both mathematical reasoning and web searching benchmarks.
### 3.1 Mathematical Reasoning
Benchmarks. We conduct our evaluation on the challenging AIME24 and AIME25 benchmarks [2], which are representative of complex, out-of-domain mathematical reasoning challenges. To ensure robust and statistically reliable results, we evaluate each problem with 32 independent runs and report the average Pass@1 score, which we denote as Mean@32.
Setup. We primarily focus on large powerful LLMs that are usually hard and expensive to be fine-tuned in real-world applications, such as DeepSeek-V3.1-Terminus [3]. We include two basic configurations: (1) Direct Prompting without tool use (a text-only input/output process), and (2) ReAct [1] with a code interpreter (CI) tool. To apply Training-Free GRPO experiments, we randomly sample 100 problems from the DAPO-Math-17K dataset [19], denoted as DAPO-100. We run the learning process for 3 epochs with a single batch per epoch (i.e., 3 steps), using a temperature of $0.7$ and a group size of $5$ during the learning phase. For out-of-domain evaluation on AIME 2024 and 2025 benchmarks, we use a temperature of $0.3$ .
Table 1: Mean@32 on AIME 2024 and AIME 2025 benchmarks (%).
| Direct | - | - | DeepSeek-V3.1-Terminus | - | 68.6 | 52.9 |
| --- | --- | --- | --- | --- | --- | --- |
| + Training-Free GRPO | $≈\mathdollar 8$ | DAPO-100 | DeepSeek-V3.1-Terminus | - | 72.6 ( $↑$ 4.0) | 54.0 ( $↑$ 1.1) |
| ReAct [1] | - | - | DeepSeek-V3.1-Terminus | CI | 80.0 | 67.9 |
| + Training-Free GRPO | $≈\mathdollar 18$ | DAPO-100 | DeepSeek-V3.1-Terminus | CI | 82.7 ( $↑$ 2.7) | 73.3 ( $↑$ 5.4) |
| ReAct [1] | - | - | DeepSeek-V3.2-Exp | CI | 71.0 | 61.8 |
| + Training-Free GRPO | $≈\mathdollar 8$ | DAPO-100 | DeepSeek-V3.2-Exp | CI | 73.1 ( $↑$ 2.1) | 63.2 ( $↑$ 1.4) |
<details>
<summary>x1.png Details</summary>

### Visual Description
## Bar Charts: Performance and Tool Calls Across Learning Steps
### Overview
The image contains two side-by-side bar charts comparing the performance and tool usage of different methods across three datasets (Train, AIME24, AIME25) over a series of learning steps. The charts share a common legend and x-axis structure.
### Components/Axes
* **Chart Type:** Grouped Bar Charts.
* **Shared X-Axis:** Labeled "Learning Step". It is divided into three main categories: **Train**, **AIME24**, and **AIME25**. Within each category, there are four sub-categories representing different methods or steps: **1**, **2**, **3**, and **ReAct**.
* **Left Chart Y-Axis:** Labeled "Performance". Scale ranges from 60 to 90 in increments of 5.
* **Right Chart Y-Axis:** Labeled "# Tool Calls". Scale ranges from 3.0 to 6.0 in increments of 0.5.
* **Legend (Bottom Center):** Defines the color coding for the bars:
* **Peach/Light Orange:** Mean@5 (Train)
* **Gray:** Mean@32 (ReAct)
* **Light Green:** Mean@32 (AIME24)
* **Light Cyan/Blue:** Mean@32 (AIME25)
### Detailed Analysis
#### Left Chart: Performance
* **Trend Verification:** The "Performance" metric generally shows an upward trend with increasing learning steps (1 -> 2 -> 3) within each dataset category, with the exception of the "ReAct" method which is consistently lower than step 3.
* **Data Points (Approximate Values):**
* **Train Category (Peach Bars):**
* Step 1: ~84.5
* Step 2: ~87.5
* Step 3: ~88.0
* **AIME24 Category (Green Bars):**
* ReAct: ~80.0
* Step 1: ~81.0
* Step 2: ~82.5
* Step 3: ~83.0
* **AIME25 Category (Cyan Bars):**
* ReAct: ~68.0
* Step 1: ~70.5
* Step 2: ~71.0
* Step 3: ~73.5
#### Right Chart: # Tool Calls
* **Trend Verification:** The "# Tool Calls" metric shows a generally decreasing trend with increasing learning steps (1 -> 2 -> 3) within each dataset category. The "ReAct" method shows a distinct pattern, with a very high value for AIME25.
* **Data Points (Approximate Values):**
* **Train Category (Peach Bars):**
* Step 1: ~3.75
* Step 2: ~3.60
* Step 3: ~3.50
* **AIME24 Category (Green Bars):**
* ReAct: ~4.40
* Step 1: ~4.10
* Step 2: ~3.80
* Step 3: ~3.80
* **AIME25 Category (Cyan Bars):**
* ReAct: ~5.60
* Step 1: ~5.10
* Step 2: ~4.85
* Step 3: ~4.75
### Key Observations
1. **Performance vs. Efficiency Trade-off:** There is a clear inverse relationship between the two charts. As learning steps progress (1->2->3), performance increases while the number of tool calls decreases, suggesting improved efficiency.
2. **ReAct Baseline:** The "ReAct" method (gray bars) serves as a baseline. The numbered learning steps (1, 2, 3) consistently outperform ReAct in the Performance metric across AIME24 and AIME25, while also using fewer tool calls in AIME24 and AIME25 (except for AIME25 ReAct's very high tool call count).
3. **Dataset Difficulty:** Performance is highest on the "Train" dataset, lower on "AIME24", and lowest on "AIME25". Conversely, tool calls are lowest on "Train" and highest on "AIME25", indicating AIME25 is the most challenging dataset.
4. **AIME25 ReAct Anomaly:** The ReAct method on the AIME25 dataset shows a particularly high number of tool calls (~5.6) coupled with the lowest performance (~68.0), highlighting its inefficiency on this specific task.
### Interpretation
The data demonstrates the effectiveness of a sequential learning process (steps 1, 2, 3) over a static ReAct approach. The progressive improvement in performance alongside a reduction in tool calls suggests the system is learning to solve problems more accurately and efficiently with each step. The consistent pattern across two different evaluation datasets (AIME24 and AIME25) indicates the learning method is robust. The significant performance drop and tool call increase for ReAct on AIME25 imply that this dataset contains problems where the standard ReAct paradigm struggles, possibly requiring more complex reasoning or planning that the step-wise learning process better accommodates. The charts collectively argue for the value of iterative learning in improving both the effectiveness (performance) and efficiency (reduced tool use) of an AI system on complex reasoning tasks.
</details>
Figure 4: Statistics at each Training-Free GRPO step with tool use and DeepSeek-V3.1-Terminus.
Main Results. As illustrated in Table 3, Training-Free GRPO achieves substantial gains in mathematical reasoning, showing a clear advantage in performance across both the tool-use and non-tool-use scenarios. The strong baseline established by DeepSeek-V3.1-Terminus with ReAct [1] yields scores of $80.0\$ on AIME24 and $67.9\$ on AIME25. Critically, applying Training-Free GRPO to the frozen DeepSeek-V3.1-Terminus elevates its performance significantly, reaching $82.7\$ on AIME24 and $73.3\$ on AIME25. This represents a substantial absolute gain of $+2.7\$ and $+5.4\$ , respectively, which is achieved with only $100$ out-of-domain training examples and zero gradient updates. This performance surpasses various state-of-the-art RL methods like ReTool [4] and AFM [16] trained on 32B LLMs (see Table 3), which typically require thousands of training examples and incur costs exceeding $\mathdollar 10,000$ . In contrast, Training-Free GRPO utilizes only $100$ data points with an approximate cost of $\mathdollar 18$ . Such outcome suggests that in real-world applications, guiding a powerful but frozen model through context-space optimization is more effective and efficient than exhaustively fine-tuning a less capable model in parameter space. For quantitative evaluation and learned experiences of Training-Free GRPO, please refer to Appendices A and C, respectively.
Learning Dynamics. As shown in Figure 4, during the 3-step learning process, we observe a steady and significant improvement in Mean@5 on the training set. Concurrently, the Mean@32 performance on both AIME24 and AIME25 also improves with each step, demonstrating that the learned experiences from only $100$ problems generalize effectively and the necessity of multi-step learning. Also, the average number of tool calls decreases during both training and out-of-domain evaluation on AIME benchmarks. This suggests that Training-Free GRPO not only encourages correct reasoning and action, but also teaches the agent to use tools more efficiently and judiciously. The learned experiential knowledge helps the agent to discover some shortcuts and avoid erroneous or redundant tool calls, validating the effectiveness of our semantic advantage guided optimization.
Table 2: Ablation study on DeepSeek-V3.1-Terminus (Mean@32, %).
| ReAct [1] | - | 80.0 | 67.9 |
| --- | --- | --- | --- |
| ReAct [1] + Directly Generated Experiences | - | 79.8 | 67.3 |
| ReAct [1] + Training-Free GRPO (w/o ground truths) | DAPO-100 | 80.7 | 68.9 |
| ReAct [1] + Training-Free GRPO (w/o group computation) | DAPO-100 | 80.4 | 69.3 |
| ReAct [1] + Training-Free GRPO | DAPO-100 | 82.7 | 73.3 |
Table 3: Mean@32 with smaller LLMs on AIME 2024 and AIME 2025 benchmarks (%).
| Method | Learning Cost | Model | Tool | AIME24 | AIME25 |
| --- | --- | --- | --- | --- | --- |
| ReAct [1] | - | Qwen2.5-32B-Instruct | CI | 29.6 | 23.1 |
| ZeroTIR [5] | $≈\mathdollar 20,000$ | Qwen2.5-32B-Instruct | CI | 56.7 | 33.3 |
| SimpleTIR [6] | $≈\mathdollar 20,000$ | Qwen2.5-32B-Instruct | CI | 59.9 | 49.2 |
| ReTool [4] | $≈\mathdollar 10,000$ | Qwen2.5-32B-Instruct | CI | 67.0 | 49.3 |
| AFM [16] | $≈\mathdollar 10,000$ | Qwen2.5-32B-Instruct | CI | 66.7 | 59.8 |
| ReAct [1] | - | Qwen3-32B (Non-Thinking) | CI | 29.1 | 19.5 |
| + Training-Free GRPO | $≈\mathdollar 4$ | Qwen3-32B (Non-Thinking) | CI | 33.5 ( $↑$ 4.4) | 25.4 ( $↑$ 5.9) |
| ReAct [1] | - | Qwen2.5-72B-Instruct | CI | 13.5 | 9.6 |
| + Training-Free GRPO | $≈\mathdollar 3$ | Qwen2.5-72B-Instruct | CI | 14.9 ( $↑$ 1.4) | 11.4 ( $↑$ 1.8) |
| ReAct [1] | - | DeepSeek-V3.1-Terminus | CI | 80.0 | 67.9 |
| + Training-Free GRPO | $≈\mathdollar 18$ | DeepSeek-V3.1-Terminus | CI | 82.7 ( $↑$ 2.7) | 73.3 ( $↑$ 5.4) |
Effectiveness of Learned Experiences. In Table 2, we also include the ReAct enhanced with the experiences directly generated by DeepSeek-V3.1-Terminus, matching the quantity learned from Training-Free GRPO. However, such directly generated experiences fail to improve the performance, highlighting the effectiveness of learned experiential knowledge from Training-Free GRPO.
Robustness to Reward Signal. Table 2 also presents a variant of Training-Free GRPO, where the ground truth answers are not provided during learning process. In such cases, the semantic advantage is directly obtained by comparing the rollouts within each group, where the LLM can only rely on implicit majority voting, self-discrimination and self-reflection to optimize the experiences. Although it does not surpass the default version with ground truths, Training-Free GRPO still achieves an impressive results of 80.7% on AIME24 and 68.9% on AIME25. It shows the robustness and applicability to domains where ground truths are scarce or unavailable, further broadening its practical utility.
Removing Group Computation. We also remove the group computation by setting the group size of $1$ in Training-Free GRPO, where the LLM can only distill semantic advantage from the single rollout of each query. The results in Table 2 show that it significantly harm the performance comparing with the default group size of $5$ . It confirms the necessity of group relative computation, which enables the LLM to compare different trajectories within each group for better semantic advantage and experience optimization.
Applicability to Different Model Sizes. By applying Training-Free GRPO on smaller LLMs, specifically Qwen3-32B [22] and Qwen2.5-72B-Instruct [23] with DAPO-100 dataset, we observe consistent improvements on the out-of-domain AIME benchmarks in Table 3. Training-Free GRPO requires significantly fewer data and much lower learning cost, contrasting sharply with recent RL methods like ZeroTIR [5], SimpleTIR [6], ReTool [4], and AFM [16], which often necessitate thousands of data points and substantial computational resources for parameter tuning. Furthermore, powered by larger models like DeepSeek-V3.1-Terminus, our approach achieves much higher Mean@32 on AIME benchmarks than all RL-trained models, while only incurring about $\mathdollar 18$ for the learning process.
### 3.2 Web Searching
Table 4: Pass@1 on WebWalkerQA (%).
| ReAct [1] + Training-Free GRPO | - AFM-100 | DeepSeek-V3.1-Terminus DeepSeek-V3.1-Terminus | 63.2 67.8 ( $↑$ 4.6) |
| --- | --- | --- | --- |
In this section, we evaluate the effectiveness of Training-Free GRPO in addressing web searching tasks by leveraging minimal experiential data to enhance agent behavior.
Datasets. For training, we constructed a minimal training set by randomly sampling 100 queries from the AFM (Chain-of-Agents) web interaction RL dataset [16], denoted as AFM-100. AFM provides high-quality, multi-turn interactions between agents and web environments, collected via reinforcement learning in realistic browsing scenarios. For evaluation, we employ WebWalkerQA benchmark [24], a widely-used dataset for assessing web agent performance. Its tasks require understanding both natural language instructions and complex web page structures, making it a rigorous evaluation framework for generalist agents.
Methods. Our proposed Training-Free GRPO is applied to DeepSeek-V3.1-Terminus without any gradient-based updates. We perform 3 epochs of training-free optimization with a group size of $G=3$ . The temperature settings follow those used in prior mathematical experiments.
Main Results. We evaluate the effectiveness of our proposed Training-Free GRPO method on the WebWalkerQA benchmark. As shown in Table 4, our method achieves a pass@1 score of 67.8% using DeepSeek-V3.1-Terminus, a significant improvement over the baseline of 63.2%. This result indicates that our approach effectively steers model behavior by leveraging learned experiential knowledge, surpassing the capabilities of static prompting ReAct strategy.
Table 5: Ablation results on WebWalkerQA subset (%).
| Method | Training Set | Model | pass@1 | pass@3 |
| --- | --- | --- | --- | --- |
| ReAct [1] | - | QwQ-32B | 27.5 | 43.1 |
| ReAct [1] + Training-Free GRPO | AFM-100 | QwQ-32B | 25.5 | 45.1 |
| ReAct [1] | - | DeepSeek-V3.1-Terminus | 66.7 | 74.5 |
| ReAct [1] + Directly Generated Experiences | AFM-100 | DeepSeek-V3.1-Terminus | 64.7 | 76.5 |
| ReAct [1] + Training-Free GRPO (w/o ground truths) | AFM-100 | DeepSeek-V3.1-Terminus | 66.7 | 78.4 |
| ReAct [1] + Training-Free GRPO | AFM-100 | DeepSeek-V3.1-Terminus | 68.6 | 78.4 |
Ablation. We conduct ablation studies on a stratified random sample of 51 instances from the WebWalkerQA test set, where the sampling is proportionally stratified by difficulty level to ensure balanced representation across different levels of complexity. All ablated models are evaluated after 2 epochs of experience optimization. The results are summarized in Table 5.
Using directly generated experiences slightly degrades over ReAct (64.7% vs. 66.7% pass@1), confirming that mere in-context examples without proper optimization may not yield gains. Training-Free GRPO without ground truth maintains the same pass@1 as ReAct (66.7%) but improves pass@3 to 78.4%, demonstrating that relative reward evaluation can enhance consistency even without ground truth. The full Training-Free GRPO achieves the best performance (68.6% pass@1 and 78.4% pass@3), highlighting the importance of combining ground truth guidance with semantic advantage and experience optimization.
Applying Training-Free GRPO to QwQ-32B [25] yields only 25.5% pass@1, significantly lower than the 66.7% achieved with DeepSeek-V3.1-Terminus, and even under performing its own ReAct baseline (27.5%). This may suggest that the effectiveness of our method is dependent on the underlying model’s reasoning and tool-use capabilities in complex tool use scenarios, indicating that model capability is a prerequisite for effective experience-based optimization.
## 4 Comparing RL Learning on Context Space and Parameter Space
### 4.1 Cross-domain Transfer Analysis
Table 6: Cross-domain transferability (Averaged pass@1, %).
| Method | Learned Domain | Math Reasoning | Web Searching | |
| --- | --- | --- | --- | --- |
| AIME24 | AIME25 | WebWalker | | |
| ReAct [1] (Qwen2.5-32B-Instruct) | - | 29.6 | 23.1 | 31.9 |
| ReTool [4] (Qwen2.5-32B-Instruct) | Math | 67.0 | 49.3 | 18.3 |
| MiroThinker [10] (Qwen3-32B) | Web | 43.5 | 36.8 | 53.6 |
| Training-Free GRPO (DeepSeek-V3.1-Terminus) | Math / Web | 82.7 | 73.3 | 67.8 |
A critical strength of Training-Free GRPO lies in its ability to achieve strong performance across diverse domains without suffering from the domain specialization trade-off observed in parameter-tuned methods. As demonstrated in Table 6, we observe the unsatisfactory performance when domain-specialized models are transferred to different domains. For instance, ReTool [4] specifically trained on mathematical reasoning tasks, achieves competitive performance on AIME24 and AIME25 within its specialized domain. However, when transferred to web searching tasks on WebWalker, its performance drops dramatically to only $18.3\$ , which is much lower than ReAct [1] without fine-tuning. Similarly, though optimized for web interactions, MiroThinker [10] significantly underperforms ReTool that is trained in the math domain on the AIME benchmarks. Such phenomenon highlights that parameter-based specialization narrows the model’s capabilities to excel in the training domain at the expense of generalizability. In contrast, Training-Free GRPO applied to the frozen LLM achieves state-of-the-art performance in both domains simultaneously by simply plugging in domain-specific learned experiences. Such cross-domain robustness makes Training-Free GRPO particularly valuable for real-world applications where agents must operate in multifaceted environments with diverse requirements.
### 4.2 Computational Costs
As shown in Figure 1, we further analyze the economic benefits of Training-Free GRPO by comparing its computational costs with a vanilla RL training approach, specifically ReTool [4], on mathematical problem-solving tasks. This comparison highlights the practical benefits of our method in scenarios characterized by limited data, constrained budgets, or volatile inference demand.
Training Cost. By replicating the training process of ReTool [4] on Qwen2.5-32B-Instruct [23], we find that it requires approximately 20,000 $×$ GPU hours with the rental price of $0.5 per GPU hour, resulting in the total training expense amounts to roughly $10,000. In contrast, Training-Free GRPO, when applied to DeepSeek-V3.1-Terminus, achieves superior performance on the AIME benchmarks (Table 3) while requiring only minimal costs. It requires only $3$ training steps over $100$ samples completed within $6$ hours, which consumes 38M input tokens and 6.6M output tokens, amounting to a total cost of approximately $18 based on the official DeepSeek AI pricing https://api-docs.deepseek.com/quick_start/pricing. Most of the input tokens qualify for the lower cache hit pricing with ReAct framework, as each processing step typically involves reusing extensive prior context.. The drastic reduction in training cost by over two orders of magnitude makes our approach especially cost-effective.
Inference Cost. Deploying a trained model like ReTool-32B entails significant fixed infrastructure costs. In a typical serving setup with 4 $×$ GPUs at the price of $0.5 per GPU-hour, vLLM-based batching requests can process about 400 problems per hour from the AIME benchmarks. The inference cost per problem averages $0.005. While this per-instance cost is relatively low, it presupposes continuous GPU availability, which becomes inefficient under fluctuating or low request volumes. In contrast, Training-Free GRPO incurs a token-based cost. On average, each request consumes 60K input tokens and 8K output tokens, totaling about $0.02 per problem with cache hit pricing for reused contexts. Although per-query inference with a large API-based model is more expensive than with a dedicated small model, many real-world applications, particularly specialized or low-traffic services, experience irregular and modest usage patterns. In such cases, maintaining a dedicated GPU cluster is economically unjustifiable. By leveraging the shared, on-demand infrastructure of large model services like DeepSeek, Training-Free GRPO eliminates fixed serving overhead and aligns costs directly with actual usage. This pay-as-you-go model is distinctly advantageous in settings where demand is unpredictable or sparse.
## 5 Related Work
LLM Agents. By leveraging external tools, Large Language Models (LLMs) can overcome inherent limitations, such as lacking real-time knowledge and precise computation. This has spurred the development of LLM agents that interleave reasoning with actions. Foundational frameworks like ReAct [1] prompt LLMs to generate explicit chain-of-thought (CoT) and actionable steps, enabling dynamic planning through tool use. Furthermore, Toolformer [26] demonstrates that LLMs can learn to self-supervise the invocation of APIs via parameter fine-tuning. Building on these principles, subsequent research has produced sophisticated single- and multi-agent systems, such as MetaGPT [27], CodeAct [28], and OWL [29], which significantly enhance the quality of planning, action execution, and tool integration.
Reinforcement Learning. Reinforcement learning (RL) has proven highly effective for aligning LLMs with complex and long-horizon goals. Foundational algorithms like Proximal Policy Optimization (PPO) [30] employ a policy model for generation and a separate critic model to estimate token-level value. Group Relative Policy Optimization (GRPO) [17] eliminates the need for a critic by estimating advantages directly from groups of responses. Recent research try to apply RL to transform LLMs from passive generators into autonomous agents that learn through environmental interaction. GiGPO [31] implements a two-level grouping mechanism for trajectories, enabling precise credit assignment at both the episode and individual step levels. ReTool [4] uses PPO to train an agent to interleave natural language with code execution for mathematical reasoning. Chain-of-Agents [16] facilitates multi-agent collaboration within a single model by using dynamic, context-aware activation of specialized tool and role-playing agents. Furthermore, Tongyi Deep Research [7] introduces synthetic data generation pipeline and conduct customized on-policy agentic RL framework. However, such parameter-updating approaches result in prohibitive computational cost, which typically restricts application to LLMs with fewer than 32B parameters. Moreover, they only achieve diminishing returns compared to simply using larger, more powerful frozen LLMs. In contrast, our proposed Training-Free GRPO method seeks to achieve comparable or even better performance on state-of-the-art LLMs without any parameter updates, drastically reducing both data and computational requirements.
Training-Free Methods. A parallel line of research aims to improve LLM behavior at inference time without updating model weights. The general approach is in-context learning (ICL) [21], which leverages external or self-generated demonstrations within a prompt to induce desired behaviors. More recent methods introduce iterative refinement mechanisms. Self-Refine [32] generates an initial output and then uses the same LLM to provide verbal feedback for subsequent revisions. Similarly, Reflexion [33] incorporates an external feedback signal to prompt the model for reflection and a new attempt. In-context reinforcement learning (ICRL) [34, 35] demonstrates that LLMs can learn from scalar reward signals by receiving prompts containing their past outputs and associated feedback. TextGrad [15] proposes a more general framework, treating optimization as a process of back-propagating textual feedback through a structured computation graph. A key characteristic of these methods is their focus on iterative, within-sample improvement for a single query. In contrast, our Training-Free GRPO more closely mirrors traditional RL by learning from a separate dataset across multiple epochs to iteratively refine a shared, high-quality experience library for out-of-domain queries. Furthermore, given each query, unlike self-critique or context updates for a single trajectory, our method explicitly compares multiple rollouts per query for a semantic advantage to compare different trajectories in each group, which has been confirmed effective in Section 3.1. Specifically in the context for optimizing agent systems, Agent KB [36] constructs a shared, hierarchical knowledge base to enable the reuse of problem-solving experiences across tasks. Unlike the complex reason-retrieve-refine process of Agent KB, Training-Free GRPO simply injects the learned experiences into the prompt. Moreover, Agent KB relies on hand-crafted examples and employs an off-policy learning paradigm only once by collecting trajectories in the different way of online inference. In contrast, our Training-Free GRPO uses a consistent pipeline and more closely mirrors on-policy RL with multi-epoch learning.
## 6 Conclusion
In this paper, we introduced Training-Free GRPO, a novel paradigm that shifts RL policy optimization from the parameter space to the context space. By leveraging group-based rollouts to iteratively distill a semantic advantage into an evolving experiential knowledge which serves as the token prior, our method successfully steers the output distribution of a frozen LLM agent, achieving significant performance gains in specialized domains. Experiments demonstrate that Training-Free GRPO not only surmounts the practical challenges of data scarcity and high computational cost but also outperforms traditional parameter-tuning methods. Our work establishes a new, highly efficient pathway for adapting powerful LLM agents, making advanced agentic capabilities more accessible and practical for real-world applications.
## Contributions
Authors Yuzheng Cai 1,2* Siqi Cai 1* Yuchen Shi 1* Zihan Xu 1* Lichao Chen 1,3 Yulei Qin 1 Xiaoyu Tan 1 Gang Li 1 Zongyi Li 1 Haojia Lin 1 Yong Mao 1 Ke Li 1Q Xing Sun 1
Affiliations 1 Tencent Youtu Lab 2 Fudan University 3 Xiamen University
*Equal Contributions Yuzheng Cai Siqi Cai Yuchen Shi Zihan Xu
## References
- Yao et al. [2023] Shunyu Yao, Jeffrey Zhao, Dian Yu, Nan Du, Izhak Shafran, Karthik Narasimhan, and Yuan Cao. React: Synergizing reasoning and acting in language models. In International Conference on Learning Representations (ICLR), 2023.
- AIME [2025] AIME. Aime problems and solutions, 2025. URL https://artofproblemsolving.com/wiki/index.php/AIME_Problems_and_Solutions.
- DeepSeek-AI [2024] DeepSeek-AI. Deepseek-v3 technical report, 2024. URL https://arxiv.org/abs/2412.19437.
- Feng et al. [2025a] Jiazhan Feng, Shijue Huang, Xingwei Qu, Ge Zhang, Yujia Qin, Baoquan Zhong, Chengquan Jiang, Jinxin Chi, and Wanjun Zhong. Retool: Reinforcement learning for strategic tool use in llms. arXiv preprint arXiv:2504.11536, 2025a.
- Mai et al. [2025] Xinji Mai, Haotian Xu, Weinong Wang, Jian Hu, Yingying Zhang, Wenqiang Zhang, et al. Agent rl scaling law: Agent rl with spontaneous code execution for mathematical problem solving. arXiv preprint arXiv:2505.07773, 2025.
- Xue et al. [2025] Zhenghai Xue, Longtao Zheng, Qian Liu, Yingru Li, Xiaosen Zheng, Zejun Ma, and Bo An. Simpletir: End-to-end reinforcement learning for multi-turn tool-integrated reasoning. arXiv preprint arXiv:2509.02479, 2025.
- Tongyi DeepResearch Team [2025] Tongyi DeepResearch Team. Tongyi-deepresearch. https://github.com/Alibaba-NLP/DeepResearch, 2025.
- Tao et al. [2025] Zhengwei Tao, Jialong Wu, Wenbiao Yin, Junkai Zhang, Baixuan Li, Haiyang Shen, Kuan Li, Liwen Zhang, Xinyu Wang, Yong Jiang, et al. Webshaper: Agentically data synthesizing via information-seeking formalization. arXiv preprint arXiv:2507.15061, 2025.
- Jin et al. [2025] Bowen Jin, Hansi Zeng, Zhenrui Yue, Jinsung Yoon, Sercan Arik, Dong Wang, Hamed Zamani, and Jiawei Han. Search-r1: Training llms to reason and leverage search engines with reinforcement learning. arXiv preprint arXiv:2503.09516, 2025.
- Team [2025a] MiroMind AI Team. Mirothinker: An open-source agentic model series trained for deep research and complex, long-horizon problem solving. https://github.com/MiroMindAI/MiroThinker, 2025a.
- Zhang et al. [2024] Kechi Zhang, Jia Li, Ge Li, Xianjie Shi, and Zhi Jin. Codeagent: Enhancing code generation with tool-integrated agent systems for real-world repo-level coding challenges. arXiv preprint arXiv:2401.07339, 2024.
- Huang et al. [2023] Dong Huang, Jie M Zhang, Michael Luck, Qingwen Bu, Yuhao Qing, and Heming Cui. Agentcoder: Multi-agent-based code generation with iterative testing and optimisation. arXiv preprint arXiv:2312.13010, 2023.
- Wang et al. [2024a] Shuai Wang, Weiwen Liu, Jingxuan Chen, Yuqi Zhou, Weinan Gan, Xingshan Zeng, Yuhan Che, Shuai Yu, Xinlong Hao, Kun Shao, et al. Gui agents with foundation models: A comprehensive survey. arXiv preprint arXiv:2411.04890, 2024a.
- Wang et al. [2024b] Junyang Wang, Haiyang Xu, Haitao Jia, Xi Zhang, Ming Yan, Weizhou Shen, Ji Zhang, Fei Huang, and Jitao Sang. Mobile-agent-v2: Mobile device operation assistant with effective navigation via multi-agent collaboration. Advances in Neural Information Processing Systems, 37:2686–2710, 2024b.
- Yuksekgonul et al. [2025] Mert Yuksekgonul, Federico Bianchi, Joseph Boen, Sheng Liu, Pan Lu, Zhi Huang, Carlos Guestrin, and James Zou. Optimizing generative ai by backpropagating language model feedback. Nature, 639(8055):609–616, 2025.
- Li et al. [2025] Weizhen Li, Jianbo Lin, Zhuosong Jiang, Jingyi Cao, Xinpeng Liu, Jiayu Zhang, Zhenqiang Huang, Qianben Chen, Weichen Sun, Qiexiang Wang, Hongxuan Lu, Tianrui Qin, Chenghao Zhu, Yi Yao, Shuying Fan, Xiaowan Li, Tiannan Wang, Pai Liu, King Zhu, He Zhu, Dingfeng Shi, Piaohong Wang, Yeyi Guan, Xiangru Tang, Minghao Liu, Yuchen Eleanor Jiang, Jian Yang, Jiaheng Liu, Ge Zhang, and Wangchunshu Zhou. Chain-of-agents: End-to-end agent foundation models via multi-agent distillation and agentic rl. 2025. URL https://arxiv.org/abs/2508.13167.
- Shao et al. [2024] Zhihong Shao, Peiyi Wang, Qihao Zhu, Runxin Xu, Junxiao Song, Xiao Bi, Haowei Zhang, Mingchuan Zhang, YK Li, Yang Wu, et al. Deepseekmath: Pushing the limits of mathematical reasoning in open language models. arXiv preprint arXiv:2402.03300, 2024.
- Liu et al. [2025] Zichen Liu, Changyu Chen, Wenjun Li, Penghui Qi, Tianyu Pang, Chao Du, Wee Sun Lee, and Min Lin. Understanding r1-zero-like training: A critical perspective. arXiv preprint arXiv:2503.20783, 2025.
- Yu et al. [2025] Qiying Yu, Zheng Zhang, Ruofei Zhu, Yufeng Yuan, Xiaochen Zuo, Yu Yue, Weinan Dai, Tiantian Fan, Gaohong Liu, Lingjun Liu, et al. Dapo: An open-source llm reinforcement learning system at scale. arXiv preprint arXiv:2503.14476, 2025.
- Zheng et al. [2025] Chujie Zheng, Shixuan Liu, Mingze Li, Xiong-Hui Chen, Bowen Yu, Chang Gao, Kai Dang, Yuqiong Liu, Rui Men, An Yang, et al. Group sequence policy optimization. arXiv preprint arXiv:2507.18071, 2025.
- Brown et al. [2020] Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
- Yang et al. [2025a] An Yang, Anfeng Li, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu, Chang Gao, Chengen Huang, Chenxu Lv, et al. Qwen3 technical report. arXiv preprint arXiv:2505.09388, 2025a.
- Yang et al. [2025b] An Yang, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu, Chengyuan Li, Dayiheng Liu, Fei Huang, Haoran Wei, Huan Lin, Jian Yang, Jianhong Tu, Jianwei Zhang, Jianxin Yang, Jiaxi Yang, Jingren Zhou, Junyang Lin, Kai Dang, Keming Lu, Keqin Bao, Kexin Yang, Le Yu, Mei Li, Mingfeng Xue, Pei Zhang, Qin Zhu, Rui Men, Runji Lin, Tianhao Li, Tianyi Tang, Tingyu Xia, Xingzhang Ren, Xuancheng Ren, Yang Fan, Yang Su, Yichang Zhang, Yu Wan, Yuqiong Liu, Zeyu Cui, Zhenru Zhang, and Zihan Qiu. Qwen2.5 technical report, 2025b. URL https://arxiv.org/abs/2412.15115.
- Wu et al. [2025] Jialong Wu, Wenbiao Yin, Yong Jiang, Zhenglin Wang, Zekun Xi, Runnan Fang, Linhai Zhang, Yulan He, Deyu Zhou, Pengjun Xie, and Fei Huang. Webwalker: Benchmarking llms in web traversal. 2025. URL https://arxiv.org/abs/2501.07572.
- Team [2025b] Qwen Team. Qwq-32b: Embracing the power of reinforcement learning, March 2025b. URL https://qwenlm.github.io/blog/qwq-32b/.
- Schick et al. [2023] Timo Schick, Jane Dwivedi-Yu, Roberto Dessì, Roberta Raileanu, Maria Lomeli, Luke Zettlemoyer, Nicola Cancedda, and Thomas Scialom. Toolformer: Language models can teach themselves to use tools, 2023. arXiv preprint arXiv:2302.04761, 2023.
- Hong et al. [2024] Sirui Hong, Mingchen Zhuge, Jonathan Chen, Xiawu Zheng, Yuheng Cheng, Ceyao Zhang, Jinlin Wang, Zili Wang, Steven Ka Shing Yau, Zijuan Lin, et al. MetaGPT: Meta programming for a multi-agent collaborative framework. International Conference on Learning Representations, ICLR, 2024.
- Wang et al. [2024c] Xingyao Wang, Yangyi Chen, Lifan Yuan, Yizhe Zhang, Yunzhu Li, Hao Peng, and Heng Ji. Executable code actions elicit better llm agents. In Forty-first International Conference on Machine Learning, 2024c.
- Hu et al. [2025] Mengkang Hu, Yuhang Zhou, Wendong Fan, Yuzhou Nie, Bowei Xia, Tao Sun, Ziyu Ye, Zhaoxuan Jin, Yingru Li, Qiguang Chen, Zeyu Zhang, Yifeng Wang, Qianshuo Ye, Bernard Ghanem, Ping Luo, and Guohao Li. OWL: Optimized workforce learning for general multi-agent assistance in real-world task automation, 2025. URL https://arxiv.org/abs/2505.23885.
- Schulman et al. [2017] John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov. Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347, 2017.
- Feng et al. [2025b] Lang Feng, Zhenghai Xue, Tingcong Liu, and Bo An. Group-in-group policy optimization for llm agent training. arXiv preprint arXiv:2505.10978, 2025b.
- Madaan et al. [2023] Aman Madaan, Niket Tandon, Prakhar Gupta, Skyler Hallinan, Luyu Gao, Sarah Wiegreffe, Uri Alon, Nouha Dziri, Shrimai Prabhumoye, Yiming Yang, et al. Self-refine: Iterative refinement with self-feedback. Advances in Neural Information Processing Systems, 36:46534–46594, 2023.
- Shinn et al. [2023] Noah Shinn, Federico Cassano, Ashwin Gopinath, Karthik Narasimhan, and Shunyu Yao. Reflexion: Language agents with verbal reinforcement learning. Advances in Neural Information Processing Systems, 36:8634–8652, 2023.
- Song et al. [2025] Kefan Song, Amir Moeini, Peng Wang, Lei Gong, Rohan Chandra, Yanjun Qi, and Shangtong Zhang. Reward is enough: Llms are in-context reinforcement learners. arXiv preprint arXiv:2506.06303, 2025.
- Monea et al. [2024] Giovanni Monea, Antoine Bosselut, Kianté Brantley, and Yoav Artzi. Llms are in-context reinforcement learners. 2024.
- Tang et al. [2025] Xiangru Tang, Tianrui Qin, Tianhao Peng, Ziyang Zhou, Daniel Shao, Tingting Du, Xinming Wei, Peng Xia, Fang Wu, He Zhu, et al. Agent KB: Leveraging cross-domain experience for agentic problem solving. arXiv preprint arXiv:2507.06229, 2025.
## Appendix A Case Study
On math reasoning and web searching scenarios, we provide examples illustrating how Training-Free GRPO enhances ReAct [1] with DeepSeek-V3.1-Terminus and corresponding tools.
### A.1 Experience-Guided Tool-Integrated Math Reasoning
We consider a geometric configuration with two rectangles $ABCD$ and $EFGH$ where $D,E,C,F$ are collinear in that order, and $A,D,H,G$ are concyclic. Given $BC=16$ , $AB=107$ , $FG=17$ , $EF=184$ , the task is to determine $CE$ .
Tool-integrated Math Reasoning (without experiences)
Question Rectangles $ABCD$ and $EFGH$ are drawn such that $D,E,C,F$ are collinear. Also, $A,D,H,G$ all lie on a circle. If $BC=16$ , $AB=107$ , $FG=17$ , and $EF=184$ , what is the length of $CE$ ?
Trajectory
[Agent]
I’ll place point $D$ at the origin $(0,0)$ and use a horizontal line for $D,E,C,F$ . For rectangle $ABCD$ , set $A=(0,16)$ , $B=(107,16)$ , $C=(107,0)$ , $D=(0,0)$ .
[Tool]
Set coordinates accordingly.
[Agent]
Because $EFGH$ is a rectangle with $EF$ horizontal and $FG=17$ , set $F=(x_f,0)$ , $E=(x_f-184,0)$ , and initially set $G=(x_f,{\color[rgb]{0.94140625,0.1875,0.625}\definecolor[named]{pgfstrokecolor}{rgb}{0.94140625,0.1875,0.625}+17})$ , $H=(x_f-184,{\color[rgb]{0.94140625,0.1875,0.625}\definecolor[named]{pgfstrokecolor}{rgb}{0.94140625,0.1875,0.625}+17})$ .
[Agent]
Impose concyclicity of $A(0,16),D(0,0),G(x_f,17),H(x_f-184,17)$ using the 4-point determinant condition, derive an equation in $x_f$ .
[Tool]
Solve the resulting quadratic for $x_f$ ; obtain $x_f≈ 184.092$ and thus $x_e=x_f-184≈ 0.092$ .
[Agent]
Compute $CE=|107-x_e|≈ 106.9076$ .
[Agent]
Swap point orderings ( $D$ – $C$ – $E$ – $F$ , $E$ – $D$ – $C$ – $F$ ) and re-derive. At one point, use an angle-sine relation and get $x_e(x_e+184)=17$ leading to $x_e=-92+√{8481}≈ 0.0923$ , hence $CE≈ 106.9076$ .
[Agent]
Present an “exact form” $CE=199-√{8481}$ as the final answer.
Outcome Incorrect: $CE≈ 106.9076$ (reported as $199-√{8481}$ ).
Figure 5: Example trajectory without experience in Math Reasoning.
Baseline (without experiences). As shown in Figure 5, the unassisted agent initializes a coordinate system with $D=(0,0)$ and models $ABCD$ as axis-aligned. For $EFGH$ , it places $E=(x_f-184,0)$ , $F=(x_f,0)$ and, critically, sets $G=(x_f, +17)$ , $H=(x_f-184, +17)$ , i.e., with a positive vertical orientation for the short side. It then enforces the four-point concyclicity of $A(0,16)$ , $D(0,0)$ , $G$ , $H$ via a determinant condition and solves for $x_f$ , yielding $x_f≈ 184.092$ and consequently $x_e=x_f-184≈ 0.092$ . From this, it reports $CE≈ 106.9076$ and an “exact” expression $199-√{8481}$ .
This trajectory exhibits three systemic issues: (i) misinterpretation of the vertical orientation (wrong sign for the $y$ -coordinates of $G,H$ ), (ii) inconsistent handling of the order $D$ – $E$ – $C$ – $F$ and the lack of a unified parameterization for segment relations, and (iii) absence of systematic, comprehensive post-solution verification—i.e., no integrated check that the final coordinates simultaneously satisfy rectangle dimensions. These issues lead to an incorrect cyclic constraint (e.g., an intermediate relation of the form $x(x+184)=17$ ) and acceptance of a spurious solution without full geometric verification. Note that although $CE≈ 106.91$ lies within $0<CE<107$ , this alone does not validate the solution; the critical failure was the lack of holistic consistency checks across all problem constraints.
Tool-integrated Math Reasoning (with experiences)
Question Rectangles $ABCD$ and $EFGH$ are drawn such that $D,E,C,F$ are collinear. Also, $A,D,H,G$ all lie on a circle. If $BC=16$ , $AB=107$ , $FG=17$ , and $EF=184$ , what is the length of $CE$ ? Experiences [1] When solving geometry problems with intersections, validate solutions lie within bounded regions or segments, not on extensions, to avoid extraneous answers. $⋯$ [7] When solving geometry with constraints (e.g., trapezoids), validate all conditions: parallel sides via slopes, convexity via cross products, to avoid extraneous solutions. $⋯$ [29] When constructing points on line extensions, carefully interpret directional phrases like ’beyond B’. $⋯$ [37] In geometry problems with points on sides of a triangle and given segment lengths, first determine all three side lengths by summing the appropriate segments. For example, if points on AB, AC, BC are given with AM, MB, AN, NC, BP, PC, then AB = AM + MB, AC = AN + NC, BC = BP + PC. This ensures the triangle is consistently defined before coordinate placement or area calculations. $⋯$
Trajectory
[Agent]
Fix the order $D$ – $E$ – $C$ – $F$ on a horizontal line. Set $CE=x$ with the segment constraint $0<x<107$ . (By [G29],[G1])
[Tool]
Record relations: $DE=107-x$ and $CF=184-x$ (from [G37]).
[Tool]
Place $D=(0,0)$ , $E=(107-x,0)$ , $C=(107,0)$ , $F=(291-x,0)$ , ensuring $EF=184$ .
[Agent]
Choose the vertical orientation consistently: since $A=(0,16)$ and $D=(0,0)$ , set the short side $FG=17$ downward, i.e., $H=(107-x,-17)$ and $G=(291-x,-17)$ . (This corrects the sign error in the no-experience path)
[Agent]
Impose concyclicity using the circle equation $x^2+y^2+Dx+Ey+F=0$ . With $A(0,16)$ and $D(0,0)$ , get $F=0$ , $E=-16$ .
[Tool]
Substitute $H(107-x,-17)$ and $G(291-x,-17)$ to obtain two equations. Subtract to eliminate constants and get $D=2x-398$ .
[Agent]
Substitute back to get the quadratic $x^2-398x+30576=0.$
[Tool]
Compute the discriminant: $398^2-4· 30576=36100=190^2$ . Roots: $x=294$ and $x=104$ .
[Agent]
Apply the segment constraint $0<x<107$ to discard $x=294$ and select $x=104$ (by [G1])
[Agent]
(Full verification by [G7]) $DE=107-104=3$ , $CF=184-104=80$ , $EF=184$ , $FG=17$ . The circle $x^2+y^2-190x-16y=0$ passes through $A=(0,16),D=(0,0),H=(3,-17),G=(187,-17)$ . All constraints satisfied.
Outcome Correct: $CE=104$ .
Figure 6: Example trajectory with experiences in Math Reasoning.
Enhanced (With Experiences). Refer to Figure 6, with a curated experience pool, the agent follows a structured pipeline:
- Directional ordering ([29]) and boundedness validation ([1]): It fixes the order $D$ – $E$ – $C$ – $F$ on a line and sets $CE=x$ with $0<x<107$ , ensuring $E$ lies on segment $DC$ and $F$ lies beyond $C$ .
- Segment-addition parameterization ([37]): It uses $DE+EC=DC=AB=107$ and $EC+CF=EF=184$ to obtain $DE=107-x$ , $CF=184-x$ , and places $D=(0,0)$ , $E=(107-x,0)$ , $C=(107,0)$ , $F=(291-x,0)$ .
- Consistent vertical orientation and cyclic modeling: Noting $A=(0,16)$ , $D=(0,0)$ , it orients the short side downward ( $FG=17$ ) so $H=(107-x,-17)$ , $G=(291-x,-17)$ . Using the circle equation $x^2+y^2+Dx+Ey+F=0$ with $A$ and $D$ yields $F=0$ , $E=-16$ . Substituting $H$ and $G$ , subtracting the two equations gives $D=2x-398$ ; back-substitution reduces to the quadratic $x^2-398x+30576=0$ , with discriminant $398^2-4· 30576=36100=190^2$ and roots $x=104, 294$ .
- Root selection and full verification ([1], [7]): Applying $0<x<107$ filters out $x=294$ , selecting $x=104$ . The agent then verifies all constraints: $DE=107-104=3$ , $CF=184-104=80$ , $EF=184$ , $FG=17$ , and confirms that the circle $x^2+y^2-190x-16y=0$ passes through $A=(0,16)$ , $D=(0,0)$ , $H=(3,-17)$ , $G=(187,-17)$ .
Comparative Analysis. This case reveals a clear causal link between experience-guided behaviors and correctness. Experience [29] eliminates directional ambiguity and enforces the correct collinearity order, directly addressing the baseline’s misplacement of $G,H$ . Experience [37] induces a clean single-variable parameterization ( $DE=107-x$ , $CF=184-x$ ), which simplifies the cyclic constraint to a solvable quadratic. Experience [1] imposes a necessary boundedness filter ( $0<x<107$ ) to discard extraneous roots. Finally, experience [7] mandates comprehensive post-solution verification (rectangle dimensions, collinearity, concyclicity), preventing acceptance of spurious solutions.
Compared to the unassisted trajectory, the experience-informed reasoning corrects the vertical orientation, resolves ordering and parameterization inconsistencies, and installs principled validation gates. This case demonstrates the positive impact of integrating domain-specific experiences on reliability and accuracy in tool-integrated mathematical reasoning.
### A.2 Experience-Guided Web Searching
We consider a web searching task from WebWalkerQA: quantify 2024 rewards for (i) creators in the Creator Program (weekly amount), and (ii) players in the Play-2-Airdrop during Private Beta 4 (total pool and duration).
Baseline (without experiences). As summarized in Figure 7, the unassisted agent issues multiple broad searches and relies heavily on result snippets and third-party summaries, delaying clicks into authoritative pages. It eventually opens the PB4 registration post to confirm “Play-2-Airdrop Campaign: 200,000 ILV over six months,” but continues to scan snippets for the Creator Program value without opening the relevant official post. Connection errors to one official page cause the agent to revert to broad searches rather than alternative primary-source strategies (e.g., site-specific queries or adjacent official posts). The trajectory remains incomplete: it reports the Play-2-Airdrop figure but fails to confirm the Creator Program’s “100 ILV weekly” from an official source, yielding an incorrect/incomplete answer.
Enhanced (With Experiences). Refer to Figure 8. With a curated experience pool, the agent follows a disciplined pipeline: (1) prioritize official sources ([14], [4]) and open the PB4 overview and registration posts to extract the “200,000 ILV over six months” and Testnet/Mainnet allocation structure; (2) refine search terms to target formal titles ([2], [12], [1]) and open “The Illuvium Creator Program Is Getting An Upgrade!”; (3) extract the exact line “The prize pool of 100 ILV tokens awaits participating Creators each week,” and (4) synthesize both verified statements into a complete answer aligned with the question requirements ([25]). This results in the correct, fully supported output: creators earn 100 ILV weekly; players have a 200,000 ILV pool distributed over six months in PB4’s Play-2-Airdrop.
Web Searching (without experiences)
Question In 2024, how much ILV reward can creators participating in the Creator Program and players engaging in the Play-2-Airdrop during Private Beta 4 on the Illuvium platform earn?
Trajectory
[Agent]
I’ll start with a broad web search for both program names and rewards in 2024.
[Tool]
google_search(”Illuvium Creator Program Play-2-Airdrop Private Beta 4 2024 rewards ILV”)
[Agent]
From snippets, PB4 mentions a “Play-2-Airdrop campaign” and “200,000 ILV over six months.” This likely answers half the question. (continue searching rather than opening)
[Agent]
I’ll search specifically for the Creator Program rewards.
[Tool]
google_search(”Ïlluvium Creator Program2̈024 rewards ILV allocation”)
[Agent]
I see references (Binance Square, media blogs) to “100 ILV per week,” but can’t confirm from illuvium.io quickly. I’ll keep searching.
[Agent]
Avoids clicking potential official announcements; keeps scanning search result snippets.
[Agent]
I’ll fetch the PB4 registration page to confirm the Play-2-Airdrop details.
[Tool]
get_content(”https://illuvium.io/news/register-now-for-private-beta-4”)
[Agent]
Extracted: “Play-2-Airdrop Campaign: 200,000 ILV tokens over six months.” That confirms the second quantity.
[Agent]
I’ll look for the Airdrop Season details page.
[Tool]
google_search(”Ïlluvium Airdrop Season Detailsb̈log post 2024”)
[Tool]
get_content(”https://illuvium.io/news/illuvium-airdrop-season-details”)
[Agent]
Connection fails; instead of trying alternative official sources or site-specific queries, returns to broad searches.
[Agent]
More searches for “Illuvium 100 ILV per week creator rewards April 2024” and “site:illuvium.io Creator Program 100 ILV” yield sparse official confirmation; I’ll defer the creator number.
[Agent]
Fails to open likely official Creator Program pages; leaves the creator number unconfirmed.
[Agent]
Synthesis: I will answer with what’s confirmed officially (Play-2-Airdrop 200,000 ILV over 6 months) and note the creator reward as unclear.
Outcome Creators: (not clearly specified; mentions unverified “100 ILV per week”). Players (PB4 Play-2-Airdrop): 200,000 ILV over six months.
Figure 7: Example trajectory without experiences in Web Searching.
Web Searching (with experiences)
Question In 2024, how much ILV reward can creators participating in the Creator Program and players engaging in the Play-2-Airdrop during Private Beta 4 on the Illuvium platform earn? Experiences [1] Document identification: Focus on extracting formal titles and collection names from official metadata and headers rather than inferring relationships from content descriptions. $⋯$ [2] Systematic exploration: Progress from broad initial queries to targeted content extraction using adaptive search strategies that balance discovery with precision. $⋯$ [4] Source prioritization: Prioritize systematic extraction from authoritative comprehensive documents over fragmented information for coherent topic coverage. $⋯$ [12] Iterative refinement: Continuously refine search terms based on emerging patterns while periodically re-evaluating previously encountered information to avoid overlooking correct solutions and maintain essential contextual elements. $⋯$ [14] Primary source prioritization: Prioritize locating official primary source documents over secondary coverage to ensure accuracy and capture exact language for definitive metadata and hierarchical relationships. [25] Partial match evaluation: Systematically evaluate partial matches from initial searches against all criteria before pursuing alternative paths. $⋯$
Trajectory
[Agent]
Start with broad search, then pivot to official illuvium.io pages for PB4 and Creator Program. (By [2], [14], [4])
[Tool]
google_search(”Illuvium Creator Program rewards 2024 Private Beta 4 ILV earnings”, num_results=5)
[Agent]
Identify official pages in results: “Illuvium Beta 4: All You Need to Know” and “Register Now for Private Beta 4”. Open them to confirm Play-2-Airdrop details. (By [14], [25])
[Tool]
get_content(”https://illuvium.io/news/illuvium-beta-4-all-you-need-to-know”)
[Agent]
Extracted from official PB4 overview: Testnet phases get 10% of the Play-2-Airdrop pool; clarifies structure and timeline.
[Tool]
get_content(”https://illuvium.io/news/register-now-for-private-beta-4”)
[Agent]
Extracted explicit official number: “Play-2-Airdrop Campaign: 200,000 ILV tokens” over six months. This covers the player reward for PB4 context.
[Agent]
Iteratively refine search for the Creator Program exact weekly reward using formal title cues like “Creator Program … upgrade”. (key behavior; By [12], [1])
[Tool]
google_search(”Ïlluvium Creator Programr̈ewards ILV 2024”, num_results=5)
[Agent]
Click the official post: “The Illuvium Creator Program Is Getting An Upgrade!”. (By [14], [1], [4])
[Tool]
get_content(”https://illuvium.io/news/illuvium-creator-program-is-getting-an-upgrade”)
[Agent]
Extracted exact statement from official source: “The prize pool of 100 ILV tokens awaits participating Creators each week.” This provides the creator reward.
Outcome Creators can earn 100 ILV weekly in the Creator Reward Program, and 200,000 ILV will be distributed over six months for players in the Play-2-Airdrop during Private Beta 4.
Figure 8: Example trajectory with experiences in Web Searching.
Comparative Analysis. Experience-guided behaviors directly address baseline deficiencies: primary source prioritization ([14], [4]) removes reliance on snippets and third-party coverage; document identification ([1]) and iterative refinement ([2], [12]) ensure the agent locates and opens the exact Creator Program post; partial match evaluation ([25]) steers the agent to confirm numerical claims at their authoritative origin. In contrast, the baseline wastes context on searches without content acquisition, leaves critical values unverified, and produces an incomplete answer.
## Appendix B Prompts
In this appendix, we provide the prompts used in math reasoning tasks. Figures 9 and 10 present the prompts for solving math tasks. Figures 11 and 12 are used in group relative semantic advantage. Figure 13 works for optimizing the experiential knowledge $E$ .
System Prompt
Solve the following problem step by step. You now have the ability to selectively write executable Python code to enhance your reasoning process, e.g., calulating numbers and verifying math computations. Never directly just printing your semantic reasoning in Python. The Python code will be executed by an external sandbox, and the output (returned as a dict with the message in the ”message” field) can be returned to aid your reasoning and help you arrive at the final answer. The Python code should be complete scripts, including necessary imports. Each code snippet is wrapped with ```python code snippet ```. The last part of your final response should be in the following format: $<$ answer $>$ $\backslash$ boxed{The final answer goes here.} $<$ /answer $>$
Figure 9: System prompt for ReAct framework in Math Reasoning.
Prompt for Involving Experiential Knowledge $E$
Please solve the problem: {problem} When solving problems, you MUST first carefully read and understand the helpful instructions and experiences: {experiences}
Figure 10: Prompt for supplementing math problems with experiential knowledge $E$ .
Prompt for Trajectory Summarization in Group Advantage Computation
An agent system may be provided with some experiences, and then it produces the following trajectory to solve the given problem. Please summarize the trajectory step-by-step: 1. For each step, describe what action is being taken, and which experience has been used in this step. 2. Given the grading of this rollout and the correct answer, identify and explain any steps that represent detours, errors, or backtracking, highlighting why they might have occurred and what their impact was on the trajectory’s progress. 3. Maintain all the core outcome of each step, even if it was part of a flawed process. $<$ trajectory $>$ {trajectory} $<$ /trajectory $>$ $<$ evaluation $>$ {whether the answer is correct or not} $<$ /evaluation $>$ $<$ groundtruth $>$ {the ground truth answer} $<$ /groundtruth $>$ Only return the trajectory summary of each step, e.g., 1. what happened in the first step and the core outcomes 2. what happened in the second step and the core outcomes 3. …
Figure 11: Prompt for summarizing each trajectory in Math Reasoning.
Prompt for Group Advantage Computation
An agent system is provided with a set of experiences and has tried to solve the problem multiple times with both successful and wrong solutions. Review these problem-solving attempt and extract generalizable experiences. Follow these steps: 1. Trajectory Analysis: $-$ For successful steps: Identify key correct decisions and insights $-$ For errors: Pinpoint where and why the reasoning went wrong $-$ Note any important patterns or strategies used/missed $-$ Review why some trajectories fail? Is there any existing experiences are missed, or experiences do not provide enough guidance? 2. Update Existing Experiences $-$ Some trajectories may be correct and others may be wrong, you should ensure there are experiences can help to run correctly $-$ You have three options: [modify, add, delete] $*$ modify: You can modify current experiences to make it helpful $*$ add: You can introduce new experiences to improve future performance $*$ delete: You can delete existing experiences $-$ You can update at most {max number of operations} clear, generalizable lessons for this case $-$ Before updating each experience, you need to: $*$ Specify when it would be most relevant $*$ List key problem features that make this experience applicable $*$ Identify similar problem patterns where this advice applies 3. Requirements for each experience that is modified or added. $-$ Begin with general background with several words in the experience $-$ Focus on strategic thinking patterns, not specific calculations $-$ Emphasize decision points that could apply to similar problems Please provide reasoning in details under the guidance of the above 3 steps. After the step-by-step reasoning, you will finish by returning in this JSON format as follows: ```json $[$ { ”option”: ”modify”, ”experience”: ”the modified experience”, ”modified_from”: ”G17” # specify the ID of experience that is modified }, { ”option”: ”add”, ”experience”: ”the added experience”, }, { ”option”: ”delete”, ”delete_id”: ”the deleted experience ID”, }, … $]$ ``` Note that your updated experiences may not need to cover all the options. You can only use one type of updates or choose to remain all experiences unchanged. $<$ problem $>$ {problem} $<$ /problem $>$ $<$ trajectories $>$ { $G$ summarized trajectories in the same group} $<$ /trajectories $>$ $<$ groundtruth $>$ {answer} $<$ /groundtruth $>$ $<$ experience $>$ {experiences} $<$ /experience $>$
Figure 12: Prompt for group advantage computation in Math Reasoning.
Prompt for the Optimization Step
An agent system is provided with a set of experiences and has tried to solve the problem multiple times. From the reflections, some suggestions on the existing experiences have been posed. Your task is to collect and think for the final experience revision plan. Each final experience must satisfy the following requirements 1. It must be clear, generalizable lessons for this case, with no more than 32 words 2. Begin with general background with several words in the experience 3. Focus on strategic thinking patterns, not specific calculations 4. Emphasize decision points that could apply to similar problems 5. Avoid repeating saying similar experience in multiple different experiences $<$ experience $>$ {experiences} $<$ /experience $>$ $<$ suggested_updates $>$ {all group advantage in current batch} $<$ /suggested_updates $>$ Please provide reasoning in each of the suggestions, and think for how to update existing experiences You have three update options: [modify, merge, delete] - modify: You can modify current experiences to make it helpful - merge: You can merge some similar experiences into a more general forms to reduce duplication - delete: You can delete an experience After generating the step-by-step reasoning, you need to give the final experience revision details by returning in this JSON format as follows: ```json $[$ { ”option”: ”modify”, ”experience”: ”the modified experience”, ”modified_from”: ”G17” # specify the ID of experience that is modified }, { ”option”: ”merge”, ”experience”: ”the merged experience”, ”merged_from”: [”C1”, ”C3”, ”S4”, …] # specify the str IDs of experiences that is merged from, at least 2 IDs are needed }, { ”option”: ”delete”, ”delete_id”: ”the deleted experience ID”, }, … $]$ ``` Note that your updated experiences may not need to cover all the options. You can only use one type of updates or choose to remain all experiences unchanged.
Figure 13: Prompt for optimizing experiential knowledge $E$ based on group advantages in the same batch in Math Reasoning.
## Appendix C Examples of Learned Experiences
In this appendix, we provide some examples in Figure 14, which are extracted from 48 learned experiences by Training-Free GRPO with tool use in Math Reasoning.
Examples of Learned Experiences
[1]. When solving geometry problems with intersections, validate solutions lie within bounded regions or segments, not on extensions, to avoid extraneous answers. [2]. When multiple locations are described relative to a common point using compass directions (e.g., southwest, southeast), analyze the angle between these directions. Southwest and southeast from the same point form a 90° angle, suggesting a right triangle configuration. This insight often allows using the Pythagorean theorem rather than more complex coordinate geometry approaches. [3]. When scoring systems reward consecutive successes, analyze whether bonuses are additive or replace base scores. Test small examples to resolve ambiguity. $⋯$ [10]. When using mathematical invariants to prove impossibility, always validate them against known achievable states or small cases. If an invariant suggests a state is unreachable but you can construct a valid sequence reaching it, re-examine your invariant analysis - you may have made an error in the mathematical reasoning or the invariant may not be preserved as claimed. [11]. When solving ’smallest prime’ problems involving divisibility of large expressions like geometric series, first handle primes 2 and 3 as special cases. Then systematically check primes in ascending order using multiplicative order properties: for prime p ¿ 3, p divides the sum if the multiplicative order of the base modulo p divides the exponent in the closed form. This avoids unnecessary direct computation of huge numbers. [12]. For expected extreme statistics in combinatorial problems, use direct enumeration for small sizes. For larger problems, develop combinatorial counting methods instead of expectation formulas. $⋯$ [21]. For complex polynomials with real parameters, separate real and imaginary parts to find when real roots exist or all are non-real. [22]. For right triangles with symmetric points on a circle, assume equal y-coordinates and opposite x to simplify area maximization. $⋯$ [32]. For points on concentric circles, collinearity along a ray minimizes the circumradius to half the max-min radius difference. [33]. In state transition problems with additive operations, consider binary representation invariants. The number of 1’s often determines extremal states or transition possibilities. $⋯$
Figure 14: Example of learned experiences from Training-Free GRPO with tool use in Math Reasoning.