# Training-Free Group Relative Policy Optimization
tristanli@tencent.com https://github.com/TencentCloudADP/youtu-agent/tree/training_free_GRPO
(October 9, 2025)
## Abstract
Recent advances in Large Language Model (LLM) agents have demonstrated their promising general capabilities. However, their performance in specialized real-world domains often degrades due to challenges in effectively integrating external tools and specific prompting strategies. While methods like agentic reinforcement learning have been proposed to address this, they typically rely on costly parameter updates, for example, through a process that uses Supervised Fine-Tuning (SFT) followed by a Reinforcement Learning (RL) phase with Group Relative Policy Optimization (GRPO) to alter the output distribution. However, we argue that LLMs can achieve a similar effect on the output distribution by learning experiential knowledge as a token prior, which is a far more lightweight approach that not only addresses practical data scarcity but also avoids the common issue of overfitting. To this end, we propose Training-Free Group Relative Policy Optimization (Training-Free GRPO), a cost-effective solution that enhances LLM agent performance without any parameter updates. Our method leverages the group relative semantic advantage instead of numerical ones within each group of rollouts, iteratively distilling high-quality experiential knowledge during multi-epoch learning on a minimal ground-truth data. Such knowledge serves as the learned token prior, which is seamlessly integrated during LLM API calls to guide model behavior. Experiments on mathematical reasoning and web searching tasks demonstrate that Training-Free GRPO, when applied to DeepSeek-V3.1-Terminus, significantly improves out-of-domain performance. With just a few dozen training samples, Training-Free GRPO outperforms fine-tuned small LLMs with marginal training data and cost.
<details>
<summary>figures/spotlight.png Details</summary>

### Visual Description
## Bar Charts: AIME Benchmarks, Training Cost, and Training Data
### Overview
The image presents three sets of bar charts comparing performance metrics for a model ("Ours") against a baseline ("W/o tool") and a baseline with an additional component ("W/o tool + Ours"). The charts cover AIME24 and AIME25 benchmarks (Mean@32), Training Cost, and Training Data (# Samples). Arrows indicate relative changes between data points.
### Components/Axes
* **AIME Benchmarks:**
* X-axis: Categories - "W/o tool", "W/o tool + Ours", "With tool", "With tool + Ours" (repeated for AIME24 and AIME25).
* Y-axis: "Mean@32 (%)", ranging from 40 to 90.
* **Training Cost:**
* X-axis: Categories - "RL Train (32B)", "Ours (671B)".
* Y-axis: "Cost ($)", ranging from 0 to 10000.
* **Training Data:**
* X-axis: Categories - "RL Train (32B)", "Ours (671B)".
* Y-axis: "# Samples", ranging from 0 to 17500.
* **Arrows:** Downward arrows indicate a decrease in value, while upward arrows indicate an increase.
* **Circles:** Highlight specific values with approximate numbers inside.
### Detailed Analysis or Content Details
**AIME Benchmarks:**
* **AIME24:**
* "W/o tool": Approximately 68.6%.
* "W/o tool + Ours": Approximately 72.6%.
* "With tool": Approximately 80%.
* "With tool + Ours": Approximately 82.7%.
* Trend: Performance increases consistently with the addition of "tool" and "Ours".
* **AIME25:**
* "W/o tool": Approximately 52.9%.
* "W/o tool + Ours": Approximately 54%.
* "With tool": Approximately 67.9%.
* "With tool + Ours": Approximately 73.3%.
* Trend: Similar to AIME24, performance increases with the addition of "tool" and "Ours".
**Training Cost:**
* "RL Train (32B)": Approximately 10000$.
* "Ours (671B)": Approximately 18$.
* Trend: A significant decrease in cost when using "Ours".
**Training Data:**
* "RL Train (32B)": Approximately 17000 samples.
* "Ours (671B)": Approximately 100 samples.
* Trend: A significant decrease in the number of samples required when using "Ours".
### Key Observations
* The addition of both "tool" and "Ours" consistently improves performance on both AIME benchmarks.
* Using "Ours" dramatically reduces both training cost and the amount of training data needed.
* The cost reduction is particularly striking, with "Ours" costing approximately 18$ compared to 10000$ for "RL Train (32B)".
* The reduction in training data is also substantial, with "Ours" requiring only 100 samples compared to 17000 for "RL Train (32B)".
### Interpretation
The data suggests that the "Ours" model, potentially in conjunction with a "tool", offers a significant improvement in performance on the AIME benchmarks while simultaneously reducing training costs and data requirements. This indicates a more efficient and effective learning process. The large difference in cost and data suggests that "Ours" may leverage techniques like transfer learning, data augmentation, or more efficient optimization algorithms. The consistent performance gains across both AIME24 and AIME25 suggest the benefits are not specific to a particular benchmark. The downward arrows visually emphasize the positive impact of "Ours" on cost and data usage, reinforcing the message of efficiency. The use of "B" in the model sizes (32B, 671B) likely refers to the number of parameters in the model, indicating that "Ours" is a much larger model, but its efficiency mitigates the expected increase in training cost.
</details>
Figure 1: Applying Training-Free GRPO on both prompting (without tools) and ReAct [1] (with tools) achieve improved Mean@32 on AIME benchmarks [2] with DeepSeek-V3.1-Terminus [3]. It consumes significantly fewer training data and lower costs on the 671B LLM than fine-tuning a 32B model [4], serving as a cost-effective alternative to RL methods. footnotetext: Full author list in contributions.
## 1 Introduction
Large Language Models (LLMs) are emerging as powerful general-purpose agents capable of interacting with complex, real-world environments. They have shown remarkable capabilities across a wide range of tasks, including complex problem-solving [4, 5, 6], advanced web research [7, 8, 9, 10], code generation and debugging [11, 12], and proficient computer use [13, 14, 15]. Despite their impressive capabilities, LLM agents often underperform in specialized, real-world domains. These scenarios typically demand the integration of external tools (e.g., calculators, APIs, databases), along with domain-specific task definitions and prompting strategies. Deploying a general-purpose agent out-of-the-box in such settings often results in suboptimal performance due to limited familiarity with domain-specific requirements or insufficient exposure to necessary tools.
To bridge this gap, agentic training has emerged as a promising strategy to facilitate the adaptation of LLM agents to specific domains and their associated tools [4, 7, 8, 16]. Recent advancements in agentic reinforcement learning (Agentic RL) approaches have employed Group Relative Policy Optimization (GRPO) [17] and its variants [18, 19, 20] to align model behaviors in the parameter space. Although these methods effectively enhance task-specific capabilities, their reliance on tuning LLM parameters poses several practical challenges:
- Computational Cost: Even for smaller models, fine-tuning demands substantial computational resources, making it both costly and environmentally unsustainable. For larger models, the costs become prohibitive. Furthermore, fine-tuned models require dedicated deployment and are often limited to specific applications, rendering them inefficient for low-frequency use cases compared to more versatile general-purpose models.
- Poor Generalization: Models optimized via parameter tuning often suffer from unsatisfactory cross-domain generalization, limiting their applicability to narrow tasks. Consequently, multiple specialized models must be deployed to handle a comprehensive set of tasks, significantly increasing system complexity and maintenance overhead.
- Data Scarcity: Fine-tuning LLMs typically necessitates large volumes of high-quality, carefully annotated data, which are often scarce and prohibitively expensive to obtain in specialized domains. Additionally, with limited samples, models are highly susceptible to overfitting, leading to poor generalization.
- Diminishing Returns: The prohibitive training costs usually compel existing approaches to fine-tune smaller LLMs with fewer than 32 billion parameters, due to resource constraints rather than optimal design choices. While larger models would be preferred, the computational expense of fine-tuning necessitates this compromise. Paradoxically, API-based or open-source larger LLMs often deliver better cost-performance ratios through scalability and continuous model updates. However, these general-purpose models underperform in specialized domains where fine-tuning is necessary, creating a cost-performance dilemma.
Such limitations inherent in parameter tuning motivate a fundamental research question: Is applying RL in parametric space the only viable approach? Can we enhance LLM agent performance in a non-parametric way with lower data and computational costs?
We answer this question affirmatively by proposing Training-Free Group Relative Policy Optimization (Training-Free GRPO), a novel and efficient method that improves LLM agent behavior in a manner similar to vanilla GRPO, while preserving the original model parameters unchanged. Our approach is motivated by the insight that LLMs already possess the fundamental capability to adapt to new scenarios, requiring only minimal practice through limited samples to achieve strong performance. Thus, instead of adapting their output distribution through parameter tuning, in-context learning [21] that leverages a lightweight token prior can also encapsulate experiential knowledge learned from a minimal training dataset.
Training-Free GRPO retains the multi-epoch learning mechanism of vanilla GRPO. In each epoch, multiple outputs are generated to deliver a group of rollouts for every query, which helps to explore the policy space and evaluate potential strategies. While vanilla GRPO relies on gradient-based parameter updates to iteratively improve policy performance, Training-Free GRPO eliminates this requirement by employing inference-only operations using LLMs. At each optimization step, rather than calculating a numerical advantage for gradient ascent within each group of rollouts, our method leverages LLMs to introspect on each group and distill a semantic advantage. Such advantage refines external experiential knowledge and guide policy outputs based on evolving contextual priors, thereby achieving policy optimization effects without modifying any model parameters.
By evaluating challenging mathematical reasoning and interactive web searching tasks, we demonstrate that our method significantly enhances the performance of frozen policy models such as DeepSeek-V3.1-Terminus [3] with only dozens of training samples. It surpasses fine-tuned 32B models in performance while requiring only a fraction of the computational resources, offering a simple and much more efficient alternative to traditional fine-tuning techniques.
Our principal contributions are threefold:
- A New Training-Free RL Paradigm: We introduce Training-Free GRPO, which shifts policy optimization from the parameter space to the context space by leveraging evolving experiential knowledge as token priors without gradient updates.
- Semantic Group Advantage: We replace numerical group advantage in vanilla GRPO with semantic group advantage, enabling LLMs to introspect their own rollouts and continuously updating experiential knowledge at multiple optimization steps.
- Data and Computational Efficiency: Experiments confirm that Training-Free GRPO effectively enhances the performance of a frozen policy with minimal training samples, offering a practical and cost-effective alternative across different domains.
- Superior Generalization: By leaving model parameters frozen and plugging in different token priors, our approach fully preserves the generalization power, eliminating the cost and complexity of deploying multiple fine-tuned specialists.
## 2 Training-Free GRPO
<details>
<summary>figures/training-free_GRPO.png Details</summary>

### Visual Description
## Diagram: Group Reciprocal Proximal Optimization (GRPO) Architectures
### Overview
The image presents a comparative diagram illustrating two architectures of Group Reciprocal Proximal Optimization (GRPO): (a) Vanilla GRPO and (b) Training-free GRPO. Both architectures share a common goal of updating a policy model based on feedback from reference and reward models, culminating in group computation and action selection. The diagram highlights the differences in how experiences are handled and utilized within each approach.
### Components/Axes
The diagram consists of several key components:
* **Policy Model:** Represented by a brain-like icon, this model is updated in both architectures.
* **Reference Model:** Present in both architectures, providing outputs (O1, O2, ⊠OG).
* **Reward Model:** Present in both architectures, generating rewards (r1, r2, ⊠rG).
* **Group Computation:** A central block in both architectures, processing outputs and rewards.
* **Actions (A1, A2, ⊠AG):** The final output of the system, representing the selected actions.
* **Experiences:** A collection of data points, only present in the Training-free GRPO architecture.
* **LLM (Large Language Model):** Used in the Training-free GRPO to summarize and extract experience.
* **Controller:** A component in the Training-free GRPO, allowing for actions like adding, deleting, keeping, or modifying elements.
* **Update Arrows:** Yellow arrows indicating the flow of updates to the policy model.
* **Data Flow Arrows:** Black arrows representing the flow of data between components.
### Detailed Analysis or Content Details
**(a) Vanilla GRPO:**
* A square box labeled "q" is positioned to the left of the policy model.
* The policy model receives an update signal (yellow arrow) from the reference and reward models.
* The reference model outputs O1, O2, and up to OG.
* The reward model outputs r1, r2, and up to rG.
* The Group Computation block processes these outputs and rewards, resulting in actions A1, A2, and up to AG.
**(b) Training-free GRPO:**
* A collection of experiences (represented by brain icons) is positioned to the left of the policy model.
* The policy model receives an update signal (yellow arrow).
* The reference model outputs Xr1, Xr2, and up to XrG.
* The reward model outputs r1, r2, and up to rG.
* An LLM is used to summarize the experiences, generating s1, s2, and up to sG.
* Another LLM is used to extract experience, generating Atext1, Atext2, and up to AtextG.
* A controller is present, with options to "ADD", "DELETE", "KEEP", and "MODIFY".
* The Group Computation block processes the outputs from the reward model and LLM, resulting in actions Atext1, Atext2, and up to AtextG.
### Key Observations
* The Training-free GRPO architecture incorporates an LLM and a controller, suggesting a more dynamic and adaptable approach compared to the Vanilla GRPO.
* The Vanilla GRPO relies directly on the outputs of the reference and reward models, while the Training-free GRPO utilizes an LLM to process experiences and generate summaries.
* The use of "Xr" in the Training-free GRPO suggests a transformation or modification of the reference model outputs.
* The actions in the Training-free GRPO are labeled "Atext", indicating they might be textual or represent a different type of output compared to the Vanilla GRPO.
### Interpretation
The diagram illustrates two distinct approaches to GRPO. The Vanilla GRPO represents a more traditional, model-based approach, where the policy model is directly updated based on the outputs of predefined reference and reward models. The Training-free GRPO, on the other hand, introduces a layer of abstraction through the use of an LLM and a controller. This allows the system to learn from experiences, adapt its behavior, and potentially generate more nuanced or context-aware actions. The inclusion of a controller suggests the ability to actively manage and refine the learning process. The "Xr" notation implies a potential pre-processing step applied to the reference model outputs before they are used in the group computation. The difference in action labeling ("A" vs. "Atext") hints at a potential shift in the nature of the actions generated by each architecture. The diagram suggests a move towards more flexible and adaptable GRPO systems that can leverage the power of LLMs to learn from experience and generate more sophisticated behaviors.
</details>
Figure 2: Comparison of vanilla GRPO and Training-Free GRPO.
<details>
<summary>figures/case.png Details</summary>

### Visual Description
\n
## Diagram: Geometry Problem Solving with Agent Rollout
### Overview
This diagram illustrates a problem-solving process involving an agent attempting to find the intersection point of a line with a triangle, and the subsequent summarization of the agent's steps and advantages gained through group computation. The diagram is divided into four main columns: "Question", "Rollout ... G", "Summarization", and "Advantage". Each column represents a stage in the problem-solving process, with multiple "trajectories" (1, 2, and 3) shown within the "Rollout" column.
### Components/Axes
The diagram doesn't have traditional axes. Instead, it uses a flow-based layout with text boxes and arrows to indicate the sequence of steps. Key components include:
* **Question:** Defines the geometric problem.
* **Rollout (Trajectory 1, 2, 3):** Shows the agent's thought process, including questions, code execution, and verification steps.
* **Summarization (Summarization 1, 2, 3):** Provides a concise summary of the agent's steps.
* **Advantage:** Highlights the benefits of group computation in identifying and correcting errors.
* **Triangle ABC:** A geometric figure with vertices A(0, 8), B(2, 0), C(8, 0).
* **Line through B:** A line cutting triangle ABC in half.
* **Line D:** A line defined by the equation 2x + y = 4.
### Detailed Analysis or Content Details
**Question Column:**
* "Triangle ABC has vertices A(0, 8), B(2, 0), C(8, 0). A line through B cuts the area of triangle ABC in half: find the sum of the slope and y-intercept of this line."
**Rollout - Trajectory 1:**
* **Step 1:** "Let me find the intersection point D on AC." (Question mark icon)
* **Step 2:** "executing code..." (Code execution icon)
* **Step 3:** "When m = -2.0, the area of triangle ABD is half the area."
* **Step 4:** "verifying the area of triangle ABD..."
* **Step 5:** "I found the solution!" (Checkmark icon)
* Diagram shows triangle ABC with point D on AC.
**Rollout - Trajectory 2:**
* **Step 1:** "For m = 2, let me verify this by checking if the line indeed divides the triangle into two equal areas." (Question mark icon)
* **Step 2:** "executing code..." (Code execution icon)
* **Step 3:** "The point D(-4,12) is actually on the line segment AC, it's an extension of it..."
* **Step 4:** "executing code..." (Code execution icon)
* **Step 5:** "The line y = 2x - 4 through point B indeed divides triangle ABC into two equal areas of 12 each."
* Diagram shows triangle ABC with point D extending beyond AC.
**Rollout - Trajectory 3:**
* Diagram shows triangle ABC with point D extending beyond AC.
**Summarization - Summarization 1:**
* **Step 1:** "The agent set up equations to find the intersection point D."
* **Step 2:** "The agent tested various slopes numerically and found that when m = -2, the area of triangle ABD equals exactly half the total area."
* **Step 3:** "The agent verified the solution divides the area in half."
**Summarization - Summarization 2:**
* **Step 1:** "The agent tests a specific slope value m = 2 to find the intersection point D."
* **Step 2:** "The agent attempts to verify the solution by calculating both areas (ABD and BCD)."
* **Step 3:** "The agent calculates the area of triangle BCD and discovers it's 36. It realizes that point D is an extension of segment AC, making the geometry invalid."
* **Step 4:** "The agent properly sets up the area equation constraint and find the intersection point."
**Summarization - Summarization 3:**
* (Content is not fully visible, appears to be a continuation of the summarization process)
**Advantage:**
* "advantage compute decision"
* "In attempt 1, the agent failed to consider the physical constraints of the triangle, the found intersection lies outside segment AC."
* "In attempt 2, the agent recognized the error in step 3 and completely changed approach, properly setting up both the area and geometric constraints simultaneously."
* "When solving geometry problems involving intersections with bounded regions, always validate that mathematical solutions satisfy geometric constraints."
### Key Observations
* The agent initially makes an error in Trajectory 1 by not considering the geometric constraints of the problem, leading to an intersection point outside the triangle.
* The agent corrects this error in Trajectory 2 by recognizing that the intersection point lies on the extension of the line segment AC.
* Group computation is highlighted as a valuable tool for identifying and correcting errors in the problem-solving process.
* The diagram emphasizes the importance of validating mathematical solutions with geometric constraints.
### Interpretation
The diagram demonstrates a problem-solving approach using an agent that iteratively refines its solution through a process of questioning, code execution, and verification. The "Rollout" column illustrates the agent's thought process, while the "Summarization" column provides a concise overview of the steps taken. The "Advantage" column highlights the benefits of group computation in identifying and correcting errors.
The diagram suggests that successful problem-solving requires not only mathematical accuracy but also a consideration of real-world constraints. The agent's initial error in Trajectory 1 underscores the importance of validating solutions with geometric constraints. The correction in Trajectory 2 demonstrates the agent's ability to learn from its mistakes and adapt its approach.
The diagram also highlights the value of collaboration and peer review in problem-solving. The "Advantage" column suggests that group computation can help to identify errors that might be missed by an individual agent. This is particularly important in complex problems where multiple factors need to be considered. The diagram is a visual representation of the iterative nature of problem-solving and the importance of both mathematical rigor and geometric intuition.
</details>
Figure 3: Example of a Training-Free GRPO learning step.
In this section, we introduce our Training-Free GRPO, a method designed to replicate the alignment benefits of the GRPO algorithm without performing any gradient-based updates to the policy modelâs parameters.
Vanilla GRPO. As shown in Figure 2, the vanilla GRPO procedure operates by first generating a group of $G$ outputs $\{o_{1},o_{2},\dots,o_{G}\}$ for a given query $q$ using the current policy LLM $\pi_{\theta}$ , i.e., $\pi_{\theta}(o_{i}\mid q)$ . Each output $o_{i}$ is then independently scored with a reward model $\mathcal{R}$ . Subsequently, with rewards $\mathbf{r}=\{r_{1},\dots,r_{G}\}$ , it calculates a group-relative advantage $\hat{A}_{i}=\frac{r_{i}-\text{mean}(\mathbf{r})}{\text{std}(\mathbf{r})}$ for each output $o_{i}$ . By combining a KL-divergence penalty against a reference model, it constructs a PPO-clipped objective function $\mathcal{J}_{\text{GRPO}}(\theta)$ , which is then maximized to update the LLM parameters $\theta$ .
Training-Free GRPO repurposes the core logic of this group-based relative evaluation but translates it into a non-parametric, inference-time process. Instead of updating the parameters $\theta$ , we leave $\theta$ permanently frozen and maintain an external experiential knowledge $\mathcal{E}$ , which is initialized to $\emptyset$ .
Rollout and Reward. As shown in Figure 2, our rollout and reward process mirrors that of GRPO exactly. Given a query $q$ , we perform a parallel rollout to generate a group of $G$ outputs $\{o_{1},o_{2},\dots,o_{G}\}$ using the LLM. Notably, while GRPO uses the current trainable policy $\pi_{\theta}$ , our policy conditions on the experiential knowledge, $\pi_{\theta}(o_{i}|q,\mathcal{E})$ . Identical to the standard GRPO setup, we score each output $o_{i}$ by the reward model $\mathcal{R}$ to obtain a scalar reward $r_{i}=\mathcal{R}(q,o_{i})$ .
Group Advantage Computation. To provide an optimization direction for policy parameters, vanilla GRPO computes a numerical advantage $\hat{A}_{i}$ that quantifies each output $o_{i}$ âs relative quality within its group. Similarly, Training-Free GRPO performs an analogous comparison within each group but produces a group relative semantic advantage in the form of natural language experience, as shown in Figure 3. Since $\hat{A}_{i}=0$ when all $G$ outputs receive identical rewards (i.e., $\text{std}(\mathbf{r})=0$ ) in vanilla GRPO, we generate such semantic advantages only for groups with both clear winners and losers. Specifically, for each output $o_{i}$ , we first ask the same LLM $\mathcal{M}$ to provide a corresponding summary $s_{i}=\mathcal{M}(p_{\text{summary}},q,o_{i})$ separately, where $p_{\text{summary}}$ is a prompt template that incorporates the query $q$ and output $o_{i}$ to form a structured summarization request. Given the summaries $\{s_{1},s_{2},\dots,s_{G}\}$ and the current experiential knowledge $\mathcal{E}$ , the LLM $\mathcal{M}$ articulates the reasons for the relative success or failure of the outputs, followed by extracting a concise natural language experience $A_{\text{text}}=\mathcal{M}({p_{\text{extract}}},q,s_{i},\mathcal{E})$ , where $p_{\text{extract}}$ is another prompt template for experience extraction. This natural language experience $A_{\text{text}}$ serves as our semantic advantage, functionally equivalent to vanilla GRPOâs $\hat{A}_{i}$ , encoding the critical experiential knowledge of what actions lead to high rewards.
Optimization. Whereas vanilla GRPO updates its model parameters $\theta$ via gradient ascent on $\mathcal{J}_{\text{GRPO}}(\theta)$ computed by all advantages in a single batch, we update our experience library $\mathcal{E}$ using all semantic advantages $A_{\text{text}}$ from the current batch. Specifically, given the existing experiences library $\mathcal{E}$ , we prompt the LLM to generate a list of operations based on all these $A_{\text{text}}$ , where each operation could be:
- Add: Directly append the experience described in $A_{\text{text}}$ to the experience library $\mathcal{E}$ .
- Delete: Based on $A_{\text{text}}$ , remove a low-quality experience from the experience library $\mathcal{E}$ .
- Modify: Refine or improve an existing experience in the experience library $\mathcal{E}$ based on insights from $A_{\text{text}}$ .
- Keep: The experience library $\mathcal{E}$ remains unchanged.
After updating the experience library $\mathcal{E}$ , the conditioned policy $\pi_{\theta}(y|q,\mathcal{E})$ produces a shifted output distribution in subsequent batches or epochs. This mirrors the effect of a GRPO policy update by steering the model towards higher-reward outputs, but achieves this by altering the context rather than the modelâs fundamental parameters. The frozen base model $\pi_{\theta}$ acts as a strong prior, ensuring output coherence and providing a built-in stability analogous to the KL-divergence constraint in GRPO that prevents the policy from deviating excessively from $\pi_{\text{ref}}$ .
## 3 Evalution
To compare Training-Free GRPO with competitive baselines, we conduct comprehensive experiments on both mathematical reasoning and web searching benchmarks.
### 3.1 Mathematical Reasoning
Benchmarks. We conduct our evaluation on the challenging AIME24 and AIME25 benchmarks [2], which are representative of complex, out-of-domain mathematical reasoning challenges. To ensure robust and statistically reliable results, we evaluate each problem with 32 independent runs and report the average Pass@1 score, which we denote as Mean@32.
Setup. We primarily focus on large powerful LLMs that are usually hard and expensive to be fine-tuned in real-world applications, such as DeepSeek-V3.1-Terminus [3]. We include two basic configurations: (1) Direct Prompting without tool use (a text-only input/output process), and (2) ReAct [1] with a code interpreter (CI) tool. To apply Training-Free GRPO experiments, we randomly sample 100 problems from the DAPO-Math-17K dataset [19], denoted as DAPO-100. We run the learning process for 3 epochs with a single batch per epoch (i.e., 3 steps), using a temperature of $0.7$ and a group size of $5$ during the learning phase. For out-of-domain evaluation on AIME 2024 and 2025 benchmarks, we use a temperature of $0.3$ .
Table 1: Mean@32 on AIME 2024 and AIME 2025 benchmarks (%).
| Direct | - | - | DeepSeek-V3.1-Terminus | - | 68.6 | 52.9 |
| --- | --- | --- | --- | --- | --- | --- |
| + Training-Free GRPO | $\approx\mathdollar 8$ | DAPO-100 | DeepSeek-V3.1-Terminus | - | 72.6 ( $\uparrow$ 4.0) | 54.0 ( $\uparrow$ 1.1) |
| ReAct [1] | - | - | DeepSeek-V3.1-Terminus | CI | 80.0 | 67.9 |
| + Training-Free GRPO | $\approx\mathdollar 18$ | DAPO-100 | DeepSeek-V3.1-Terminus | CI | 82.7 ( $\uparrow$ 2.7) | 73.3 ( $\uparrow$ 5.4) |
| ReAct [1] | - | - | DeepSeek-V3.2-Exp | CI | 71.0 | 61.8 |
| + Training-Free GRPO | $\approx\mathdollar 8$ | DAPO-100 | DeepSeek-V3.2-Exp | CI | 73.1 ( $\uparrow$ 2.1) | 63.2 ( $\uparrow$ 1.4) |
<details>
<summary>x1.png Details</summary>

### Visual Description
\n
## Bar Charts: Performance and Tool Calls vs. Learning Step
### Overview
The image presents two sets of bar charts, side-by-side. The left chart displays "Performance" metrics, while the right chart shows the "# Tool Calls". Both charts compare performance across different "Learning Steps" (1, 2, 3, and ReAct) and for different models: "Train", "AIME24", and "AIME25". The charts use different colors to represent each model's performance.
### Components/Axes
* **X-axis (Both Charts):** "Learning Step" with categories: 1, 2, 3, ReAct.
* **Y-axis (Left Chart):** "Performance" ranging from approximately 60 to 90.
* **Y-axis (Right Chart):** "# Tool Calls" ranging from approximately 3.0 to 6.0.
* **Legend (Bottom):**
* Peach: "Mean@5 (Train)"
* Gray: "Mean@32 (ReAct)"
* Light Green: "Mean@32 (AIME24)"
* Light Blue: "Mean@32 (AIME25)"
* **Titles (Top):** "Train", "AIME24", "AIME25" are placed above each set of bars, indicating the model being evaluated.
### Detailed Analysis or Content Details
**Left Chart (Performance):**
* **Train (Peach):**
* Step 1: Approximately 84.
* Step 2: Approximately 88.
* Step 3: Approximately 87.
* **ReAct (Gray):**
* Step 1: Approximately 79.
* Step 2: Approximately 72.
* Step 3: Approximately 68.
* **AIME24 (Light Green):**
* Step 1: Approximately 79.
* Step 2: Approximately 82.
* Step 3: Approximately 84.
* **AIME25 (Light Blue):**
* Step 1: Approximately 72.
* Step 2: Approximately 75.
* Step 3: Approximately 70.
**Right Chart (# Tool Calls):**
* **Train (Peach):**
* Step 1: Approximately 3.3.
* Step 2: Approximately 3.6.
* Step 3: Approximately 3.5.
* **ReAct (Gray):**
* Step 1: Approximately 4.2.
* Step 2: Approximately 5.7.
* Step 3: Approximately 4.0.
* **AIME24 (Light Green):**
* Step 1: Approximately 4.5.
* Step 2: Approximately 4.8.
* Step 3: Approximately 4.7.
* **AIME25 (Light Blue):**
* Step 1: Approximately 4.8.
* Step 2: Approximately 5.1.
* Step 3: Approximately 5.0.
### Key Observations
* **Performance Trend:** For the "Train" model, performance increases from Step 1 to Step 2, then slightly decreases at Step 3. "ReAct" shows a consistent decline in performance across all steps. "AIME24" shows a steady increase in performance. "AIME25" shows a slight increase from Step 1 to Step 2, then a decrease at Step 3.
* **Tool Call Trend:** The number of tool calls generally increases from Step 1 to Step 2, and then plateaus or slightly decreases at Step 3 for all models.
* **ReAct Outlier:** The "ReAct" model consistently exhibits the lowest performance and a relatively high number of tool calls.
* **Train vs. AIME Models:** The "Train" model generally achieves the highest performance, while the "AIME" models show lower, but relatively stable, performance.
### Interpretation
The data suggests that increasing the "Learning Step" does not necessarily lead to improved performance, particularly for the "ReAct" model. The "Train" model demonstrates the best performance, but its gains diminish after Step 2. The "AIME" models show a more consistent, though lower, level of performance. The higher number of tool calls associated with the "ReAct" model, coupled with its declining performance, may indicate that the model is struggling to effectively utilize the tools available to it. The relationship between tool calls and performance is not straightforward; more tool calls do not automatically translate to better performance. The data could be used to inform decisions about which learning steps and models to prioritize for further development. The "ReAct" model may require further investigation to understand why it is underperforming. The consistent performance of the AIME models suggests they may be more robust to changes in the learning step.
</details>
Figure 4: Statistics at each Training-Free GRPO step with tool use and DeepSeek-V3.1-Terminus.
Main Results. As illustrated in Table 3, Training-Free GRPO achieves substantial gains in mathematical reasoning, showing a clear advantage in performance across both the tool-use and non-tool-use scenarios. The strong baseline established by DeepSeek-V3.1-Terminus with ReAct [1] yields scores of $80.0\$ on AIME24 and $67.9\$ on AIME25. Critically, applying Training-Free GRPO to the frozen DeepSeek-V3.1-Terminus elevates its performance significantly, reaching $82.7\$ on AIME24 and $73.3\$ on AIME25. This represents a substantial absolute gain of $+2.7\$ and $+5.4\$ , respectively, which is achieved with only $100$ out-of-domain training examples and zero gradient updates. This performance surpasses various state-of-the-art RL methods like ReTool [4] and AFM [16] trained on 32B LLMs (see Table 3), which typically require thousands of training examples and incur costs exceeding $\mathdollar 10,000$ . In contrast, Training-Free GRPO utilizes only $100$ data points with an approximate cost of $\mathdollar 18$ . Such outcome suggests that in real-world applications, guiding a powerful but frozen model through context-space optimization is more effective and efficient than exhaustively fine-tuning a less capable model in parameter space. For quantitative evaluation and learned experiences of Training-Free GRPO, please refer to Appendices A and C, respectively.
Learning Dynamics. As shown in Figure 4, during the 3-step learning process, we observe a steady and significant improvement in Mean@5 on the training set. Concurrently, the Mean@32 performance on both AIME24 and AIME25 also improves with each step, demonstrating that the learned experiences from only $100$ problems generalize effectively and the necessity of multi-step learning. Also, the average number of tool calls decreases during both training and out-of-domain evaluation on AIME benchmarks. This suggests that Training-Free GRPO not only encourages correct reasoning and action, but also teaches the agent to use tools more efficiently and judiciously. The learned experiential knowledge helps the agent to discover some shortcuts and avoid erroneous or redundant tool calls, validating the effectiveness of our semantic advantage guided optimization.
Table 2: Ablation study on DeepSeek-V3.1-Terminus (Mean@32, %).
| ReAct [1] | - | 80.0 | 67.9 |
| --- | --- | --- | --- |
| ReAct [1] + Directly Generated Experiences | - | 79.8 | 67.3 |
| ReAct [1] + Training-Free GRPO (w/o ground truths) | DAPO-100 | 80.7 | 68.9 |
| ReAct [1] + Training-Free GRPO (w/o group computation) | DAPO-100 | 80.4 | 69.3 |
| ReAct [1] + Training-Free GRPO | DAPO-100 | 82.7 | 73.3 |
Table 3: Mean@32 with smaller LLMs on AIME 2024 and AIME 2025 benchmarks (%).
| Method | Learning Cost | Model | Tool | AIME24 | AIME25 |
| --- | --- | --- | --- | --- | --- |
| ReAct [1] | - | Qwen2.5-32B-Instruct | CI | 29.6 | 23.1 |
| ZeroTIR [5] | $\approx\mathdollar 20,000$ | Qwen2.5-32B-Instruct | CI | 56.7 | 33.3 |
| SimpleTIR [6] | $\approx\mathdollar 20,000$ | Qwen2.5-32B-Instruct | CI | 59.9 | 49.2 |
| ReTool [4] | $\approx\mathdollar 10,000$ | Qwen2.5-32B-Instruct | CI | 67.0 | 49.3 |
| AFM [16] | $\approx\mathdollar 10,000$ | Qwen2.5-32B-Instruct | CI | 66.7 | 59.8 |
| ReAct [1] | - | Qwen3-32B (Non-Thinking) | CI | 29.1 | 19.5 |
| + Training-Free GRPO | $\approx\mathdollar 4$ | Qwen3-32B (Non-Thinking) | CI | 33.5 ( $\uparrow$ 4.4) | 25.4 ( $\uparrow$ 5.9) |
| ReAct [1] | - | Qwen2.5-72B-Instruct | CI | 13.5 | 9.6 |
| + Training-Free GRPO | $\approx\mathdollar 3$ | Qwen2.5-72B-Instruct | CI | 14.9 ( $\uparrow$ 1.4) | 11.4 ( $\uparrow$ 1.8) |
| ReAct [1] | - | DeepSeek-V3.1-Terminus | CI | 80.0 | 67.9 |
| + Training-Free GRPO | $\approx\mathdollar 18$ | DeepSeek-V3.1-Terminus | CI | 82.7 ( $\uparrow$ 2.7) | 73.3 ( $\uparrow$ 5.4) |
Effectiveness of Learned Experiences. In Table 2, we also include the ReAct enhanced with the experiences directly generated by DeepSeek-V3.1-Terminus, matching the quantity learned from Training-Free GRPO. However, such directly generated experiences fail to improve the performance, highlighting the effectiveness of learned experiential knowledge from Training-Free GRPO.
Robustness to Reward Signal. Table 2 also presents a variant of Training-Free GRPO, where the ground truth answers are not provided during learning process. In such cases, the semantic advantage is directly obtained by comparing the rollouts within each group, where the LLM can only rely on implicit majority voting, self-discrimination and self-reflection to optimize the experiences. Although it does not surpass the default version with ground truths, Training-Free GRPO still achieves an impressive results of 80.7% on AIME24 and 68.9% on AIME25. It shows the robustness and applicability to domains where ground truths are scarce or unavailable, further broadening its practical utility.
Removing Group Computation. We also remove the group computation by setting the group size of $1$ in Training-Free GRPO, where the LLM can only distill semantic advantage from the single rollout of each query. The results in Table 2 show that it significantly harm the performance comparing with the default group size of $5$ . It confirms the necessity of group relative computation, which enables the LLM to compare different trajectories within each group for better semantic advantage and experience optimization.
Applicability to Different Model Sizes. By applying Training-Free GRPO on smaller LLMs, specifically Qwen3-32B [22] and Qwen2.5-72B-Instruct [23] with DAPO-100 dataset, we observe consistent improvements on the out-of-domain AIME benchmarks in Table 3. Training-Free GRPO requires significantly fewer data and much lower learning cost, contrasting sharply with recent RL methods like ZeroTIR [5], SimpleTIR [6], ReTool [4], and AFM [16], which often necessitate thousands of data points and substantial computational resources for parameter tuning. Furthermore, powered by larger models like DeepSeek-V3.1-Terminus, our approach achieves much higher Mean@32 on AIME benchmarks than all RL-trained models, while only incurring about $\mathdollar 18$ for the learning process.
### 3.2 Web Searching
Table 4: Pass@1 on WebWalkerQA (%).
| ReAct [1] + Training-Free GRPO | - AFM-100 | DeepSeek-V3.1-Terminus DeepSeek-V3.1-Terminus | 63.2 67.8 ( $\uparrow$ 4.6) |
| --- | --- | --- | --- |
In this section, we evaluate the effectiveness of Training-Free GRPO in addressing web searching tasks by leveraging minimal experiential data to enhance agent behavior.
Datasets. For training, we constructed a minimal training set by randomly sampling 100 queries from the AFM (Chain-of-Agents) web interaction RL dataset [16], denoted as AFM-100. AFM provides high-quality, multi-turn interactions between agents and web environments, collected via reinforcement learning in realistic browsing scenarios. For evaluation, we employ WebWalkerQA benchmark [24], a widely-used dataset for assessing web agent performance. Its tasks require understanding both natural language instructions and complex web page structures, making it a rigorous evaluation framework for generalist agents.
Methods. Our proposed Training-Free GRPO is applied to DeepSeek-V3.1-Terminus without any gradient-based updates. We perform 3 epochs of training-free optimization with a group size of $G=3$ . The temperature settings follow those used in prior mathematical experiments.
Main Results. We evaluate the effectiveness of our proposed Training-Free GRPO method on the WebWalkerQA benchmark. As shown in Table 4, our method achieves a pass@1 score of 67.8% using DeepSeek-V3.1-Terminus, a significant improvement over the baseline of 63.2%. This result indicates that our approach effectively steers model behavior by leveraging learned experiential knowledge, surpassing the capabilities of static prompting ReAct strategy.
Table 5: Ablation results on WebWalkerQA subset (%).
| Method | Training Set | Model | pass@1 | pass@3 |
| --- | --- | --- | --- | --- |
| ReAct [1] | - | QwQ-32B | 27.5 | 43.1 |
| ReAct [1] + Training-Free GRPO | AFM-100 | QwQ-32B | 25.5 | 45.1 |
| ReAct [1] | - | DeepSeek-V3.1-Terminus | 66.7 | 74.5 |
| ReAct [1] + Directly Generated Experiences | AFM-100 | DeepSeek-V3.1-Terminus | 64.7 | 76.5 |
| ReAct [1] + Training-Free GRPO (w/o ground truths) | AFM-100 | DeepSeek-V3.1-Terminus | 66.7 | 78.4 |
| ReAct [1] + Training-Free GRPO | AFM-100 | DeepSeek-V3.1-Terminus | 68.6 | 78.4 |
Ablation. We conduct ablation studies on a stratified random sample of 51 instances from the WebWalkerQA test set, where the sampling is proportionally stratified by difficulty level to ensure balanced representation across different levels of complexity. All ablated models are evaluated after 2 epochs of experience optimization. The results are summarized in Table 5.
Using directly generated experiences slightly degrades over ReAct (64.7% vs. 66.7% pass@1), confirming that mere in-context examples without proper optimization may not yield gains. Training-Free GRPO without ground truth maintains the same pass@1 as ReAct (66.7%) but improves pass@3 to 78.4%, demonstrating that relative reward evaluation can enhance consistency even without ground truth. The full Training-Free GRPO achieves the best performance (68.6% pass@1 and 78.4% pass@3), highlighting the importance of combining ground truth guidance with semantic advantage and experience optimization.
Applying Training-Free GRPO to QwQ-32B [25] yields only 25.5% pass@1, significantly lower than the 66.7% achieved with DeepSeek-V3.1-Terminus, and even under performing its own ReAct baseline (27.5%). This may suggest that the effectiveness of our method is dependent on the underlying modelâs reasoning and tool-use capabilities in complex tool use scenarios, indicating that model capability is a prerequisite for effective experience-based optimization.
## 4 Comparing RL Learning on Context Space and Parameter Space
### 4.1 Cross-domain Transfer Analysis
Table 6: Cross-domain transferability (Averaged pass@1, %).
| Method | Learned Domain | Math Reasoning | Web Searching | |
| --- | --- | --- | --- | --- |
| AIME24 | AIME25 | WebWalker | | |
| ReAct [1] (Qwen2.5-32B-Instruct) | - | 29.6 | 23.1 | 31.9 |
| ReTool [4] (Qwen2.5-32B-Instruct) | Math | 67.0 | 49.3 | 18.3 |
| MiroThinker [10] (Qwen3-32B) | Web | 43.5 | 36.8 | 53.6 |
| Training-Free GRPO (DeepSeek-V3.1-Terminus) | Math / Web | 82.7 | 73.3 | 67.8 |
A critical strength of Training-Free GRPO lies in its ability to achieve strong performance across diverse domains without suffering from the domain specialization trade-off observed in parameter-tuned methods. As demonstrated in Table 6, we observe the unsatisfactory performance when domain-specialized models are transferred to different domains. For instance, ReTool [4] specifically trained on mathematical reasoning tasks, achieves competitive performance on AIME24 and AIME25 within its specialized domain. However, when transferred to web searching tasks on WebWalker, its performance drops dramatically to only $18.3\$ , which is much lower than ReAct [1] without fine-tuning. Similarly, though optimized for web interactions, MiroThinker [10] significantly underperforms ReTool that is trained in the math domain on the AIME benchmarks. Such phenomenon highlights that parameter-based specialization narrows the modelâs capabilities to excel in the training domain at the expense of generalizability. In contrast, Training-Free GRPO applied to the frozen LLM achieves state-of-the-art performance in both domains simultaneously by simply plugging in domain-specific learned experiences. Such cross-domain robustness makes Training-Free GRPO particularly valuable for real-world applications where agents must operate in multifaceted environments with diverse requirements.
### 4.2 Computational Costs
As shown in Figure 1, we further analyze the economic benefits of Training-Free GRPO by comparing its computational costs with a vanilla RL training approach, specifically ReTool [4], on mathematical problem-solving tasks. This comparison highlights the practical benefits of our method in scenarios characterized by limited data, constrained budgets, or volatile inference demand.
Training Cost. By replicating the training process of ReTool [4] on Qwen2.5-32B-Instruct [23], we find that it requires approximately 20,000 $\times$ GPU hours with the rental price of $0.5 per GPU hour, resulting in the total training expense amounts to roughly $10,000. In contrast, Training-Free GRPO, when applied to DeepSeek-V3.1-Terminus, achieves superior performance on the AIME benchmarks (Table 3) while requiring only minimal costs. It requires only $3$ training steps over $100$ samples completed within $6$ hours, which consumes 38M input tokens and 6.6M output tokens, amounting to a total cost of approximately $18 based on the official DeepSeek AI pricing https://api-docs.deepseek.com/quick_start/pricing. Most of the input tokens qualify for the lower cache hit pricing with ReAct framework, as each processing step typically involves reusing extensive prior context.. The drastic reduction in training cost by over two orders of magnitude makes our approach especially cost-effective.
Inference Cost. Deploying a trained model like ReTool-32B entails significant fixed infrastructure costs. In a typical serving setup with 4 $\times$ GPUs at the price of $0.5 per GPU-hour, vLLM-based batching requests can process about 400 problems per hour from the AIME benchmarks. The inference cost per problem averages $0.005. While this per-instance cost is relatively low, it presupposes continuous GPU availability, which becomes inefficient under fluctuating or low request volumes. In contrast, Training-Free GRPO incurs a token-based cost. On average, each request consumes 60K input tokens and 8K output tokens, totaling about $0.02 per problem with cache hit pricing for reused contexts. Although per-query inference with a large API-based model is more expensive than with a dedicated small model, many real-world applications, particularly specialized or low-traffic services, experience irregular and modest usage patterns. In such cases, maintaining a dedicated GPU cluster is economically unjustifiable. By leveraging the shared, on-demand infrastructure of large model services like DeepSeek, Training-Free GRPO eliminates fixed serving overhead and aligns costs directly with actual usage. This pay-as-you-go model is distinctly advantageous in settings where demand is unpredictable or sparse.
## 5 Related Work
LLM Agents. By leveraging external tools, Large Language Models (LLMs) can overcome inherent limitations, such as lacking real-time knowledge and precise computation. This has spurred the development of LLM agents that interleave reasoning with actions. Foundational frameworks like ReAct [1] prompt LLMs to generate explicit chain-of-thought (CoT) and actionable steps, enabling dynamic planning through tool use. Furthermore, Toolformer [26] demonstrates that LLMs can learn to self-supervise the invocation of APIs via parameter fine-tuning. Building on these principles, subsequent research has produced sophisticated single- and multi-agent systems, such as MetaGPT [27], CodeAct [28], and OWL [29], which significantly enhance the quality of planning, action execution, and tool integration.
Reinforcement Learning. Reinforcement learning (RL) has proven highly effective for aligning LLMs with complex and long-horizon goals. Foundational algorithms like Proximal Policy Optimization (PPO) [30] employ a policy model for generation and a separate critic model to estimate token-level value. Group Relative Policy Optimization (GRPO) [17] eliminates the need for a critic by estimating advantages directly from groups of responses. Recent research try to apply RL to transform LLMs from passive generators into autonomous agents that learn through environmental interaction. GiGPO [31] implements a two-level grouping mechanism for trajectories, enabling precise credit assignment at both the episode and individual step levels. ReTool [4] uses PPO to train an agent to interleave natural language with code execution for mathematical reasoning. Chain-of-Agents [16] facilitates multi-agent collaboration within a single model by using dynamic, context-aware activation of specialized tool and role-playing agents. Furthermore, Tongyi Deep Research [7] introduces synthetic data generation pipeline and conduct customized on-policy agentic RL framework. However, such parameter-updating approaches result in prohibitive computational cost, which typically restricts application to LLMs with fewer than 32B parameters. Moreover, they only achieve diminishing returns compared to simply using larger, more powerful frozen LLMs. In contrast, our proposed Training-Free GRPO method seeks to achieve comparable or even better performance on state-of-the-art LLMs without any parameter updates, drastically reducing both data and computational requirements.
Training-Free Methods. A parallel line of research aims to improve LLM behavior at inference time without updating model weights. The general approach is in-context learning (ICL) [21], which leverages external or self-generated demonstrations within a prompt to induce desired behaviors. More recent methods introduce iterative refinement mechanisms. Self-Refine [32] generates an initial output and then uses the same LLM to provide verbal feedback for subsequent revisions. Similarly, Reflexion [33] incorporates an external feedback signal to prompt the model for reflection and a new attempt. In-context reinforcement learning (ICRL) [34, 35] demonstrates that LLMs can learn from scalar reward signals by receiving prompts containing their past outputs and associated feedback. TextGrad [15] proposes a more general framework, treating optimization as a process of back-propagating textual feedback through a structured computation graph. A key characteristic of these methods is their focus on iterative, within-sample improvement for a single query. In contrast, our Training-Free GRPO more closely mirrors traditional RL by learning from a separate dataset across multiple epochs to iteratively refine a shared, high-quality experience library for out-of-domain queries. Furthermore, given each query, unlike self-critique or context updates for a single trajectory, our method explicitly compares multiple rollouts per query for a semantic advantage to compare different trajectories in each group, which has been confirmed effective in Section 3.1. Specifically in the context for optimizing agent systems, Agent KB [36] constructs a shared, hierarchical knowledge base to enable the reuse of problem-solving experiences across tasks. Unlike the complex reason-retrieve-refine process of Agent KB, Training-Free GRPO simply injects the learned experiences into the prompt. Moreover, Agent KB relies on hand-crafted examples and employs an off-policy learning paradigm only once by collecting trajectories in the different way of online inference. In contrast, our Training-Free GRPO uses a consistent pipeline and more closely mirrors on-policy RL with multi-epoch learning.
## 6 Conclusion
In this paper, we introduced Training-Free GRPO, a novel paradigm that shifts RL policy optimization from the parameter space to the context space. By leveraging group-based rollouts to iteratively distill a semantic advantage into an evolving experiential knowledge which serves as the token prior, our method successfully steers the output distribution of a frozen LLM agent, achieving significant performance gains in specialized domains. Experiments demonstrate that Training-Free GRPO not only surmounts the practical challenges of data scarcity and high computational cost but also outperforms traditional parameter-tuning methods. Our work establishes a new, highly efficient pathway for adapting powerful LLM agents, making advanced agentic capabilities more accessible and practical for real-world applications.
## Contributions
Authors Yuzheng Cai 1,2* Siqi Cai 1* Yuchen Shi 1* Zihan Xu 1* Lichao Chen 1,3 Yulei Qin 1 Xiaoyu Tan 1 Gang Li 1 Zongyi Li 1 Haojia Lin 1 Yong Mao 1 Ke Li 1Q Xing Sun 1
Affiliations 1 Tencent Youtu Lab 2 Fudan University 3 Xiamen University
*Equal Contributions Yuzheng Cai Siqi Cai Yuchen Shi Zihan Xu
## References
- Yao et al. [2023] Shunyu Yao, Jeffrey Zhao, Dian Yu, Nan Du, Izhak Shafran, Karthik Narasimhan, and Yuan Cao. React: Synergizing reasoning and acting in language models. In International Conference on Learning Representations (ICLR), 2023.
- AIME [2025] AIME. Aime problems and solutions, 2025. URL https://artofproblemsolving.com/wiki/index.php/AIME_Problems_and_Solutions.
- DeepSeek-AI [2024] DeepSeek-AI. Deepseek-v3 technical report, 2024. URL https://arxiv.org/abs/2412.19437.
- Feng et al. [2025a] Jiazhan Feng, Shijue Huang, Xingwei Qu, Ge Zhang, Yujia Qin, Baoquan Zhong, Chengquan Jiang, Jinxin Chi, and Wanjun Zhong. Retool: Reinforcement learning for strategic tool use in llms. arXiv preprint arXiv:2504.11536, 2025a.
- Mai et al. [2025] Xinji Mai, Haotian Xu, Weinong Wang, Jian Hu, Yingying Zhang, Wenqiang Zhang, et al. Agent rl scaling law: Agent rl with spontaneous code execution for mathematical problem solving. arXiv preprint arXiv:2505.07773, 2025.
- Xue et al. [2025] Zhenghai Xue, Longtao Zheng, Qian Liu, Yingru Li, Xiaosen Zheng, Zejun Ma, and Bo An. Simpletir: End-to-end reinforcement learning for multi-turn tool-integrated reasoning. arXiv preprint arXiv:2509.02479, 2025.
- Tongyi DeepResearch Team [2025] Tongyi DeepResearch Team. Tongyi-deepresearch. https://github.com/Alibaba-NLP/DeepResearch, 2025.
- Tao et al. [2025] Zhengwei Tao, Jialong Wu, Wenbiao Yin, Junkai Zhang, Baixuan Li, Haiyang Shen, Kuan Li, Liwen Zhang, Xinyu Wang, Yong Jiang, et al. Webshaper: Agentically data synthesizing via information-seeking formalization. arXiv preprint arXiv:2507.15061, 2025.
- Jin et al. [2025] Bowen Jin, Hansi Zeng, Zhenrui Yue, Jinsung Yoon, Sercan Arik, Dong Wang, Hamed Zamani, and Jiawei Han. Search-r1: Training llms to reason and leverage search engines with reinforcement learning. arXiv preprint arXiv:2503.09516, 2025.
- Team [2025a] MiroMind AI Team. Mirothinker: An open-source agentic model series trained for deep research and complex, long-horizon problem solving. https://github.com/MiroMindAI/MiroThinker, 2025a.
- Zhang et al. [2024] Kechi Zhang, Jia Li, Ge Li, Xianjie Shi, and Zhi Jin. Codeagent: Enhancing code generation with tool-integrated agent systems for real-world repo-level coding challenges. arXiv preprint arXiv:2401.07339, 2024.
- Huang et al. [2023] Dong Huang, Jie M Zhang, Michael Luck, Qingwen Bu, Yuhao Qing, and Heming Cui. Agentcoder: Multi-agent-based code generation with iterative testing and optimisation. arXiv preprint arXiv:2312.13010, 2023.
- Wang et al. [2024a] Shuai Wang, Weiwen Liu, Jingxuan Chen, Yuqi Zhou, Weinan Gan, Xingshan Zeng, Yuhan Che, Shuai Yu, Xinlong Hao, Kun Shao, et al. Gui agents with foundation models: A comprehensive survey. arXiv preprint arXiv:2411.04890, 2024a.
- Wang et al. [2024b] Junyang Wang, Haiyang Xu, Haitao Jia, Xi Zhang, Ming Yan, Weizhou Shen, Ji Zhang, Fei Huang, and Jitao Sang. Mobile-agent-v2: Mobile device operation assistant with effective navigation via multi-agent collaboration. Advances in Neural Information Processing Systems, 37:2686â2710, 2024b.
- Yuksekgonul et al. [2025] Mert Yuksekgonul, Federico Bianchi, Joseph Boen, Sheng Liu, Pan Lu, Zhi Huang, Carlos Guestrin, and James Zou. Optimizing generative ai by backpropagating language model feedback. Nature, 639(8055):609â616, 2025.
- Li et al. [2025] Weizhen Li, Jianbo Lin, Zhuosong Jiang, Jingyi Cao, Xinpeng Liu, Jiayu Zhang, Zhenqiang Huang, Qianben Chen, Weichen Sun, Qiexiang Wang, Hongxuan Lu, Tianrui Qin, Chenghao Zhu, Yi Yao, Shuying Fan, Xiaowan Li, Tiannan Wang, Pai Liu, King Zhu, He Zhu, Dingfeng Shi, Piaohong Wang, Yeyi Guan, Xiangru Tang, Minghao Liu, Yuchen Eleanor Jiang, Jian Yang, Jiaheng Liu, Ge Zhang, and Wangchunshu Zhou. Chain-of-agents: End-to-end agent foundation models via multi-agent distillation and agentic rl. 2025. URL https://arxiv.org/abs/2508.13167.
- Shao et al. [2024] Zhihong Shao, Peiyi Wang, Qihao Zhu, Runxin Xu, Junxiao Song, Xiao Bi, Haowei Zhang, Mingchuan Zhang, YK Li, Yang Wu, et al. Deepseekmath: Pushing the limits of mathematical reasoning in open language models. arXiv preprint arXiv:2402.03300, 2024.
- Liu et al. [2025] Zichen Liu, Changyu Chen, Wenjun Li, Penghui Qi, Tianyu Pang, Chao Du, Wee Sun Lee, and Min Lin. Understanding r1-zero-like training: A critical perspective. arXiv preprint arXiv:2503.20783, 2025.
- Yu et al. [2025] Qiying Yu, Zheng Zhang, Ruofei Zhu, Yufeng Yuan, Xiaochen Zuo, Yu Yue, Weinan Dai, Tiantian Fan, Gaohong Liu, Lingjun Liu, et al. Dapo: An open-source llm reinforcement learning system at scale. arXiv preprint arXiv:2503.14476, 2025.
- Zheng et al. [2025] Chujie Zheng, Shixuan Liu, Mingze Li, Xiong-Hui Chen, Bowen Yu, Chang Gao, Kai Dang, Yuqiong Liu, Rui Men, An Yang, et al. Group sequence policy optimization. arXiv preprint arXiv:2507.18071, 2025.
- Brown et al. [2020] Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Language models are few-shot learners. Advances in neural information processing systems, 33:1877â1901, 2020.
- Yang et al. [2025a] An Yang, Anfeng Li, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu, Chang Gao, Chengen Huang, Chenxu Lv, et al. Qwen3 technical report. arXiv preprint arXiv:2505.09388, 2025a.
- Yang et al. [2025b] An Yang, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu, Chengyuan Li, Dayiheng Liu, Fei Huang, Haoran Wei, Huan Lin, Jian Yang, Jianhong Tu, Jianwei Zhang, Jianxin Yang, Jiaxi Yang, Jingren Zhou, Junyang Lin, Kai Dang, Keming Lu, Keqin Bao, Kexin Yang, Le Yu, Mei Li, Mingfeng Xue, Pei Zhang, Qin Zhu, Rui Men, Runji Lin, Tianhao Li, Tianyi Tang, Tingyu Xia, Xingzhang Ren, Xuancheng Ren, Yang Fan, Yang Su, Yichang Zhang, Yu Wan, Yuqiong Liu, Zeyu Cui, Zhenru Zhang, and Zihan Qiu. Qwen2.5 technical report, 2025b. URL https://arxiv.org/abs/2412.15115.
- Wu et al. [2025] Jialong Wu, Wenbiao Yin, Yong Jiang, Zhenglin Wang, Zekun Xi, Runnan Fang, Linhai Zhang, Yulan He, Deyu Zhou, Pengjun Xie, and Fei Huang. Webwalker: Benchmarking llms in web traversal. 2025. URL https://arxiv.org/abs/2501.07572.
- Team [2025b] Qwen Team. Qwq-32b: Embracing the power of reinforcement learning, March 2025b. URL https://qwenlm.github.io/blog/qwq-32b/.
- Schick et al. [2023] Timo Schick, Jane Dwivedi-Yu, Roberto DessĂŹ, Roberta Raileanu, Maria Lomeli, Luke Zettlemoyer, Nicola Cancedda, and Thomas Scialom. Toolformer: Language models can teach themselves to use tools, 2023. arXiv preprint arXiv:2302.04761, 2023.
- Hong et al. [2024] Sirui Hong, Mingchen Zhuge, Jonathan Chen, Xiawu Zheng, Yuheng Cheng, Ceyao Zhang, Jinlin Wang, Zili Wang, Steven Ka Shing Yau, Zijuan Lin, et al. MetaGPT: Meta programming for a multi-agent collaborative framework. International Conference on Learning Representations, ICLR, 2024.
- Wang et al. [2024c] Xingyao Wang, Yangyi Chen, Lifan Yuan, Yizhe Zhang, Yunzhu Li, Hao Peng, and Heng Ji. Executable code actions elicit better llm agents. In Forty-first International Conference on Machine Learning, 2024c.
- Hu et al. [2025] Mengkang Hu, Yuhang Zhou, Wendong Fan, Yuzhou Nie, Bowei Xia, Tao Sun, Ziyu Ye, Zhaoxuan Jin, Yingru Li, Qiguang Chen, Zeyu Zhang, Yifeng Wang, Qianshuo Ye, Bernard Ghanem, Ping Luo, and Guohao Li. OWL: Optimized workforce learning for general multi-agent assistance in real-world task automation, 2025. URL https://arxiv.org/abs/2505.23885.
- Schulman et al. [2017] John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov. Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347, 2017.
- Feng et al. [2025b] Lang Feng, Zhenghai Xue, Tingcong Liu, and Bo An. Group-in-group policy optimization for llm agent training. arXiv preprint arXiv:2505.10978, 2025b.
- Madaan et al. [2023] Aman Madaan, Niket Tandon, Prakhar Gupta, Skyler Hallinan, Luyu Gao, Sarah Wiegreffe, Uri Alon, Nouha Dziri, Shrimai Prabhumoye, Yiming Yang, et al. Self-refine: Iterative refinement with self-feedback. Advances in Neural Information Processing Systems, 36:46534â46594, 2023.
- Shinn et al. [2023] Noah Shinn, Federico Cassano, Ashwin Gopinath, Karthik Narasimhan, and Shunyu Yao. Reflexion: Language agents with verbal reinforcement learning. Advances in Neural Information Processing Systems, 36:8634â8652, 2023.
- Song et al. [2025] Kefan Song, Amir Moeini, Peng Wang, Lei Gong, Rohan Chandra, Yanjun Qi, and Shangtong Zhang. Reward is enough: Llms are in-context reinforcement learners. arXiv preprint arXiv:2506.06303, 2025.
- Monea et al. [2024] Giovanni Monea, Antoine Bosselut, Kianté Brantley, and Yoav Artzi. Llms are in-context reinforcement learners. 2024.
- Tang et al. [2025] Xiangru Tang, Tianrui Qin, Tianhao Peng, Ziyang Zhou, Daniel Shao, Tingting Du, Xinming Wei, Peng Xia, Fang Wu, He Zhu, et al. Agent KB: Leveraging cross-domain experience for agentic problem solving. arXiv preprint arXiv:2507.06229, 2025.
## Appendix A Case Study
On math reasoning and web searching scenarios, we provide examples illustrating how Training-Free GRPO enhances ReAct [1] with DeepSeek-V3.1-Terminus and corresponding tools.
### A.1 Experience-Guided Tool-Integrated Math Reasoning
We consider a geometric configuration with two rectangles $ABCD$ and $EFGH$ where $D,E,C,F$ are collinear in that order, and $A,D,H,G$ are concyclic. Given $BC=16$ , $AB=107$ , $FG=17$ , $EF=184$ , the task is to determine $CE$ .
Tool-integrated Math Reasoning (without experiences)
Question Rectangles $ABCD$ and $EFGH$ are drawn such that $D,E,C,F$ are collinear. Also, $A,D,H,G$ all lie on a circle. If $BC=16$ , $AB=107$ , $FG=17$ , and $EF=184$ , what is the length of $CE$ ?
Trajectory
[Agent]
Iâll place point $D$ at the origin $(0,0)$ and use a horizontal line for $D,E,C,F$ . For rectangle $ABCD$ , set $A=(0,16)$ , $B=(107,16)$ , $C=(107,0)$ , $D=(0,0)$ .
[Tool]
Set coordinates accordingly.
[Agent]
Because $EFGH$ is a rectangle with $EF$ horizontal and $FG=17$ , set $F=(x_{f},0)$ , $E=(x_{f}-184,0)$ , and initially set $G=(x_{f},{\color[rgb]{0.94140625,0.1875,0.625}\definecolor[named]{pgfstrokecolor}{rgb}{0.94140625,0.1875,0.625}+17})$ , $H=(x_{f}-184,{\color[rgb]{0.94140625,0.1875,0.625}\definecolor[named]{pgfstrokecolor}{rgb}{0.94140625,0.1875,0.625}+17})$ .
[Agent]
Impose concyclicity of $A(0,16),D(0,0),G(x_{f},17),H(x_{f}-184,17)$ using the 4-point determinant condition, derive an equation in $x_{f}$ .
[Tool]
Solve the resulting quadratic for $x_{f}$ ; obtain $x_{f}\approx 184.092$ and thus $x_{e}=x_{f}-184\approx 0.092$ .
[Agent]
Compute $CE=|107-x_{e}|\approx 106.9076$ .
[Agent]
Swap point orderings ( $D$ â $C$ â $E$ â $F$ , $E$ â $D$ â $C$ â $F$ ) and re-derive. At one point, use an angle-sine relation and get $x_{e}(x_{e}+184)=17$ leading to $x_{e}=-92+\sqrt{8481}\approx 0.0923$ , hence $CE\approx 106.9076$ .
[Agent]
Present an âexact formâ $CE=199-\sqrt{8481}$ as the final answer.
Outcome Incorrect: $CE\approx 106.9076$ (reported as $199-\sqrt{8481}$ ).
Figure 5: Example trajectory without experience in Math Reasoning.
Baseline (without experiences). As shown in Figure 5, the unassisted agent initializes a coordinate system with $D=(0,0)$ and models $ABCD$ as axis-aligned. For $EFGH$ , it places $E=(x_{f}-184,0)$ , $F=(x_{f},0)$ and, critically, sets $G=(x_{f},\,+17)$ , $H=(x_{f}-184,\,+17)$ , i.e., with a positive vertical orientation for the short side. It then enforces the four-point concyclicity of $A(0,16)$ , $D(0,0)$ , $G$ , $H$ via a determinant condition and solves for $x_{f}$ , yielding $x_{f}\approx 184.092$ and consequently $x_{e}=x_{f}-184\approx 0.092$ . From this, it reports $CE\approx 106.9076$ and an âexactâ expression $199-\sqrt{8481}$ .
This trajectory exhibits three systemic issues: (i) misinterpretation of the vertical orientation (wrong sign for the $y$ -coordinates of $G,H$ ), (ii) inconsistent handling of the order $D$ â $E$ â $C$ â $F$ and the lack of a unified parameterization for segment relations, and (iii) absence of systematic, comprehensive post-solution verificationâi.e., no integrated check that the final coordinates simultaneously satisfy rectangle dimensions. These issues lead to an incorrect cyclic constraint (e.g., an intermediate relation of the form $x(x+184)=17$ ) and acceptance of a spurious solution without full geometric verification. Note that although $CE\approx 106.91$ lies within $0<CE<107$ , this alone does not validate the solution; the critical failure was the lack of holistic consistency checks across all problem constraints.
Tool-integrated Math Reasoning (with experiences)
Question Rectangles $ABCD$ and $EFGH$ are drawn such that $D,E,C,F$ are collinear. Also, $A,D,H,G$ all lie on a circle. If $BC=16$ , $AB=107$ , $FG=17$ , and $EF=184$ , what is the length of $CE$ ? Experiences [1] When solving geometry problems with intersections, validate solutions lie within bounded regions or segments, not on extensions, to avoid extraneous answers. $\cdots$ [7] When solving geometry with constraints (e.g., trapezoids), validate all conditions: parallel sides via slopes, convexity via cross products, to avoid extraneous solutions. $\cdots$ [29] When constructing points on line extensions, carefully interpret directional phrases like âbeyond Bâ. $\cdots$ [37] In geometry problems with points on sides of a triangle and given segment lengths, first determine all three side lengths by summing the appropriate segments. For example, if points on AB, AC, BC are given with AM, MB, AN, NC, BP, PC, then AB = AM + MB, AC = AN + NC, BC = BP + PC. This ensures the triangle is consistently defined before coordinate placement or area calculations. $\cdots$
Trajectory
[Agent]
Fix the order $D$ â $E$ â $C$ â $F$ on a horizontal line. Set $CE=x$ with the segment constraint $0<x<107$ . (By [G29],[G1])
[Tool]
Record relations: $DE=107-x$ and $CF=184-x$ (from [G37]).
[Tool]
Place $D=(0,0)$ , $E=(107-x,0)$ , $C=(107,0)$ , $F=(291-x,0)$ , ensuring $EF=184$ .
[Agent]
Choose the vertical orientation consistently: since $A=(0,16)$ and $D=(0,0)$ , set the short side $FG=17$ downward, i.e., $H=(107-x,-17)$ and $G=(291-x,-17)$ . (This corrects the sign error in the no-experience path)
[Agent]
Impose concyclicity using the circle equation $x^{2}+y^{2}+Dx+Ey+F=0$ . With $A(0,16)$ and $D(0,0)$ , get $F=0$ , $E=-16$ .
[Tool]
Substitute $H(107-x,-17)$ and $G(291-x,-17)$ to obtain two equations. Subtract to eliminate constants and get $D=2x-398$ .
[Agent]
Substitute back to get the quadratic $x^{2}-398x+30576=0.$
[Tool]
Compute the discriminant: $398^{2}-4\cdot 30576=36100=190^{2}$ . Roots: $x=294$ and $x=104$ .
[Agent]
Apply the segment constraint $0<x<107$ to discard $x=294$ and select $x=104$ (by [G1])
[Agent]
(Full verification by [G7]) $DE=107-104=3$ , $CF=184-104=80$ , $EF=184$ , $FG=17$ . The circle $x^{2}+y^{2}-190x-16y=0$ passes through $A=(0,16),D=(0,0),H=(3,-17),G=(187,-17)$ . All constraints satisfied.
Outcome Correct: $CE=104$ .
Figure 6: Example trajectory with experiences in Math Reasoning.
Enhanced (With Experiences). Refer to Figure 6, with a curated experience pool, the agent follows a structured pipeline:
- Directional ordering ([29]) and boundedness validation ([1]): It fixes the order $D$ â $E$ â $C$ â $F$ on a line and sets $CE=x$ with $0<x<107$ , ensuring $E$ lies on segment $DC$ and $F$ lies beyond $C$ .
- Segment-addition parameterization ([37]): It uses $DE+EC=DC=AB=107$ and $EC+CF=EF=184$ to obtain $DE=107-x$ , $CF=184-x$ , and places $D=(0,0)$ , $E=(107-x,0)$ , $C=(107,0)$ , $F=(291-x,0)$ .
- Consistent vertical orientation and cyclic modeling: Noting $A=(0,16)$ , $D=(0,0)$ , it orients the short side downward ( $FG=17$ ) so $H=(107-x,-17)$ , $G=(291-x,-17)$ . Using the circle equation $x^{2}+y^{2}+Dx+Ey+F=0$ with $A$ and $D$ yields $F=0$ , $E=-16$ . Substituting $H$ and $G$ , subtracting the two equations gives $D=2x-398$ ; back-substitution reduces to the quadratic $x^{2}-398x+30576=0$ , with discriminant $398^{2}-4\cdot 30576=36100=190^{2}$ and roots $x=104,\,294$ .
- Root selection and full verification ([1], [7]): Applying $0<x<107$ filters out $x=294$ , selecting $x=104$ . The agent then verifies all constraints: $DE=107-104=3$ , $CF=184-104=80$ , $EF=184$ , $FG=17$ , and confirms that the circle $x^{2}+y^{2}-190x-16y=0$ passes through $A=(0,16)$ , $D=(0,0)$ , $H=(3,-17)$ , $G=(187,-17)$ .
Comparative Analysis. This case reveals a clear causal link between experience-guided behaviors and correctness. Experience [29] eliminates directional ambiguity and enforces the correct collinearity order, directly addressing the baselineâs misplacement of $G,H$ . Experience [37] induces a clean single-variable parameterization ( $DE=107-x$ , $CF=184-x$ ), which simplifies the cyclic constraint to a solvable quadratic. Experience [1] imposes a necessary boundedness filter ( $0<x<107$ ) to discard extraneous roots. Finally, experience [7] mandates comprehensive post-solution verification (rectangle dimensions, collinearity, concyclicity), preventing acceptance of spurious solutions.
Compared to the unassisted trajectory, the experience-informed reasoning corrects the vertical orientation, resolves ordering and parameterization inconsistencies, and installs principled validation gates. This case demonstrates the positive impact of integrating domain-specific experiences on reliability and accuracy in tool-integrated mathematical reasoning.
### A.2 Experience-Guided Web Searching
We consider a web searching task from WebWalkerQA: quantify 2024 rewards for (i) creators in the Creator Program (weekly amount), and (ii) players in the Play-2-Airdrop during Private Beta 4 (total pool and duration).
Baseline (without experiences). As summarized in Figure 7, the unassisted agent issues multiple broad searches and relies heavily on result snippets and third-party summaries, delaying clicks into authoritative pages. It eventually opens the PB4 registration post to confirm âPlay-2-Airdrop Campaign: 200,000 ILV over six months,â but continues to scan snippets for the Creator Program value without opening the relevant official post. Connection errors to one official page cause the agent to revert to broad searches rather than alternative primary-source strategies (e.g., site-specific queries or adjacent official posts). The trajectory remains incomplete: it reports the Play-2-Airdrop figure but fails to confirm the Creator Programâs â100 ILV weeklyâ from an official source, yielding an incorrect/incomplete answer.
Enhanced (With Experiences). Refer to Figure 8. With a curated experience pool, the agent follows a disciplined pipeline: (1) prioritize official sources ([14], [4]) and open the PB4 overview and registration posts to extract the â200,000 ILV over six monthsâ and Testnet/Mainnet allocation structure; (2) refine search terms to target formal titles ([2], [12], [1]) and open âThe Illuvium Creator Program Is Getting An Upgrade!â; (3) extract the exact line âThe prize pool of 100 ILV tokens awaits participating Creators each week,â and (4) synthesize both verified statements into a complete answer aligned with the question requirements ([25]). This results in the correct, fully supported output: creators earn 100 ILV weekly; players have a 200,000 ILV pool distributed over six months in PB4âs Play-2-Airdrop.
Web Searching (without experiences)
Question In 2024, how much ILV reward can creators participating in the Creator Program and players engaging in the Play-2-Airdrop during Private Beta 4 on the Illuvium platform earn?
Trajectory
[Agent]
Iâll start with a broad web search for both program names and rewards in 2024.
[Tool]
google_search(âIlluvium Creator Program Play-2-Airdrop Private Beta 4 2024 rewards ILVâ)
[Agent]
From snippets, PB4 mentions a âPlay-2-Airdrop campaignâ and â200,000 ILV over six months.â This likely answers half the question. (continue searching rather than opening)
[Agent]
Iâll search specifically for the Creator Program rewards.
[Tool]
google_search(âĂlluvium Creator Program2Ì024 rewards ILV allocationâ)
[Agent]
I see references (Binance Square, media blogs) to â100 ILV per week,â but canât confirm from illuvium.io quickly. Iâll keep searching.
[Agent]
Avoids clicking potential official announcements; keeps scanning search result snippets.
[Agent]
Iâll fetch the PB4 registration page to confirm the Play-2-Airdrop details.
[Tool]
get_content(âhttps://illuvium.io/news/register-now-for-private-beta-4â)
[Agent]
Extracted: âPlay-2-Airdrop Campaign: 200,000 ILV tokens over six months.â That confirms the second quantity.
[Agent]
Iâll look for the Airdrop Season details page.
[Tool]
google_search(âĂlluvium Airdrop Season DetailsbÌlog post 2024â)
[Tool]
get_content(âhttps://illuvium.io/news/illuvium-airdrop-season-detailsâ)
[Agent]
Connection fails; instead of trying alternative official sources or site-specific queries, returns to broad searches.
[Agent]
More searches for âIlluvium 100 ILV per week creator rewards April 2024â and âsite:illuvium.io Creator Program 100 ILVâ yield sparse official confirmation; Iâll defer the creator number.
[Agent]
Fails to open likely official Creator Program pages; leaves the creator number unconfirmed.
[Agent]
Synthesis: I will answer with whatâs confirmed officially (Play-2-Airdrop 200,000 ILV over 6 months) and note the creator reward as unclear.
Outcome Creators: (not clearly specified; mentions unverified â100 ILV per weekâ). Players (PB4 Play-2-Airdrop): 200,000 ILV over six months.
Figure 7: Example trajectory without experiences in Web Searching.
Web Searching (with experiences)
Question In 2024, how much ILV reward can creators participating in the Creator Program and players engaging in the Play-2-Airdrop during Private Beta 4 on the Illuvium platform earn? Experiences [1] Document identification: Focus on extracting formal titles and collection names from official metadata and headers rather than inferring relationships from content descriptions. $\cdots$ [2] Systematic exploration: Progress from broad initial queries to targeted content extraction using adaptive search strategies that balance discovery with precision. $\cdots$ [4] Source prioritization: Prioritize systematic extraction from authoritative comprehensive documents over fragmented information for coherent topic coverage. $\cdots$ [12] Iterative refinement: Continuously refine search terms based on emerging patterns while periodically re-evaluating previously encountered information to avoid overlooking correct solutions and maintain essential contextual elements. $\cdots$ [14] Primary source prioritization: Prioritize locating official primary source documents over secondary coverage to ensure accuracy and capture exact language for definitive metadata and hierarchical relationships. [25] Partial match evaluation: Systematically evaluate partial matches from initial searches against all criteria before pursuing alternative paths. $\cdots$
Trajectory
[Agent]
Start with broad search, then pivot to official illuvium.io pages for PB4 and Creator Program. (By [2], [14], [4])
[Tool]
google_search(âIlluvium Creator Program rewards 2024 Private Beta 4 ILV earningsâ, num_results=5)
[Agent]
Identify official pages in results: âIlluvium Beta 4: All You Need to Knowâ and âRegister Now for Private Beta 4â. Open them to confirm Play-2-Airdrop details. (By [14], [25])
[Tool]
get_content(âhttps://illuvium.io/news/illuvium-beta-4-all-you-need-to-knowâ)
[Agent]
Extracted from official PB4 overview: Testnet phases get 10% of the Play-2-Airdrop pool; clarifies structure and timeline.
[Tool]
get_content(âhttps://illuvium.io/news/register-now-for-private-beta-4â)
[Agent]
Extracted explicit official number: âPlay-2-Airdrop Campaign: 200,000 ILV tokensâ over six months. This covers the player reward for PB4 context.
[Agent]
Iteratively refine search for the Creator Program exact weekly reward using formal title cues like âCreator Program ⊠upgradeâ. (key behavior; By [12], [1])
[Tool]
google_search(âĂlluvium Creator ProgramrÌewards ILV 2024â, num_results=5)
[Agent]
Click the official post: âThe Illuvium Creator Program Is Getting An Upgrade!â. (By [14], [1], [4])
[Tool]
get_content(âhttps://illuvium.io/news/illuvium-creator-program-is-getting-an-upgradeâ)
[Agent]
Extracted exact statement from official source: âThe prize pool of 100 ILV tokens awaits participating Creators each week.â This provides the creator reward.
Outcome Creators can earn 100 ILV weekly in the Creator Reward Program, and 200,000 ILV will be distributed over six months for players in the Play-2-Airdrop during Private Beta 4.
Figure 8: Example trajectory with experiences in Web Searching.
Comparative Analysis. Experience-guided behaviors directly address baseline deficiencies: primary source prioritization ([14], [4]) removes reliance on snippets and third-party coverage; document identification ([1]) and iterative refinement ([2], [12]) ensure the agent locates and opens the exact Creator Program post; partial match evaluation ([25]) steers the agent to confirm numerical claims at their authoritative origin. In contrast, the baseline wastes context on searches without content acquisition, leaves critical values unverified, and produces an incomplete answer.
## Appendix B Prompts
In this appendix, we provide the prompts used in math reasoning tasks. Figures 9 and 10 present the prompts for solving math tasks. Figures 11 and 12 are used in group relative semantic advantage. Figure 13 works for optimizing the experiential knowledge $\mathcal{E}$ .
System Prompt
Solve the following problem step by step. You now have the ability to selectively write executable Python code to enhance your reasoning process, e.g., calulating numbers and verifying math computations. Never directly just printing your semantic reasoning in Python. The Python code will be executed by an external sandbox, and the output (returned as a dict with the message in the âmessageâ field) can be returned to aid your reasoning and help you arrive at the final answer. The Python code should be complete scripts, including necessary imports. Each code snippet is wrapped with ```python code snippet ```. The last part of your final response should be in the following format: $<$ answer $>$ $\backslash$ boxed{The final answer goes here.} $<$ /answer $>$
Figure 9: System prompt for ReAct framework in Math Reasoning.
Prompt for Involving Experiential Knowledge $\mathcal{E}$
Please solve the problem: {problem} When solving problems, you MUST first carefully read and understand the helpful instructions and experiences: {experiences}
Figure 10: Prompt for supplementing math problems with experiential knowledge $\mathcal{E}$ .
Prompt for Trajectory Summarization in Group Advantage Computation
An agent system may be provided with some experiences, and then it produces the following trajectory to solve the given problem. Please summarize the trajectory step-by-step: 1. For each step, describe what action is being taken, and which experience has been used in this step. 2. Given the grading of this rollout and the correct answer, identify and explain any steps that represent detours, errors, or backtracking, highlighting why they might have occurred and what their impact was on the trajectoryâs progress. 3. Maintain all the core outcome of each step, even if it was part of a flawed process. $<$ trajectory $>$ {trajectory} $<$ /trajectory $>$ $<$ evaluation $>$ {whether the answer is correct or not} $<$ /evaluation $>$ $<$ groundtruth $>$ {the ground truth answer} $<$ /groundtruth $>$ Only return the trajectory summary of each step, e.g., 1. what happened in the first step and the core outcomes 2. what happened in the second step and the core outcomes 3. âŠ
Figure 11: Prompt for summarizing each trajectory in Math Reasoning.
Prompt for Group Advantage Computation
An agent system is provided with a set of experiences and has tried to solve the problem multiple times with both successful and wrong solutions. Review these problem-solving attempt and extract generalizable experiences. Follow these steps: 1. Trajectory Analysis: $\quad-$ For successful steps: Identify key correct decisions and insights $\quad-$ For errors: Pinpoint where and why the reasoning went wrong $\quad-$ Note any important patterns or strategies used/missed $\quad-$ Review why some trajectories fail? Is there any existing experiences are missed, or experiences do not provide enough guidance? 2. Update Existing Experiences $\quad-$ Some trajectories may be correct and others may be wrong, you should ensure there are experiences can help to run correctly $\quad-$ You have three options: [modify, add, delete] $\quad\quad*$ modify: You can modify current experiences to make it helpful $\quad\quad*$ add: You can introduce new experiences to improve future performance $\quad\quad*$ delete: You can delete existing experiences $\quad-$ You can update at most {max number of operations} clear, generalizable lessons for this case $\quad-$ Before updating each experience, you need to: $\quad\quad*$ Specify when it would be most relevant $\quad\quad*$ List key problem features that make this experience applicable $\quad\quad*$ Identify similar problem patterns where this advice applies 3. Requirements for each experience that is modified or added. $\quad-$ Begin with general background with several words in the experience $\quad-$ Focus on strategic thinking patterns, not specific calculations $\quad-$ Emphasize decision points that could apply to similar problems Please provide reasoning in details under the guidance of the above 3 steps. After the step-by-step reasoning, you will finish by returning in this JSON format as follows: ```json $[$ { âoptionâ: âmodifyâ, âexperienceâ: âthe modified experienceâ, âmodified_fromâ: âG17â # specify the ID of experience that is modified }, { âoptionâ: âaddâ, âexperienceâ: âthe added experienceâ, }, { âoptionâ: âdeleteâ, âdelete_idâ: âthe deleted experience IDâ, }, ⊠$]$ ``` Note that your updated experiences may not need to cover all the options. You can only use one type of updates or choose to remain all experiences unchanged. $<$ problem $>$ {problem} $<$ /problem $>$ $<$ trajectories $>$ { $G$ summarized trajectories in the same group} $<$ /trajectories $>$ $<$ groundtruth $>$ {answer} $<$ /groundtruth $>$ $<$ experience $>$ {experiences} $<$ /experience $>$
Figure 12: Prompt for group advantage computation in Math Reasoning.
Prompt for the Optimization Step
An agent system is provided with a set of experiences and has tried to solve the problem multiple times. From the reflections, some suggestions on the existing experiences have been posed. Your task is to collect and think for the final experience revision plan. Each final experience must satisfy the following requirements 1. It must be clear, generalizable lessons for this case, with no more than 32 words 2. Begin with general background with several words in the experience 3. Focus on strategic thinking patterns, not specific calculations 4. Emphasize decision points that could apply to similar problems 5. Avoid repeating saying similar experience in multiple different experiences $<$ experience $>$ {experiences} $<$ /experience $>$ $<$ suggested_updates $>$ {all group advantage in current batch} $<$ /suggested_updates $>$ Please provide reasoning in each of the suggestions, and think for how to update existing experiences You have three update options: [modify, merge, delete] - modify: You can modify current experiences to make it helpful - merge: You can merge some similar experiences into a more general forms to reduce duplication - delete: You can delete an experience After generating the step-by-step reasoning, you need to give the final experience revision details by returning in this JSON format as follows: ```json $[$ { âoptionâ: âmodifyâ, âexperienceâ: âthe modified experienceâ, âmodified_fromâ: âG17â # specify the ID of experience that is modified }, { âoptionâ: âmergeâ, âexperienceâ: âthe merged experienceâ, âmerged_fromâ: [âC1â, âC3â, âS4â, âŠ] # specify the str IDs of experiences that is merged from, at least 2 IDs are needed }, { âoptionâ: âdeleteâ, âdelete_idâ: âthe deleted experience IDâ, }, ⊠$]$ ``` Note that your updated experiences may not need to cover all the options. You can only use one type of updates or choose to remain all experiences unchanged.
Figure 13: Prompt for optimizing experiential knowledge $\mathcal{E}$ based on group advantages in the same batch in Math Reasoning.
## Appendix C Examples of Learned Experiences
In this appendix, we provide some examples in Figure 14, which are extracted from 48 learned experiences by Training-Free GRPO with tool use in Math Reasoning.
Examples of Learned Experiences
[1]. When solving geometry problems with intersections, validate solutions lie within bounded regions or segments, not on extensions, to avoid extraneous answers. [2]. When multiple locations are described relative to a common point using compass directions (e.g., southwest, southeast), analyze the angle between these directions. Southwest and southeast from the same point form a 90° angle, suggesting a right triangle configuration. This insight often allows using the Pythagorean theorem rather than more complex coordinate geometry approaches. [3]. When scoring systems reward consecutive successes, analyze whether bonuses are additive or replace base scores. Test small examples to resolve ambiguity. $\cdots$ [10]. When using mathematical invariants to prove impossibility, always validate them against known achievable states or small cases. If an invariant suggests a state is unreachable but you can construct a valid sequence reaching it, re-examine your invariant analysis - you may have made an error in the mathematical reasoning or the invariant may not be preserved as claimed. [11]. When solving âsmallest primeâ problems involving divisibility of large expressions like geometric series, first handle primes 2 and 3 as special cases. Then systematically check primes in ascending order using multiplicative order properties: for prime p Âż 3, p divides the sum if the multiplicative order of the base modulo p divides the exponent in the closed form. This avoids unnecessary direct computation of huge numbers. [12]. For expected extreme statistics in combinatorial problems, use direct enumeration for small sizes. For larger problems, develop combinatorial counting methods instead of expectation formulas. $\cdots$ [21]. For complex polynomials with real parameters, separate real and imaginary parts to find when real roots exist or all are non-real. [22]. For right triangles with symmetric points on a circle, assume equal y-coordinates and opposite x to simplify area maximization. $\cdots$ [32]. For points on concentric circles, collinearity along a ray minimizes the circumradius to half the max-min radius difference. [33]. In state transition problems with additive operations, consider binary representation invariants. The number of 1âs often determines extremal states or transition possibilities. $\cdots$
Figure 14: Example of learned experiences from Training-Free GRPO with tool use in Math Reasoning.