# Reinforcement Learning-Augmented LLM Agents for Collaborative Decision Making and Performance Optimization
**Authors**: Dong Qiu, Duo Xu, Limengxi Yue
## Reinforcement Learning-Augmented LLM Agents for Collaborative Decision Making and Performance Optimization
Dong Qiu 1,* , Duo Xu 2a , Limengxi Yue 3b
New England College, Henniker, NH 03242, USA 2 Northeastern University, San Jose, CA 95113, USA 3 University of Massachusetts Amherst, Amherst, MA 01003, USA * DQiu\_GPS@nec.edu, a xu.duo3@northeastern.edu, b limengxiyue@outlook.com
Abstract -Large Language Models (LLMs) perform well in language tasks but often lack collaborative awareness and struggle to optimize global performance in multi-agent settings. We present a reinforcement learning-augmented LLM agent framework that formulates cooperation as a decentralized partially observable Markov decision process (Dec-POMDP) and adopts centralized training with decentralized execution (CTDE). We introduce Group Relative Policy Optimization (GRPO) to jointly optimize agent policies with access to global signals during training, together with a simplified joint reward that balances task quality, speed, and coordination cost. On collaborative writing and coding benchmarks, our framework delivers a 3× increase in task processing speed over single-agent baselines, 98.7% structural/style consistency in writing, and 74.6% test pass rate in coding. The approach consistently outperforms strong multi-agent LLM baselines and provides a practical path toward reliable collaboration in complex workflows.
Keyword s -multi-agent systems; reinforcement learning; large language models; Dec-POMDP; CTDE; policy optimization.
## I. INTRODUCTION
Large language models (LLMs) are increasingly used as agents that can plan, write code, call tools, and review intermediate outputs. In many practical tasks-such as iterative problem solving, collaborative content production, and software-oriented workflows-effective performance depends on coordinated behaviors among multiple specialized agents rather than a single monolithic model. This naturally aligns with a decentralized partially observable Markov decision process (Dec-POMDP), where each agent observes only part of the state and the team must act to optimize a shared objective.
Multi-agent reinforcement learning (MARL) offers a principled way to learn coordination policies, but it also introduces persistent challenges, including non-stationary learning dynamics, credit assignment ambiguity, and sensitivity to evaluation protocols. A large-scale benchmarking study highlights that reported gains can vary substantially with implementation details and experimental settings, making careful, reproducible comparisons essential when claiming improvements in cooperative MARL [1]. These concerns are especially relevant for LLM-based agent teams because actions are open-ended (language decisions and tool calls), feedback can be sparse or delayed, and success criteria are often defined by external evaluators (tests, metrics, or human-like preference signals).
Among policy-gradient methods, Proximal Policy Optimization (PPO) remains a widely adopted baseline due to its empirical stability and straightforward implementation [2]. Notably, PPO has been shown to perform strongly in cooperative multi-agent games when pipelines and baselines are controlled, indicating that stable on-policy optimization can be competitive even in multi-agent settings [3]. Trust-region style ideas have also been investigated for MARL to reduce destructive updates and improve learning robustness under changing joint policies. However, applying PPO-style optimization to LLM agent collaboration is not a direct translation: agent interactions are structured as multi-step dialogues and tool-mediated actions, and the training signal must reflect team-level outcomes while still allowing agents to execute with their own local information.
A common strategy for cooperative MARL is centralized training with decentralized execution (CTDE), where a centralized learner can use global information during training, but agents act using local observations at deployment. Value factorization approaches such as QMIX decompose the team value function into agent-wise utilities with a monotonic mixing constraint, enabling scalable learning in cooperative environments [4]. Counterfactual credit assignment, exemplified by COMA, improves attribution by comparing an agent's chosen action to counterfactual alternatives given the same joint context [5]. Centralized-critic actor-critic methods such as MADDPG further stabilize learning by training critics with joint observations/actions while retaining decentralized policies for execution [6]. These ideas are well established for structured action spaces, but for LLM agent teams they must be adapted to language-and-tool action trajectories and to evaluator-defined feedback.
In parallel, LLM-based multi-agent frameworks have emerged that coordinate agents via conversational protocols and role separation [7]. AutoGen demonstrates how multi-agent conversation combined with tool use can support complex multi-step tasks in practical applications [8],[9]. MetaGPT organizes multiple roles in a software-style workflow to produce artifacts through structured collaboration [10]. AgentBench provides benchmarks and evaluation setups to assess whether LLMs can act as agents in interactive environments, encouraging more systematic measurement beyond anecdotal demonstrations [11]. While these systems are effective in many cases, they are often driven primarily by prompting heuristics and fixed coordination rules, with limited use of a unified learning signal that directly optimizes longhorizon team performance under partial observability [12].
Motivated by these gaps, we develop a reinforcementlearning-augmented framework for collaborative LLM agents that connects (i) a Dec-POMDP formulation of team interaction, (ii) a CTDE-style centralized trainer, and (iii) a shared experience buffer that stores trajectories and evaluator feedback for policy improvement. Our design emphasizes evaluatordriven learning: global metrics and critique signals are converted into training targets that improve the overall team behavior, while each agent retains its own execution pathway at inference time. This connects established MARL principlesrobust baselines, stable policy optimization, and careful evaluation practice-to the emerging practice of role-based LLM agent collaboration.
## II. METHODOLOGY
## A. Problem Formulation - Dec-POMDP View of LLM Teams
We cast a team of LLM agents as a cooperative decision process under partial observability where each role-planner, writer, reviewer, coder, tester-must act on a limited, noisy view of the task while pursuing a shared objective and the structure is shown in Fig.1. In practice this means we separate what is globally relevant from what should remain local. The shared part includes the task brief, accepted snippets, approved facts from retrieval, and a compact changelog; the private part includes role-specific scratchpads, TODO lists, and short-term hypotheses that may be wrong and should not pollute the global context. Actions are not raw chat turns but a small set of structured primitives such as 'plan,' 'draft section,' 'integrate,' 'lint,' 'unit-test,' 'repair,' and 'finalize,' each with explicit pre- and post-conditions. Time is episodic and ends on a successful 'finalize,' an explicit 'handoff' to a human, or a safety timeout. This formulation forces us to think about who is allowed to speak, with what budget, and to whom, which in turn reduces chatter loops and makes downstream credit assignment feasible. It also clarifies how tools are modeled: a retrieval query, a unit test, or a linter is an environment effect that returns observable signals and leaves traces for later auditing. The outcome is a clean scaffold where collaboration is not just many messages but a sequence of purposeful steps with measurable impacts on global progress, quality, and cost.
Fig. 1. RL-Augmented LLM Team (CTDE + GRPO)
<details>
<summary>Image 1 Details</summary>

### Visual Description
## Flowchart: System Operational and Control Process
### Overview
The diagram illustrates a technical system workflow involving data acquisition, processing, resource allocation, and control signal generation. It depicts bidirectional data flows between physical systems, operational data, and computational processes, with feedback loops and parallel processing paths.
### Components/Axes
1. **Main Components**:
- Sending-End Physical System
- Operational Data (Hreq, D_i, req)
- System Identification and Linearization (A(θ))
- Wideband Analysis and Demand Calculation (M_i, I_i)
- Resource Mapping and Allocation
- Device Control and Allocation
- Control Signal
- POD (Proof of Delivery)
2. **Data Flows**:
- Arrows indicate sequential processing (e.g., "Sending-End Physical System" → "Operational Data")
- Dashed lines suggest optional or parallel processing paths
- Bidirectional feedback between "Wideband Analysis" and "System Identification"
### Detailed Analysis
1. **Left Pathway**:
- **Sending-End Physical System** → **Operational Data** (Hreq, D_i, req)
- **Operational Data** branches to:
- **Resource Mapping and Allocation**
- **POD** (Hreq, D_i, req)
2. **Right Pathway**:
- **System Identification and Linearization** (A(θ)) → **Wideband Analysis and Demand Calculation** (M_i, I_i)
- **Wideband Analysis** feeds back to **System Identification**
- **Wideband Analysis** → **Device Control and Allocation** → **Control Signal**
3. **Key Parameters**:
- **A(θ)**: System identification function with angular parameter θ
- **M_i, I_i**: Outputs from wideband analysis (likely matrix and intensity metrics)
- **Hreq, D_i, req**: Operational data inputs for POD
### Key Observations
1. **Parallel Processing**: Operational data splits into two streams (resource allocation vs. POD)
2. **Feedback Loop**: Wideband analysis iteratively refines system identification
3. **Control Signal Generation**: Final output depends on both resource allocation and device control processes
4. **Bidirectional Dependency**: System identification and wideband analysis form a closed-loop optimization system
### Interpretation
This diagram represents a closed-loop control system for managing physical infrastructure (e.g., power grids, communication networks). The system:
1. Acquires real-time operational data from physical endpoints
2. Uses system identification to model physical characteristics (A(θ))
3. Performs wideband analysis to calculate demand metrics (M_i, I_i)
4. Allocates resources based on both operational needs (POD) and demand calculations
5. Generates control signals through device-level allocation
The bidirectional feedback between system identification and wideband analysis suggests an adaptive optimization process, where demand calculations continuously refine system models. The parallel processing of operational data indicates the system handles multiple objectives simultaneously: immediate resource needs (POD) and long-term demand planning (wideband analysis).
Notably, the absence of explicit time-series data suggests this is a static process model rather than a dynamic simulation. The use of matrix notation (M_i) implies complex, multidimensional data processing in the wideband analysis stage.
</details>
## B. Centralized Training, Decentralized Execution - Why CTDE Fits LLM Workflows
Centralized training with decentralized execution is a natural fit for language agents because the expensive partlearning how to coordinate-benefits from seeing everything, while the runtime part must stay lean and privacy-aware. During training, a centralized critic inspects the full transcript, tool logs, and intermediate artifacts, building a calibrated sense of whether the team is moving toward completion or stuck in a loop. This global vantage point allows the critic to learn patterns that no single role can see: redundant reviews that add tokens but no value, premature coding before requirements stabilize, or over-eager planners who spawn tasks that nobody closes. At inference time we remove this omniscient view; each agent receives a bounded observation that includes the current artifact slice, a minimal cross-role summary, and a compact history window. The separation keeps prompts small, respects access controls, and limits leakage. We further enforce two practical devices: a soft handoff clock that nudges agents to pass work instead of hoarding it, and a message-budget token that forces roles to prioritize what to say. Together they create a rhythm: plan just enough, draft with discipline, review with purpose, test early, and converge without ping-pong. CTDE thus turns openended conversation into a controlled production pipeline where learning occurs with global information but execution remains local, efficient, and auditable.
## C. Group Relative Policy Optimization - Stable Credit for Teams
We extend standard policy optimization with a grouprelative baseline designed for teams. The core idea is simple: when judging one agent's contribution, do not compare it to a fixed average but to what the group would have achieved without that agent's current move. This leave-one-out perspective dampens variance and discourages blame-shifting. In language terms, if the reviewer adds a short, high-leverage comment that enables the coder to fix a failing test, the reviewer's decision is credited even if the overall reward arrives later at commit time. Conversely, if two agents repeat the same suggestion, the coordination cost increases and both see diminished gains. We keep the optimization conservative with clipped updates, modest entropy to sustain exploration, and a small Kullback-Leibler penalty to stay close to the supervised prior so that style and safety do not drift. The training batch mixes short and long episodes to avoid bias toward quick wins, and we randomize role orderings to prevent brittle conventions such as 'the planner always speaks first.' Over time the optimizer learns crisp patterns: planners constrain scope early, writers propose structured drafts instead of long monologues, reviewers focus on schema and factual alignment, coders run tests before large edits, and testers suggest targeted repairs instead of reopening design debates.
## D. Joint Reward Design - Compact, Normalized, and Actionable
A practical reward must be easy to compute, robust to noise, and aligned with operator goals. We therefore combine four signals. First is task quality: for writing, structure rationality and style consistency after a lightweight review pass; for coding, unit-test pass fraction with partial credit for near-misses.
Second is speed: wall-clock improvements normalized within a batch so teams that move faster under comparable conditions receive higher credit without incentivizing reckless shortcuts. Third is coordination efficiency: a gentle penalty for redundant turns, over-long messages, and cross-role conflicts that reopen settled decisions. Fourth is compliance and safety: negative marks for schema violations, unsafe tool calls, hallucinated citations, or style drift below a configured threshold. Each component is scaled to a similar range and normalized per batch to prevent any single signal from dominating as tasks vary in length. During early training we up-weight coordination to stop chatter early; later we shift mass to quality once the team learns to keep messages tight. The reward is also auditable: we persist per-turn fragments explaining what contributed to credit or penalty so that failures can be diagnosed and human policy can be updated. In ablations we find that removing the coordination term slows convergence and that normalizing by batch improves stability across prompt difficulties.
## E. Observation and Action Primitives - Small Interfaces, Big Effects
The observation design aims to be compact yet sufficient. Every role receives three slices: the brief (problem statement and constraints), the current artifact view (only the portion it must touch, such as a section outline or a code file diff), and a role-local memory that stores checklists, unresolved risks, and short notes. Cross-role visibility is mediated by a 'summary rail,' a rolling, human-readable synopsis that captures only decisions and blockers, not full messages. This rail keeps everyone aligned without inflating token budgets. On the action side, we intentionally restrict the verb set. 'Plan' creates or edits a short outline with owners and acceptance criteria. 'Draft section' or 'implement function' modifies only the assigned slice; large-scale refactors require an explicit 'propose change' followed by a review accept. 'Integrate' merges approved pieces and resolves conflicts with traceable choices. 'Lint' and 'test' trigger deterministic tools and report outcomes in a terse, machine-readable ledger. 'Repair' is allowed only after a failing lint or test to keep edits grounded in evidence. 'Finalize' produces a complete artifact with a conformance checklist. These primitives are easy to instrument, constrain token growth, and make it straightforward to compute rewards, because every turn either pushes the artifact forward or documents why it could not. The result is less verbosity, fewer ambiguous handoffs, and a log that practitioners can actually read.
## F. Implementation Details - Practical Training Loop and Systems Notes
We build policies from instruction-tuned backbones with lightweight adapters for role conditioning so that a single base model can serve multiple roles without duplicating parameters. The critic is smaller and focuses on the global transcript and tool traces, using pooled representations and simple attentional pooling to keep latency low. Training runs in mixed precision with gradient accumulation to reach effective batch sizes on modest hardware. We maintain a shared experience buffer that stores trajectories, tool outcomes, and concise self-critiques; these critiques help the critic learn what mattered without leaking private scratch content into the shared context. A curriculum grows episode length over time and interleaves writing and coding tasks, which prevents overfitting to either dialogue-heavy or tool-heavy regimes. We gate deployments with three safeguards: a budget cap per episode, a safety filter that screens tool calls and red-flags high-risk prompts, and a 'coach' monitor that can pause the team when it detects unproductive loops or style violations. For reproducibility we ship fixed seeds, prompt packs, and ablation switches, and we log per-role latency, token counts, and decision rationales so that cost accounting and error analysis are routine rather than ad hoc. In production-like runs we also shard transcripts, compress summaries, and cache tool outputs to keep costs predictable while preserving traceability.
## III. EXPERIMENT
## A. Benchmarks and Tasks - What We Measure and Why
We evaluate the framework on two complementary collaborative task families that jointly probe knowledge generation and verifiable execution. The first family is Collaborative Writing , consisting of 150 prompts spanning technical reports, research proposals, executive summaries, and end-user 'how-to' guides. Each prompt specifies genre, audience, and style constraints, and comes with a bounded retrieval pool to prevent unstructured information drift. This task stresses the team's ability to coordinate on structural organization, maintain style consistency, and keep factual alignment-e.g., whether the planner constrains scope early, whether the writer adheres to a shared outline, and whether the reviewer converges within limited rounds. The second family is Role-Split Coding , with 120 problems covering common data structures, string/array routines, lightweight API stub implementations, and targeted unit-test repair. This task emphasizes a closed loop from plan → implement → test → repair and reveals timing issues around 'who speaks first,' 'who freezes requirements,' and 'who triggers tools.' For both families we define strict termination conditions (either a passing outcome or a timeout) and preserve a full audit trail (per-turn action, tool receipts, and a succinct human-readable summary), enabling reproducible measurement of speed, quality, and coordination efficiency.
## B. Metrics and Protocols -Throughput, Quality, and Coordination
Our metric suite unifies throughput, quality, and cost . Throughput is normalized as processing speed (×) relative to a Single-LLM baseline; we also log message turns and total tokens as coordination costs. Quality in writing is measured by a composite of structural rationality and style consistency using a lightweight review form (headings hierarchy, paragraph granularity, terminology unification, and citation/format conformance); in coding, quality is measured by unit-test pass rate (%) and repair iterations to success. Cost is captured by wall-clock latency and token usage, under matched model budgets. For protocol control, each sample has a maximum dialogue length and wall-clock cap; all methods use the same backbone capacity and tool privileges. On writing, three expert raters provide majority adjudication; on coding, hidden unit tests and partial credit scoring reduce overfitting to test names or ordering. We preserve per-turn contribution notes, tool receipts, and 'why' summaries to support post-hoc attribution and faithful reproduction.
## C. Baselines and Controls - What We Compare Against
We compare three primary systems: Single LLM (one model executes plan and implementation end-to-end), AutoGen Team (two conversational agents with tools), and Proposed (GRPO) . To avoid confounds from capacity and call frequency, we hold the backbone size constant, cap concurrent tool calls, and enforce equal budget (calls and maximum context) across methods. We also run two controlled variants of our method: one removes the coordination-cost penalty (quality + speed only), and the other replaces the grouprelative baseline with a constant baseline to test creditassignment stability. Engineering-wise, we fix random seeds, mix short and long samples in batches, and interleave writing and coding tasks to avoid overfitting to either dialogue-heavy or tool-heavy regimes.
Fig. 2. Comparison Across Methods on Collaborative Tasks
<details>
<summary>Image 2 Details</summary>

### Visual Description
## Bar Chart: Comparison Across Methods on Collaborative Tasks
### Overview
The chart compares three methods (Single LLM, AutoGen Team, Proposed (GRPO)) across three metrics: Processing Speed (x), Writing Score (%), and Coding Pass Rate (%). Each method is represented by grouped bars, with values plotted on a y-axis ranging from 0 to 100.
### Components/Axes
- **X-axis**: Categories (methods) labeled as "Single LLM," "AutoGen Team," and "Proposed (GRPO)."
- **Y-axis**: "Value" (0–100), with no explicit unit for Processing Speed (x).
- **Legend**:
- Orange: Processing Speed (x)
- Blue: Writing Score (%)
- Green: Coding Pass Rate (%)
- **Bar Colors**:
- Orange bars (Processing Speed) are the shortest in each group.
- Blue bars (Writing Score) are the tallest in each group.
- Green bars (Coding Pass Rate) are intermediate in height.
### Detailed Analysis
- **Single LLM**:
- Processing Speed: 1.0x (orange bar, ~1.0)
- Writing Score: 90.1% (blue bar, ~90.1)
- Coding Pass Rate: 61.3% (green bar, ~61.3)
- **AutoGen Team**:
- Processing Speed: 1.8x (orange bar, ~1.8)
- Writing Score: 94.2% (blue bar, ~94.2)
- Coding Pass Rate: 68.1% (green bar, ~68.1)
- **Proposed (GRPO)**:
- Processing Speed: 3.0x (orange bar, ~3.0)
- Writing Score: 98.7% (blue bar, ~98.7)
- Coding Pass Rate: 74.6% (green bar, ~74.6)
### Key Observations
1. **Processing Speed** increases linearly across methods: 1.0x → 1.8x → 3.0x.
2. **Writing Score** improves steadily: 90.1% → 94.2% → 98.7%.
3. **Coding Pass Rate** also rises: 61.3% → 68.1% → 74.6%.
4. All metrics show the highest values for the "Proposed (GRPO)" method.
### Interpretation
The data demonstrates that the "Proposed (GRPO)" method outperforms both "Single LLM" and "AutoGen Team" across all evaluated metrics. The most significant improvement is in **Processing Speed**, which triples from the baseline (Single LLM) to the proposed method. **Writing Score** and **Coding Pass Rate** also show consistent gains, suggesting that GRPO enhances efficiency and effectiveness in collaborative tasks. The uniform upward trend across all metrics implies that GRPO optimizes both speed and quality, making it the most robust solution among the compared methods.
</details>
Fig. 3. Learning Curves on Writing and Coding Benchmarks
<details>
<summary>Image 3 Details</summary>

### Visual Description
## Line Graph: Performance Metrics Over Training Steps
### Overview
The image is a line graph comparing the performance of two methods ("Proposed (GRPO)" and "AutoGen Team") across two metrics: "Writing Quality" and "Coding Pass Rate" as training steps increase. The x-axis represents training steps (scaled by 1,000), and the y-axis represents scores/pass rates in percentage. Four lines are plotted, with distinct colors for each metric-method combination.
### Components/Axes
- **X-axis**: "Training Steps (x10³)" with markers at 0, 25, 50, 75, 100, 125, 150, 175, and 200.
- **Y-axis**: "Score / Pass Rate (%)" with markers at 50, 60, 70, 80, 90.
- **Legend**: Located in the bottom-right corner, with four entries:
1. **Orange**: Writing Quality — Proposed (GRPO)
2. **Blue**: Writing Quality — AutoGen Team
3. **Green**: Coding Pass Rate — Proposed (GRPO)
4. **Yellow**: Coding Pass Rate — AutoGen Team
### Detailed Analysis
1. **Writing Quality — Proposed (GRPO)** (Orange):
- Starts at ~80% at 0 steps.
- Rises sharply to ~95% by 200 steps.
- Slope: Steep upward trend, indicating rapid improvement.
2. **Writing Quality — AutoGen Team** (Blue):
- Starts at ~78% at 0 steps.
- Increases gradually to ~88% by 200 steps.
- Slope: Gentle upward trend, slower improvement than GRPO.
3. **Coding Pass Rate — Proposed (GRPO)** (Green):
- Starts at ~55% at 0 steps.
- Rises steadily to ~75% by 200 steps.
- Slope: Moderate upward trend, outperforming AutoGen Team.
4. **Coding Pass Rate — AutoGen Team** (Yellow):
- Starts at ~52% at 0 steps.
- Increases slowly to ~64% by 200 steps.
- Slope: Very gradual upward trend, minimal improvement.
### Key Observations
- **Performance Gaps**:
- Proposed (GRPO) consistently outperforms AutoGen Team in both metrics.
- Largest gap in "Writing Quality" (~7% at 200 steps).
- Smaller gap in "Coding Pass Rate" (~11% at 200 steps).
- **Trend Acceleration**:
- GRPO lines show steeper slopes, suggesting faster learning or optimization.
- AutoGen Team lines plateau earlier, indicating diminishing returns.
### Interpretation
The data demonstrates that the **Proposed (GRPO)** method significantly outperforms the **AutoGen Team** in both writing quality and coding pass rates. The steeper slopes of GRPO lines suggest it achieves higher efficiency in training, likely due to better optimization or architectural advantages. The AutoGen Team's gradual improvement implies reliance on slower, incremental learning. The smaller gap in coding pass rates may reflect inherent differences in task complexity or evaluation criteria. These results highlight GRPO's potential as a superior approach for tasks requiring rapid performance gains.
</details>
## D. Main Results - Team-Level Gains without Token Bloat
The headline outcomes appear in Table I and Fig. 2 . Under matched budgets, Proposed (GRPO) achieves 98.7% structure/style in Collaborative Writing and pushes processing speed to 3.0× ; in Role-Split Coding, the unit-test pass rate reaches 74.6% , outperforming both Single LLM and AutoGen Team. Crucially, the improvements do not come from 'talking more.' We see 20-35% fewer message turns and ~18-22% fewer tokens at similar or better quality, indicating that the group-relative optimization and coordination cost effectively suppress redundant chatter. The Paretod data is shown in table II and plot in Fig. 4 situates each method on the speedquality plane with bubble area proportional to tokens, showing our method consistently on a better front: higher quality, faster completion, and smaller token footprint-without inflating the number of agents or relying on brittle hand-written scripts.
<details>
<summary>Image 4 Details</summary>

### Visual Description
## Line Chart: Speed and Quality Across Methods (Line Chart; tokens annotated)
### Overview
The chart compares two metrics—**Processing Speed (x)** and **Writing Structure/Style (%)**—across three methods: **Single LLM**, **AutoGen Team**, and **Proposed (GRPO)**. Both metrics are plotted on a line chart with numerical values on the x-axis (Processing Speed) and percentage values on the y-axis (Writing Structure/Style). Tokens are annotated at each data point, indicating the relative scale of token usage.
---
### Components/Axes
- **X-axis (Horizontal)**:
- Label: **"Processing Speed (x)"**
- Scale: Numerical values from **1.00** to **3.00** (with intermediate markers at 1.75, 2.75).
- Categories: **Single LLM**, **AutoGen Team**, **Proposed (GRPO)** (positioned at 1.00, 1.75, and 2.75, respectively).
- **Y-axis (Vertical)**:
- Label: **"Writing Structure/Style (%)"**
- Scale: Percentage values from **90%** to **100%** (with intermediate markers at 92%, 94%, 96%, 98%).
- **Legend**:
- Located in the **top-left corner**.
- **Orange line**: **"Processing Speed (x)"**
- **Brown line**: **"Writing Structure/Style (%)"**
- **Annotations**:
- **Tokens**:
- **Single LLM**: "Tokens: 1.00x"
- **AutoGen Team**: "Tokens: 0.90x"
- **Proposed (GRPO)**: "Tokens: 0.78x"
---
### Detailed Analysis
#### Processing Speed (x) (Orange Line)
- **Single LLM**: 1.00 (lowest value).
- **AutoGen Team**: 1.75 (moderate increase).
- **Proposed (GRPO)**: 2.75 (highest value, approaching 3.00).
- **Trend**: Steadily increasing from left to right.
#### Writing Structure/Style (%) (Brown Line)
- **Single LLM**: 90% (lowest value).
- **AutoGen Team**: 92% (moderate improvement).
- **Proposed (GRPO)**: 98% (highest value, near 100%).
- **Trend**: Steadily increasing from left to right.
#### Tokens (Annotations)
- **Single LLM**: 1.00x (baseline).
- **AutoGen Team**: 0.90x (slight decrease).
- **Proposed (GRPO)**: 0.78x (further decrease).
- **Trend**: Decreasing as methods progress.
---
### Key Observations
1. **Positive Correlation**: Both **Processing Speed (x)** and **Writing Structure/Style (%)** increase with the progression from **Single LLM** to **Proposed (GRPO)**.
2. **Token Efficiency**: Token usage decreases as methods improve, suggesting higher efficiency in token utilization for advanced methods.
3. **Proposed (GRPO) Dominance**: This method achieves the highest values for both metrics, indicating superior performance.
---
### Interpretation
The chart demonstrates that the **Proposed (GRPO)** method outperforms the other two in both **processing speed** and **writing quality**. The **tokens** metric, however, shows a decline, which could imply that the Proposed method optimizes token usage while maintaining or improving performance. This suggests a trade-off between token efficiency and performance gains. The **AutoGen Team** method represents an intermediate step, showing moderate improvements over the **Single LLM** baseline. The **Single LLM** method serves as the reference point, with the lowest values for all metrics.
The data highlights the importance of method selection in balancing speed, quality, and resource efficiency. The **Proposed (GRPO)** method appears to be the most effective, though further analysis of token usage patterns would be necessary to fully understand its implications.
</details>
Method
Fig. 4. Speed and Quality Across Methods
Table I. Overall Results (mean ± s.e.m.)
| Method | Processing Speed (×) | Writing Structure/St yle (%) | Coding Pass Rate (%) | Turns |
|-----------------|------------------------|--------------------------------|------------------------|---------|
| Single LLM | 1.00 ± 0.05 | 90.1 ± 2.1 | 61.3 ± 2.8 | 14.2 |
| AutoGen Team | 1.80 ± 0.07 | 94.2 ± 1.9 | 68.1 ± 2.5 | 11.3 |
| Proposed (GRPO) | 3.00 ± 0.09 | 98.7 ± 0.7 | 74.6 ± 2.1 | 8.6 |
## E. Learning Dynamics - How Teams Improve Over Time
To visualize how teams progress from 'busy but inefficient' to 'concise and effective,' we track writing quality and coding pass rate versus training steps ( Fig. 3 ). GRPO quickly curbs unproductive exchanges, then steadily lifts structural/style scores and pass rates, continuing to consolidate with modest exploration in later stages. By contrast, AutoGen Team exhibits gains but shows late-stage volatility-'review repetition' and 'premature coding' oscillations are more frequent. These curves align with log-level observations: under GRPO, planners tend to freeze scope earlier, writers respect granularity constraints, reviewers focus on schema and factual alignment rather than re-raising resolved points; on the coding side, tests are triggered earlier and more frequently, and repairs are targeted to failing assertions rather than broad rewrites. In short, training elevates averages and also establishes a healthier team cadence.
Table II. GRPO Ablations on Writing (n=150 prompts)
| Variant | Struct./Style (%) | Turns | Speed (×) |
|--------------------------|---------------------|---------|-------------|
| Full GRPO (ours) | 98.7 | 8.3 | 3 |
| w/o Group Baseline | 96.9 | 9.8 | 2.5 |
| w/o Coord. Cost | 96.1 | 10.6 | 2.3 |
| Joint Reward →Local Only | 94 | 12.1 | 2 |
## F. Robustness and Breakdown - Task Types, Difficulty, and Failure Modes
We further break down performance by genre and difficulty ( Table III ). On the writing side, technical reports and research proposals achieve the highest structural scores thanks to clearer constraints; how-to guides place stricter demands on terminology consistency, where GRPO's early 'term freeze' reduces style drift. On the coding side, data structures and string routines benefit most because unit-test failure messages provide direct evidence for targeted repair; in API stub tasks, if the planner fails to freeze requirements promptly, rework increases. The principal failure modes we observe are (1) overplanning that delays execution, (2) review repetition that inflates turns, and (3) late testing that forces 'big repairs' near the end. Under GRPO, the frequencies of these modes drop by roughly 41% , 36% , and 29% , respectively, and when they do occur, episodes recover faster because the coach/monitor agent interrupts loops and re-anchors decisions to evidence.
Table III. Performance by Task Type and Difficulty
| Subset | Single LLM (Speed × / Qual. %) | AutoGen Team (× / %) | Proposed (× / %) |
|-----------------------------|----------------------------------|------------------------|--------------------|
| Writing -Tech Report (easy) | 1.05 / 91.2 | 1.86 / 95.1 | 3.08 / 99.0 |
| Writing- Proposal (medium) | 0.98 / 89.4 | 1.77 / 94.0 | 2.95 / 98.4 |
| Writing -How- to (hard) | 0.94 / 88.8 | 1.72 / 93.2 | 2.87 / 97.5 |
| Coding- DS/Strings (easy) | 1.02 / 63.0 | 1.85 / 69.5 | 3.05 / 76.9 |
| Coding -API Stubs (medium) | 0.97 / 60.2 | 1.74 / 67.0 | 2.92 / 73.8 |
| Coding -Mixed (hard) | 0.93 / 58.9 | 1.68 / 66.1 | 2.81 / 72.3 |
## G. Cost and Efficiency - Doing More with Less
We decompose end-to-end costs into thinking cost (tokens) and waiting cost (wall-clock) . At matched QoS, Proposed (GRPO) reduces wall-clock by ~60-70% versus Single LLM and ~40-50% versus AutoGen Team, while tokens drop by ~18-22% . Log inspection shows two major sources of savings: earlier outline freezing that prevents 'tear-downs,' and tight test-then-repair loops that ground edits in evidence rather than broad free-form debates. Fig. 4 plots all methods on the speedquality plane with bubble area indicating token cost, visually summarizing the Pareto advantage: higher quality, faster throughput, and smaller budgets-without adding more agents or relying on brittle, hand-crafted playbooks.
## IV. DISCUSSION
Our gains come from changing how the team works, not from letting models talk more. In writing, the planner locks a short outline early, the writer sticks to that shape, and the reviewer checks for structure and facts instead of nitpicking style. That alone removes a lot of churn. In coding, tests move to the front of the loop and fixes are tied to failing assertions, so edits are smaller and verification is immediate. You can see the effect in the logs: fewer long speeches, more commits that actually advance the artifact, and a steady climb in quality without a spike in tokens.
CTDE is the right fit for this. During training, a centralized critic sees the whole transcript and the tool traces, so it learns patterns a single role would miss-scope creep, repeated reviews, late testing. At inference, each role runs small: a tight view of the artifact, a short summary of decisions, and its own scratchpad. That constraint is a feature. It keeps prompts lean, reduces leakage, and forces clearer turns. GRPO then fixes the credit problem that usually plagues teams: we reward the marginal contribution relative to what the group would have done otherwise. Short, high-leverage moves get credit; redundant messages don't. The tone of the team shifts from debate to evidence.
There are limits. Structure helps, but very long documents or codebases can strain the observation slices, and noisy tools make credit messy. Evaluations on style still carry some subjectivity, even with rubrics and blind review. And our setup favors methods that use tokens well; with unlimited budgets, conversational baselines can close some of the gap. These are reasonable trade-offs for production settings, but they're worth calling out.
For deployment, start simple. Use templates that already have checklists, wrap tools so their outputs are short and machine-readable, cap budgets per episode, and keep a lightweight 'coach' to interrupt loops. Log the basics-turns, tokens, latency, why a decision was made-and review a sample of episodes each week to tune penalties and rails. In practice, teams end up talking less and finishing more. That's the point.
## V. CONCLUSION
We presented a reinforcement learning-augmented LLM framework for cooperative decision making. By casting collaboration as a Dec-POMDP and training with a centralized critic plus group-relative advantages, we improved throughput and quality on writing and coding tasks. The approach is simple to implement, compatible with standard PPO tooling, and practical for real pipelines. We anticipate broader impact in document production, software engineering, and operations where teams of agents must coordinate under partial information.
## REFERENCES
- [1] Papoudakis, G., Christianos, F., Schäfer, L., & Albrecht, S. V. (2021). Benchmarking Multi-Agent Deep Reinforcement Learning Algorithms. Journal of Artificial Intelligence Research, 72, 1-52. https://doi.org/10.1613/jair.1.12742
- [2] Schulman, J., Wolski, F., Dhariwal, P., Radford, A., & Klimov, O. Proximal Policy Optimization Algorithms. Advances in Neural Information Processing Systems (NeurIPS), 30, 2017.
- [3] Yu, C., Velu, A., Vinitsky, E., Wang, Y., Bayen, A., & Wu, Y. (2022). The Surprising Effectiveness of PPO in Cooperative Multi-Agent Games. Advances in Neural Information Processing Systems. https://doi.org/10.5555/3600270.3602057
- [4] Rashid, T., Samvelyan, M., Schroeder, C., Farquhar, G., Foerster, J., & Whiteson, S. 'QMIX: Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning,' Proceedings of the 35th International Conference on Machine Learning (ICML), 2018.
- [5] Foerster, J., Farquhar, G., Afouras, T., Nardelli, N., & Whiteson, S. (2018). Counterfactual Multi-Agent Policy Gradients. Proceedings of the AAAI Conference on Artificial Intelligence, 32(1). https://doi.org/10.1609/aaai.v32i1.11794
- [6] Lowe, R., Wu, Y., Tamar, A., Harb, J., Abbeel, P., & Mordatch, I. (2017). Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments. Advances in Neural Information Processing Systems, 30.https://doi.org/10.5555/3295222.3295385
- [7] Kuba, M., Schäfer, L., Albrecht, S. V., & Peters, J. (2022). Trust Region Policy Optimisation in Multi-Agent Reinforcement Learning. International Conference on Learning Representations.
- [8] Fu, J., Zhang, H., & Xu, D. (2022). Revisiting Some Common Practices in Cooperative Multi-Agent Reinforcement Learning. Proceedings of the 39th International Conference on Machine Learning. https://doi.org/10.5555/3524938.3525086
- [9] Wu, Z., Zhang, S., Chen, M., et al. (2023). AutoGen: Enabling Next-Gen LLM Applications via Multi-Agent Conversation Framework. arXiv:2308.08155
- [10] Hong, S., Zheng, L., Chen, Y., et al. (2023). MetaGPT: Meta Programming for A Multi-Agent Collaborative Framework. arXiv:2308.00352
- [11] Li, G., Hammoud, H. A., Itani, H., et al. (2023). AgentBench: Evaluating LLMs as Agents. arXiv:2308.03688
- [12] Ning, Z., & Xie, M. (2024). A Survey on Multi-Agent Reinforcement Learning: Challenges, Applications and Recent Advances. Journal of Automation and Intelligence.