# Reinforcement Learning-Augmented LLM Agents for Collaborative Decision Making and Performance Optimization
**Authors**: Dong Qiu, Duo Xu, Limengxi Yue
## Reinforcement Learning-Augmented LLM Agents for Collaborative Decision Making and Performance Optimization
Dong Qiu 1,* , Duo Xu 2a , Limengxi Yue 3b
New England College, Henniker, NH 03242, USA 2 Northeastern University, San Jose, CA 95113, USA 3 University of Massachusetts Amherst, Amherst, MA 01003, USA * DQiu\_GPS@nec.edu, a xu.duo3@northeastern.edu, b limengxiyue@outlook.com
Abstract -Large Language Models (LLMs) perform well in language tasks but often lack collaborative awareness and struggle to optimize global performance in multi-agent settings. We present a reinforcement learning-augmented LLM agent framework that formulates cooperation as a decentralized partially observable Markov decision process (Dec-POMDP) and adopts centralized training with decentralized execution (CTDE). We introduce Group Relative Policy Optimization (GRPO) to jointly optimize agent policies with access to global signals during training, together with a simplified joint reward that balances task quality, speed, and coordination cost. On collaborative writing and coding benchmarks, our framework delivers a 3× increase in task processing speed over single-agent baselines, 98.7% structural/style consistency in writing, and 74.6% test pass rate in coding. The approach consistently outperforms strong multi-agent LLM baselines and provides a practical path toward reliable collaboration in complex workflows.
Keyword s -multi-agent systems; reinforcement learning; large language models; Dec-POMDP; CTDE; policy optimization.
## I. INTRODUCTION
Large language models (LLMs) are increasingly used as agents that can plan, write code, call tools, and review intermediate outputs. In many practical tasks-such as iterative problem solving, collaborative content production, and software-oriented workflows-effective performance depends on coordinated behaviors among multiple specialized agents rather than a single monolithic model. This naturally aligns with a decentralized partially observable Markov decision process (Dec-POMDP), where each agent observes only part of the state and the team must act to optimize a shared objective.
Multi-agent reinforcement learning (MARL) offers a principled way to learn coordination policies, but it also introduces persistent challenges, including non-stationary learning dynamics, credit assignment ambiguity, and sensitivity to evaluation protocols. A large-scale benchmarking study highlights that reported gains can vary substantially with implementation details and experimental settings, making careful, reproducible comparisons essential when claiming improvements in cooperative MARL [1]. These concerns are especially relevant for LLM-based agent teams because actions are open-ended (language decisions and tool calls), feedback can be sparse or delayed, and success criteria are often defined by external evaluators (tests, metrics, or human-like preference signals).
Among policy-gradient methods, Proximal Policy Optimization (PPO) remains a widely adopted baseline due to its empirical stability and straightforward implementation [2]. Notably, PPO has been shown to perform strongly in cooperative multi-agent games when pipelines and baselines are controlled, indicating that stable on-policy optimization can be competitive even in multi-agent settings [3]. Trust-region style ideas have also been investigated for MARL to reduce destructive updates and improve learning robustness under changing joint policies. However, applying PPO-style optimization to LLM agent collaboration is not a direct translation: agent interactions are structured as multi-step dialogues and tool-mediated actions, and the training signal must reflect team-level outcomes while still allowing agents to execute with their own local information.
A common strategy for cooperative MARL is centralized training with decentralized execution (CTDE), where a centralized learner can use global information during training, but agents act using local observations at deployment. Value factorization approaches such as QMIX decompose the team value function into agent-wise utilities with a monotonic mixing constraint, enabling scalable learning in cooperative environments [4]. Counterfactual credit assignment, exemplified by COMA, improves attribution by comparing an agent's chosen action to counterfactual alternatives given the same joint context [5]. Centralized-critic actor-critic methods such as MADDPG further stabilize learning by training critics with joint observations/actions while retaining decentralized policies for execution [6]. These ideas are well established for structured action spaces, but for LLM agent teams they must be adapted to language-and-tool action trajectories and to evaluator-defined feedback.
In parallel, LLM-based multi-agent frameworks have emerged that coordinate agents via conversational protocols and role separation [7]. AutoGen demonstrates how multi-agent conversation combined with tool use can support complex multi-step tasks in practical applications [8],[9]. MetaGPT organizes multiple roles in a software-style workflow to produce artifacts through structured collaboration [10]. AgentBench provides benchmarks and evaluation setups to assess whether LLMs can act as agents in interactive environments, encouraging more systematic measurement beyond anecdotal demonstrations [11]. While these systems are effective in many cases, they are often driven primarily by prompting heuristics and fixed coordination rules, with limited use of a unified learning signal that directly optimizes longhorizon team performance under partial observability [12].
Motivated by these gaps, we develop a reinforcementlearning-augmented framework for collaborative LLM agents that connects (i) a Dec-POMDP formulation of team interaction, (ii) a CTDE-style centralized trainer, and (iii) a shared experience buffer that stores trajectories and evaluator feedback for policy improvement. Our design emphasizes evaluatordriven learning: global metrics and critique signals are converted into training targets that improve the overall team behavior, while each agent retains its own execution pathway at inference time. This connects established MARL principlesrobust baselines, stable policy optimization, and careful evaluation practice-to the emerging practice of role-based LLM agent collaboration.
## II. METHODOLOGY
## A. Problem Formulation - Dec-POMDP View of LLM Teams
We cast a team of LLM agents as a cooperative decision process under partial observability where each role-planner, writer, reviewer, coder, tester-must act on a limited, noisy view of the task while pursuing a shared objective and the structure is shown in Fig.1. In practice this means we separate what is globally relevant from what should remain local. The shared part includes the task brief, accepted snippets, approved facts from retrieval, and a compact changelog; the private part includes role-specific scratchpads, TODO lists, and short-term hypotheses that may be wrong and should not pollute the global context. Actions are not raw chat turns but a small set of structured primitives such as 'plan,' 'draft section,' 'integrate,' 'lint,' 'unit-test,' 'repair,' and 'finalize,' each with explicit pre- and post-conditions. Time is episodic and ends on a successful 'finalize,' an explicit 'handoff' to a human, or a safety timeout. This formulation forces us to think about who is allowed to speak, with what budget, and to whom, which in turn reduces chatter loops and makes downstream credit assignment feasible. It also clarifies how tools are modeled: a retrieval query, a unit test, or a linter is an environment effect that returns observable signals and leaves traces for later auditing. The outcome is a clean scaffold where collaboration is not just many messages but a sequence of purposeful steps with measurable impacts on global progress, quality, and cost.
Fig. 1. RL-Augmented LLM Team (CTDE + GRPO)
<details>
<summary>Image 1 Details</summary>

### Visual Description
\n
## Diagram: System Flow for Sending-End Physical System
### Overview
The image depicts a block diagram illustrating the flow of data and control signals within a system, likely related to communication or signal processing. The diagram shows a series of interconnected blocks representing different stages of processing, starting with a "Sending-End Physical System" and culminating in a "control signal".
### Components/Axes
The diagram consists of the following blocks:
* **Sending-End Physical System:** Top-left block.
* **System Identification and Linearization:** Two blocks positioned to the right of "Sending-End Physical System".
* **Wideband Analysis and Demand Calculation:** Central block, receiving input from "System Identification and Linearization" and "Sending-End Physical System".
* **Resource Mapping and Allocation:** Bottom-center block, receiving input from "Wideband Analysis and Demand Calculation".
* **Device Control and Allocation:** Bottom-right block, receiving input from "Resource Mapping and Allocation".
Additionally, the diagram includes the following labels representing data/signal flow:
* **(operational data)**: Flowing from "Sending-End Physical System" to "Wideband Analysis and Demand Calculation".
* **A(θ)**: Flowing from "System Identification and Linearization" to "Wideband Analysis and Demand Calculation".
* **(M₁, I₁)**: Flowing from "System Identification and Linearization" to "Wideband Analysis and Demand Calculation".
* **(Hreq, D₁, req)**: Flowing from "Wideband Analysis and Demand Calculation" to "Resource Mapping and Allocation".
* **POD**: Label associated with the flow from "Wideband Analysis and Demand Calculation" to "Resource Mapping and Allocation".
* **control signal**: Flowing from "Device Control and Allocation".
### Detailed Analysis or Content Details
The diagram illustrates a sequential process. The "Sending-End Physical System" generates "operational data" which is fed into the "Wideband Analysis and Demand Calculation" block. Simultaneously, the "System Identification and Linearization" block (appearing twice) provides inputs "A(θ)" and "(M₁, I₁)" to the "Wideband Analysis and Demand Calculation" block. The output of the "Wideband Analysis and Demand Calculation" block, labeled "(Hreq, D₁, req)" and "POD", is then passed to the "Resource Mapping and Allocation" block. Finally, the "Resource Mapping and Allocation" block sends a "control signal" to the "Device Control and Allocation" block.
### Key Observations
The diagram highlights a feedback-like structure, where system characteristics are identified and linearized to inform the analysis and demand calculation. The presence of two "System Identification and Linearization" blocks suggests either redundancy or parallel processing of different aspects of system identification. The labels within parentheses indicate specific data or parameters being transmitted between blocks.
### Interpretation
This diagram likely represents a control system for a physical system, possibly a transmitter or communication device. The "Sending-End Physical System" represents the actual hardware, while the subsequent blocks perform signal processing, resource allocation, and control functions. The "System Identification and Linearization" blocks are crucial for adapting the control strategy to the specific characteristics of the physical system. The "Wideband Analysis and Demand Calculation" block likely determines the required resources (bandwidth, power, etc.) based on the input signal and system conditions. The "Resource Mapping and Allocation" block then assigns these resources, and the "Device Control and Allocation" block implements the control actions. The diagram suggests a closed-loop control system where the system's performance is continuously monitored and adjusted to meet the desired specifications. The use of mathematical notations like A(θ), M₁, I₁, Hreq, D₁, and req indicates a quantitative approach to system analysis and control. The diagram does not provide specific numerical data or performance metrics, but it clearly outlines the functional relationships between different components of the system.
</details>
## B. Centralized Training, Decentralized Execution - Why CTDE Fits LLM Workflows
Centralized training with decentralized execution is a natural fit for language agents because the expensive partlearning how to coordinate-benefits from seeing everything, while the runtime part must stay lean and privacy-aware. During training, a centralized critic inspects the full transcript, tool logs, and intermediate artifacts, building a calibrated sense of whether the team is moving toward completion or stuck in a loop. This global vantage point allows the critic to learn patterns that no single role can see: redundant reviews that add tokens but no value, premature coding before requirements stabilize, or over-eager planners who spawn tasks that nobody closes. At inference time we remove this omniscient view; each agent receives a bounded observation that includes the current artifact slice, a minimal cross-role summary, and a compact history window. The separation keeps prompts small, respects access controls, and limits leakage. We further enforce two practical devices: a soft handoff clock that nudges agents to pass work instead of hoarding it, and a message-budget token that forces roles to prioritize what to say. Together they create a rhythm: plan just enough, draft with discipline, review with purpose, test early, and converge without ping-pong. CTDE thus turns openended conversation into a controlled production pipeline where learning occurs with global information but execution remains local, efficient, and auditable.
## C. Group Relative Policy Optimization - Stable Credit for Teams
We extend standard policy optimization with a grouprelative baseline designed for teams. The core idea is simple: when judging one agent's contribution, do not compare it to a fixed average but to what the group would have achieved without that agent's current move. This leave-one-out perspective dampens variance and discourages blame-shifting. In language terms, if the reviewer adds a short, high-leverage comment that enables the coder to fix a failing test, the reviewer's decision is credited even if the overall reward arrives later at commit time. Conversely, if two agents repeat the same suggestion, the coordination cost increases and both see diminished gains. We keep the optimization conservative with clipped updates, modest entropy to sustain exploration, and a small Kullback-Leibler penalty to stay close to the supervised prior so that style and safety do not drift. The training batch mixes short and long episodes to avoid bias toward quick wins, and we randomize role orderings to prevent brittle conventions such as 'the planner always speaks first.' Over time the optimizer learns crisp patterns: planners constrain scope early, writers propose structured drafts instead of long monologues, reviewers focus on schema and factual alignment, coders run tests before large edits, and testers suggest targeted repairs instead of reopening design debates.
## D. Joint Reward Design - Compact, Normalized, and Actionable
A practical reward must be easy to compute, robust to noise, and aligned with operator goals. We therefore combine four signals. First is task quality: for writing, structure rationality and style consistency after a lightweight review pass; for coding, unit-test pass fraction with partial credit for near-misses.
Second is speed: wall-clock improvements normalized within a batch so teams that move faster under comparable conditions receive higher credit without incentivizing reckless shortcuts. Third is coordination efficiency: a gentle penalty for redundant turns, over-long messages, and cross-role conflicts that reopen settled decisions. Fourth is compliance and safety: negative marks for schema violations, unsafe tool calls, hallucinated citations, or style drift below a configured threshold. Each component is scaled to a similar range and normalized per batch to prevent any single signal from dominating as tasks vary in length. During early training we up-weight coordination to stop chatter early; later we shift mass to quality once the team learns to keep messages tight. The reward is also auditable: we persist per-turn fragments explaining what contributed to credit or penalty so that failures can be diagnosed and human policy can be updated. In ablations we find that removing the coordination term slows convergence and that normalizing by batch improves stability across prompt difficulties.
## E. Observation and Action Primitives - Small Interfaces, Big Effects
The observation design aims to be compact yet sufficient. Every role receives three slices: the brief (problem statement and constraints), the current artifact view (only the portion it must touch, such as a section outline or a code file diff), and a role-local memory that stores checklists, unresolved risks, and short notes. Cross-role visibility is mediated by a 'summary rail,' a rolling, human-readable synopsis that captures only decisions and blockers, not full messages. This rail keeps everyone aligned without inflating token budgets. On the action side, we intentionally restrict the verb set. 'Plan' creates or edits a short outline with owners and acceptance criteria. 'Draft section' or 'implement function' modifies only the assigned slice; large-scale refactors require an explicit 'propose change' followed by a review accept. 'Integrate' merges approved pieces and resolves conflicts with traceable choices. 'Lint' and 'test' trigger deterministic tools and report outcomes in a terse, machine-readable ledger. 'Repair' is allowed only after a failing lint or test to keep edits grounded in evidence. 'Finalize' produces a complete artifact with a conformance checklist. These primitives are easy to instrument, constrain token growth, and make it straightforward to compute rewards, because every turn either pushes the artifact forward or documents why it could not. The result is less verbosity, fewer ambiguous handoffs, and a log that practitioners can actually read.
## F. Implementation Details - Practical Training Loop and Systems Notes
We build policies from instruction-tuned backbones with lightweight adapters for role conditioning so that a single base model can serve multiple roles without duplicating parameters. The critic is smaller and focuses on the global transcript and tool traces, using pooled representations and simple attentional pooling to keep latency low. Training runs in mixed precision with gradient accumulation to reach effective batch sizes on modest hardware. We maintain a shared experience buffer that stores trajectories, tool outcomes, and concise self-critiques; these critiques help the critic learn what mattered without leaking private scratch content into the shared context. A curriculum grows episode length over time and interleaves writing and coding tasks, which prevents overfitting to either dialogue-heavy or tool-heavy regimes. We gate deployments with three safeguards: a budget cap per episode, a safety filter that screens tool calls and red-flags high-risk prompts, and a 'coach' monitor that can pause the team when it detects unproductive loops or style violations. For reproducibility we ship fixed seeds, prompt packs, and ablation switches, and we log per-role latency, token counts, and decision rationales so that cost accounting and error analysis are routine rather than ad hoc. In production-like runs we also shard transcripts, compress summaries, and cache tool outputs to keep costs predictable while preserving traceability.
## III. EXPERIMENT
## A. Benchmarks and Tasks - What We Measure and Why
We evaluate the framework on two complementary collaborative task families that jointly probe knowledge generation and verifiable execution. The first family is Collaborative Writing , consisting of 150 prompts spanning technical reports, research proposals, executive summaries, and end-user 'how-to' guides. Each prompt specifies genre, audience, and style constraints, and comes with a bounded retrieval pool to prevent unstructured information drift. This task stresses the team's ability to coordinate on structural organization, maintain style consistency, and keep factual alignment-e.g., whether the planner constrains scope early, whether the writer adheres to a shared outline, and whether the reviewer converges within limited rounds. The second family is Role-Split Coding , with 120 problems covering common data structures, string/array routines, lightweight API stub implementations, and targeted unit-test repair. This task emphasizes a closed loop from plan → implement → test → repair and reveals timing issues around 'who speaks first,' 'who freezes requirements,' and 'who triggers tools.' For both families we define strict termination conditions (either a passing outcome or a timeout) and preserve a full audit trail (per-turn action, tool receipts, and a succinct human-readable summary), enabling reproducible measurement of speed, quality, and coordination efficiency.
## B. Metrics and Protocols -Throughput, Quality, and Coordination
Our metric suite unifies throughput, quality, and cost . Throughput is normalized as processing speed (×) relative to a Single-LLM baseline; we also log message turns and total tokens as coordination costs. Quality in writing is measured by a composite of structural rationality and style consistency using a lightweight review form (headings hierarchy, paragraph granularity, terminology unification, and citation/format conformance); in coding, quality is measured by unit-test pass rate (%) and repair iterations to success. Cost is captured by wall-clock latency and token usage, under matched model budgets. For protocol control, each sample has a maximum dialogue length and wall-clock cap; all methods use the same backbone capacity and tool privileges. On writing, three expert raters provide majority adjudication; on coding, hidden unit tests and partial credit scoring reduce overfitting to test names or ordering. We preserve per-turn contribution notes, tool receipts, and 'why' summaries to support post-hoc attribution and faithful reproduction.
## C. Baselines and Controls - What We Compare Against
We compare three primary systems: Single LLM (one model executes plan and implementation end-to-end), AutoGen Team (two conversational agents with tools), and Proposed (GRPO) . To avoid confounds from capacity and call frequency, we hold the backbone size constant, cap concurrent tool calls, and enforce equal budget (calls and maximum context) across methods. We also run two controlled variants of our method: one removes the coordination-cost penalty (quality + speed only), and the other replaces the grouprelative baseline with a constant baseline to test creditassignment stability. Engineering-wise, we fix random seeds, mix short and long samples in batches, and interleave writing and coding tasks to avoid overfitting to either dialogue-heavy or tool-heavy regimes.
Fig. 2. Comparison Across Methods on Collaborative Tasks
<details>
<summary>Image 2 Details</summary>

### Visual Description
\n
## Bar Chart: Comparison Across Methods on Collaborative Tasks
### Overview
This image presents a bar chart comparing three methods – Single LLM, AutoGen Team, and Proposed (GRPO) – across three metrics: Processing Speed, Writing Score, and Coding Pass Rate. The chart uses bar heights to represent values, with each method having three bars corresponding to the three metrics.
### Components/Axes
* **Title:** "Fig. 2. Comparison Across Methods on Collaborative Tasks" (Top-center)
* **X-axis:** Methods – "Single LLM", "AutoGen Team", "Proposed (GRPO)" (Bottom-center)
* **Y-axis:** "Value" (Left-center), ranging from 0 to 100, with increments of 20.
* **Legend:** Located in the top-right corner.
* Orange: "Processing Speed (x)"
* Blue: "Writing Score (%)"
* Green: "Coding Pass Rate (%)"
### Detailed Analysis
The chart consists of three groups of bars, one for each method. Within each group, there are three bars representing the three metrics.
**Single LLM:**
* Processing Speed (Orange): Approximately 1.0
* Writing Score (Blue): Approximately 90.1
* Coding Pass Rate (Green): Approximately 61.3
**AutoGen Team:**
* Processing Speed (Orange): Approximately 1.8
* Writing Score (Blue): Approximately 94.2
* Coding Pass Rate (Green): Approximately 68.1
**Proposed (GRPO):**
* Processing Speed (Orange): Approximately 3.0
* Writing Score (Blue): Approximately 98.7
* Coding Pass Rate (Green): Approximately 74.6
The bars for each metric within a method group are visually distinct based on the legend colors. The Writing Score bars are consistently the highest for all three methods. The Coding Pass Rate bars are the lowest. The Processing Speed bars are significantly lower than the other two metrics.
### Key Observations
* The "Proposed (GRPO)" method consistently outperforms the other two methods across all three metrics.
* The "AutoGen Team" method shows improvement over the "Single LLM" method in all metrics.
* Processing Speed is significantly lower than both Writing Score and Coding Pass Rate for all methods.
* The difference in Writing Score between "Single LLM" and "AutoGen Team" is relatively small, but the difference between "AutoGen Team" and "Proposed (GRPO)" is more substantial.
* The Coding Pass Rate shows the most significant improvement from "Single LLM" to "Proposed (GRPO)".
### Interpretation
The data suggests that the "Proposed (GRPO)" method is the most effective approach for collaborative tasks, demonstrating superior performance in processing speed, writing quality, and coding accuracy. The improvement from "Single LLM" to "AutoGen Team" indicates the benefits of utilizing a team-based approach. The substantial gains achieved by "Proposed (GRPO)" suggest that the specific techniques employed in this method are particularly well-suited for enhancing collaborative task performance.
The low values for Processing Speed relative to the other metrics might indicate that the speed is a limiting factor, or that the other metrics are more important for the overall success of the tasks. The increasing Coding Pass Rate across the methods suggests that the methods are becoming more reliable in generating correct code. The consistent high Writing Score indicates that all methods are capable of producing high-quality written content.
The chart provides a clear visual comparison of the strengths and weaknesses of each method, allowing for informed decision-making regarding the selection of the most appropriate approach for collaborative tasks.
</details>
Fig. 3. Learning Curves on Writing and Coding Benchmarks
<details>
<summary>Image 3 Details</summary>

### Visual Description
\n
## Line Chart: Performance Comparison of Proposed (GRPO) and AutoGen Team
### Overview
This line chart compares the performance of a "Proposed (GRPO)" method against an "AutoGen Team" method across two metrics: "Writing Quality" and "Coding Pass Rate". Performance is measured as a percentage ("Score / Pass Rate (%)") over "Training Steps" (expressed as multiples of 10^3).
### Components/Axes
* **X-axis:** Training Steps (x10^3), ranging from 0 to 200, with tick marks at 0, 25, 50, 75, 100, 125, 150, and 200.
* **Y-axis:** Score / Pass Rate (%), ranging from 50% to 90%, with tick marks at 50, 60, 70, 80, and 90.
* **Legend:** Located in the bottom-right corner. Contains the following entries:
* Yellow Line: "Writing Quality – Proposed (GRPO)"
* Blue Line: "Writing Quality – AutoGen Team"
* Light Green Line: "Coding Pass Rate – Proposed (GRPO)"
* Light Blue Line: "Coding Pass Rate – AutoGen Team"
### Detailed Analysis
**Writing Quality – Proposed (GRPO) (Yellow Line):**
The yellow line shows an upward trend, starting at approximately 68% at 0 training steps. It increases rapidly initially, then plateaus.
* At 25 training steps: ~75%
* At 50 training steps: ~82%
* At 75 training steps: ~87%
* At 100 training steps: ~89%
* At 125 training steps: ~90%
* At 150 training steps: ~91%
* At 200 training steps: ~91%
**Writing Quality – AutoGen Team (Blue Line):**
The blue line also shows an upward trend, but it is less steep than the yellow line. It starts at approximately 62% at 0 training steps.
* At 25 training steps: ~66%
* At 50 training steps: ~72%
* At 75 training steps: ~76%
* At 100 training steps: ~78%
* At 125 training steps: ~80%
* At 150 training steps: ~82%
* At 200 training steps: ~84%
**Coding Pass Rate – Proposed (GRPO) (Light Green Line):**
The light green line shows a significant upward trend, starting at approximately 58% at 0 training steps. It increases rapidly and then begins to level off.
* At 25 training steps: ~64%
* At 50 training steps: ~72%
* At 75 training steps: ~78%
* At 100 training steps: ~82%
* At 125 training steps: ~85%
* At 150 training steps: ~87%
* At 200 training steps: ~88%
**Coding Pass Rate – AutoGen Team (Light Blue Line):**
The light blue line shows an upward trend, but it is less pronounced than the light green line. It starts at approximately 55% at 0 training steps.
* At 25 training steps: ~59%
* At 50 training steps: ~65%
* At 75 training steps: ~70%
* At 100 training steps: ~74%
* At 125 training steps: ~76%
* At 150 training steps: ~78%
* At 200 training steps: ~80%
### Key Observations
* The "Proposed (GRPO)" method consistently outperforms the "AutoGen Team" method for both "Writing Quality" and "Coding Pass Rate".
* The "Coding Pass Rate" shows a larger performance gap between the two methods than "Writing Quality".
* Both metrics exhibit diminishing returns with increasing training steps, indicating that the performance gains start to plateau after a certain point.
* The "Proposed (GRPO)" method reaches a higher plateau for both metrics.
### Interpretation
The data suggests that the "Proposed (GRPO)" method is more effective than the "AutoGen Team" method in both writing quality and coding pass rate. The larger performance difference in coding pass rate indicates that the "Proposed (GRPO)" method may be particularly beneficial for tasks requiring higher accuracy and reliability. The diminishing returns observed with increasing training steps suggest that there is an optimal training duration beyond which further training does not significantly improve performance. This could be due to the model reaching its capacity or the training data becoming saturated. The chart provides a quantitative comparison of the two methods, allowing for informed decision-making regarding which method to employ based on the specific requirements of the task. The consistent outperformance of the "Proposed (GRPO)" method suggests a superior underlying algorithm or training strategy.
</details>
## D. Main Results - Team-Level Gains without Token Bloat
The headline outcomes appear in Table I and Fig. 2 . Under matched budgets, Proposed (GRPO) achieves 98.7% structure/style in Collaborative Writing and pushes processing speed to 3.0× ; in Role-Split Coding, the unit-test pass rate reaches 74.6% , outperforming both Single LLM and AutoGen Team. Crucially, the improvements do not come from 'talking more.' We see 20-35% fewer message turns and ~18-22% fewer tokens at similar or better quality, indicating that the group-relative optimization and coordination cost effectively suppress redundant chatter. The Paretod data is shown in table II and plot in Fig. 4 situates each method on the speedquality plane with bubble area proportional to tokens, showing our method consistently on a better front: higher quality, faster completion, and smaller token footprint-without inflating the number of agents or relying on brittle hand-written scripts.
<details>
<summary>Image 4 Details</summary>

### Visual Description
\n
## Line Chart: Speed and Quality Across Methods
### Overview
This image presents a line chart comparing the Processing Speed and Writing Structure/Style across three methods: Single LLM, AutoGen Team, and Proposed (GRPO). The chart uses two y-axes to represent these metrics, with the x-axis representing the method. The chart also includes annotations indicating the number of tokens used for each method.
### Components/Axes
* **Title:** "Fig. 4. Speed and Quality Across Methods (Line Chart; tokens annotated)"
* **X-axis:** "Method" with categories: "Single LLM", "AutoGen Team", "Proposed (GRPO)"
* **Left Y-axis:** "Processing Speed (x)" ranging from 1.00 to 3.00, with increments of 0.25.
* **Right Y-axis:** "Writing Structure/Style (%)" ranging from 90 to 98, with increments of 2.
* **Legend:** Located in the top-left corner.
* "Processing Speed (x)" - represented by a solid gold line.
* "Writing Structure/Style (%)" - represented by a dashed gold line.
* **Annotations:**
* "Tokens: 1.00x" positioned near "Single LLM".
* "Tokens: 0.90x" positioned near "AutoGen Team".
* "Tokens: 0.78x" positioned near "Proposed (GRPO)".
### Detailed Analysis
* **Processing Speed (x):**
* The line slopes upward from left to right.
* At "Single LLM", the processing speed is approximately 1.00x.
* At "AutoGen Team", the processing speed is approximately 1.75x.
* At "Proposed (GRPO)", the processing speed is approximately 3.00x.
* **Writing Structure/Style (%):**
* The line also slopes upward from left to right, but with a less steep gradient than the processing speed line.
* At "Single LLM", the writing structure/style is approximately 92%.
* At "AutoGen Team", the writing structure/style is approximately 94%.
* At "Proposed (GRPO)", the writing structure/style is approximately 96%.
### Key Observations
* The "Proposed (GRPO)" method demonstrates the highest processing speed and writing structure/style.
* The "AutoGen Team" method shows a significant improvement in processing speed compared to the "Single LLM" method, with a moderate improvement in writing structure/style.
* The number of tokens used decreases as the method progresses from "Single LLM" to "Proposed (GRPO)".
### Interpretation
The chart suggests that the "Proposed (GRPO)" method offers the best performance in terms of both processing speed and writing quality, while also being the most efficient in terms of token usage. The increasing trend of both metrics indicates a positive correlation between the method and its performance. The decreasing token usage suggests that the "Proposed (GRPO)" method is more optimized for resource utilization. The difference in slope between the two lines indicates that processing speed improves at a faster rate than writing structure/style as the method changes. This could imply that the "Proposed (GRPO)" method prioritizes speed over stylistic refinement, or that the stylistic improvements are less pronounced than the speed gains. The annotations regarding token usage provide additional context, suggesting that the improved performance of the "Proposed (GRPO)" method is achieved with fewer tokens, indicating greater efficiency.
</details>
Method
Fig. 4. Speed and Quality Across Methods
Table I. Overall Results (mean ± s.e.m.)
| Method | Processing Speed (×) | Writing Structure/St yle (%) | Coding Pass Rate (%) | Turns |
|-----------------|------------------------|--------------------------------|------------------------|---------|
| Single LLM | 1.00 ± 0.05 | 90.1 ± 2.1 | 61.3 ± 2.8 | 14.2 |
| AutoGen Team | 1.80 ± 0.07 | 94.2 ± 1.9 | 68.1 ± 2.5 | 11.3 |
| Proposed (GRPO) | 3.00 ± 0.09 | 98.7 ± 0.7 | 74.6 ± 2.1 | 8.6 |
## E. Learning Dynamics - How Teams Improve Over Time
To visualize how teams progress from 'busy but inefficient' to 'concise and effective,' we track writing quality and coding pass rate versus training steps ( Fig. 3 ). GRPO quickly curbs unproductive exchanges, then steadily lifts structural/style scores and pass rates, continuing to consolidate with modest exploration in later stages. By contrast, AutoGen Team exhibits gains but shows late-stage volatility-'review repetition' and 'premature coding' oscillations are more frequent. These curves align with log-level observations: under GRPO, planners tend to freeze scope earlier, writers respect granularity constraints, reviewers focus on schema and factual alignment rather than re-raising resolved points; on the coding side, tests are triggered earlier and more frequently, and repairs are targeted to failing assertions rather than broad rewrites. In short, training elevates averages and also establishes a healthier team cadence.
Table II. GRPO Ablations on Writing (n=150 prompts)
| Variant | Struct./Style (%) | Turns | Speed (×) |
|--------------------------|---------------------|---------|-------------|
| Full GRPO (ours) | 98.7 | 8.3 | 3 |
| w/o Group Baseline | 96.9 | 9.8 | 2.5 |
| w/o Coord. Cost | 96.1 | 10.6 | 2.3 |
| Joint Reward →Local Only | 94 | 12.1 | 2 |
## F. Robustness and Breakdown - Task Types, Difficulty, and Failure Modes
We further break down performance by genre and difficulty ( Table III ). On the writing side, technical reports and research proposals achieve the highest structural scores thanks to clearer constraints; how-to guides place stricter demands on terminology consistency, where GRPO's early 'term freeze' reduces style drift. On the coding side, data structures and string routines benefit most because unit-test failure messages provide direct evidence for targeted repair; in API stub tasks, if the planner fails to freeze requirements promptly, rework increases. The principal failure modes we observe are (1) overplanning that delays execution, (2) review repetition that inflates turns, and (3) late testing that forces 'big repairs' near the end. Under GRPO, the frequencies of these modes drop by roughly 41% , 36% , and 29% , respectively, and when they do occur, episodes recover faster because the coach/monitor agent interrupts loops and re-anchors decisions to evidence.
Table III. Performance by Task Type and Difficulty
| Subset | Single LLM (Speed × / Qual. %) | AutoGen Team (× / %) | Proposed (× / %) |
|-----------------------------|----------------------------------|------------------------|--------------------|
| Writing -Tech Report (easy) | 1.05 / 91.2 | 1.86 / 95.1 | 3.08 / 99.0 |
| Writing- Proposal (medium) | 0.98 / 89.4 | 1.77 / 94.0 | 2.95 / 98.4 |
| Writing -How- to (hard) | 0.94 / 88.8 | 1.72 / 93.2 | 2.87 / 97.5 |
| Coding- DS/Strings (easy) | 1.02 / 63.0 | 1.85 / 69.5 | 3.05 / 76.9 |
| Coding -API Stubs (medium) | 0.97 / 60.2 | 1.74 / 67.0 | 2.92 / 73.8 |
| Coding -Mixed (hard) | 0.93 / 58.9 | 1.68 / 66.1 | 2.81 / 72.3 |
## G. Cost and Efficiency - Doing More with Less
We decompose end-to-end costs into thinking cost (tokens) and waiting cost (wall-clock) . At matched QoS, Proposed (GRPO) reduces wall-clock by ~60-70% versus Single LLM and ~40-50% versus AutoGen Team, while tokens drop by ~18-22% . Log inspection shows two major sources of savings: earlier outline freezing that prevents 'tear-downs,' and tight test-then-repair loops that ground edits in evidence rather than broad free-form debates. Fig. 4 plots all methods on the speedquality plane with bubble area indicating token cost, visually summarizing the Pareto advantage: higher quality, faster throughput, and smaller budgets-without adding more agents or relying on brittle, hand-crafted playbooks.
## IV. DISCUSSION
Our gains come from changing how the team works, not from letting models talk more. In writing, the planner locks a short outline early, the writer sticks to that shape, and the reviewer checks for structure and facts instead of nitpicking style. That alone removes a lot of churn. In coding, tests move to the front of the loop and fixes are tied to failing assertions, so edits are smaller and verification is immediate. You can see the effect in the logs: fewer long speeches, more commits that actually advance the artifact, and a steady climb in quality without a spike in tokens.
CTDE is the right fit for this. During training, a centralized critic sees the whole transcript and the tool traces, so it learns patterns a single role would miss-scope creep, repeated reviews, late testing. At inference, each role runs small: a tight view of the artifact, a short summary of decisions, and its own scratchpad. That constraint is a feature. It keeps prompts lean, reduces leakage, and forces clearer turns. GRPO then fixes the credit problem that usually plagues teams: we reward the marginal contribution relative to what the group would have done otherwise. Short, high-leverage moves get credit; redundant messages don't. The tone of the team shifts from debate to evidence.
There are limits. Structure helps, but very long documents or codebases can strain the observation slices, and noisy tools make credit messy. Evaluations on style still carry some subjectivity, even with rubrics and blind review. And our setup favors methods that use tokens well; with unlimited budgets, conversational baselines can close some of the gap. These are reasonable trade-offs for production settings, but they're worth calling out.
For deployment, start simple. Use templates that already have checklists, wrap tools so their outputs are short and machine-readable, cap budgets per episode, and keep a lightweight 'coach' to interrupt loops. Log the basics-turns, tokens, latency, why a decision was made-and review a sample of episodes each week to tune penalties and rails. In practice, teams end up talking less and finishing more. That's the point.
## V. CONCLUSION
We presented a reinforcement learning-augmented LLM framework for cooperative decision making. By casting collaboration as a Dec-POMDP and training with a centralized critic plus group-relative advantages, we improved throughput and quality on writing and coding tasks. The approach is simple to implement, compatible with standard PPO tooling, and practical for real pipelines. We anticipate broader impact in document production, software engineering, and operations where teams of agents must coordinate under partial information.
## REFERENCES
- [1] Papoudakis, G., Christianos, F., Schäfer, L., & Albrecht, S. V. (2021). Benchmarking Multi-Agent Deep Reinforcement Learning Algorithms. Journal of Artificial Intelligence Research, 72, 1-52. https://doi.org/10.1613/jair.1.12742
- [2] Schulman, J., Wolski, F., Dhariwal, P., Radford, A., & Klimov, O. Proximal Policy Optimization Algorithms. Advances in Neural Information Processing Systems (NeurIPS), 30, 2017.
- [3] Yu, C., Velu, A., Vinitsky, E., Wang, Y., Bayen, A., & Wu, Y. (2022). The Surprising Effectiveness of PPO in Cooperative Multi-Agent Games. Advances in Neural Information Processing Systems. https://doi.org/10.5555/3600270.3602057
- [4] Rashid, T., Samvelyan, M., Schroeder, C., Farquhar, G., Foerster, J., & Whiteson, S. 'QMIX: Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning,' Proceedings of the 35th International Conference on Machine Learning (ICML), 2018.
- [5] Foerster, J., Farquhar, G., Afouras, T., Nardelli, N., & Whiteson, S. (2018). Counterfactual Multi-Agent Policy Gradients. Proceedings of the AAAI Conference on Artificial Intelligence, 32(1). https://doi.org/10.1609/aaai.v32i1.11794
- [6] Lowe, R., Wu, Y., Tamar, A., Harb, J., Abbeel, P., & Mordatch, I. (2017). Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments. Advances in Neural Information Processing Systems, 30.https://doi.org/10.5555/3295222.3295385
- [7] Kuba, M., Schäfer, L., Albrecht, S. V., & Peters, J. (2022). Trust Region Policy Optimisation in Multi-Agent Reinforcement Learning. International Conference on Learning Representations.
- [8] Fu, J., Zhang, H., & Xu, D. (2022). Revisiting Some Common Practices in Cooperative Multi-Agent Reinforcement Learning. Proceedings of the 39th International Conference on Machine Learning. https://doi.org/10.5555/3524938.3525086
- [9] Wu, Z., Zhang, S., Chen, M., et al. (2023). AutoGen: Enabling Next-Gen LLM Applications via Multi-Agent Conversation Framework. arXiv:2308.08155
- [10] Hong, S., Zheng, L., Chen, Y., et al. (2023). MetaGPT: Meta Programming for A Multi-Agent Collaborative Framework. arXiv:2308.00352
- [11] Li, G., Hammoud, H. A., Itani, H., et al. (2023). AgentBench: Evaluating LLMs as Agents. arXiv:2308.03688
- [12] Ning, Z., & Xie, M. (2024). A Survey on Multi-Agent Reinforcement Learning: Challenges, Applications and Recent Advances. Journal of Automation and Intelligence.