2601.08129v3
Model: healer-alpha-free
# Emergent Coordination in Multi-Agent Systems via Pressure Fields and Temporal Decay
[1] Roland R. Rodriguez, Jr.
1] Independent Researcher
## Abstract
Current multi-agent large language model (LLM) frameworks rely on explicit orchestration patterns borrowed from human organizational structures: planners delegate to executors, managers coordinate workers, and hierarchical control flow governs agent interactions. These approaches suffer from coordination overhead that scales poorly with agent count and task complexity. We propose a fundamentally different paradigm inspired by natural coordination mechanisms: agents operate locally on a shared artifact, guided only by pressure gradients derived from measurable quality signals, with temporal decay preventing premature convergence. We formalize this as optimization over a pressure landscape and prove convergence guarantees under mild conditions.
Empirically, on meeting room scheduling across 1350 total trials (270 per strategy), pressure-field coordination achieves $4Ă$ higher solve rates than conversation-based coordination and over $30Ă$ higher than hierarchical control (48.5% vs 11.1% vs 1.5%; all pairwise comparisons $p<0.001$ ). Ablation studies suggest temporal decay is beneficial. On easy problems, pressure-field achieves 86.7% solve rate compared to 33.3% for the next-best baseline. Foundation models enable this approach: their broad pretraining and zero-shot reasoning allow quality-improving patches from local pressure signals alone, without domain-specific coordination protocols. This suggests that constraint-driven emergence offers a simpler and more effective foundation for multi-agent AI.
keywords: multi-agent systems, emergent coordination, decentralized optimization, LLM agents, stigmergy
## 1 Introduction
Multi-agent systems built on Large Language Models address complex task automation [wu2023autogen, hong2023metagpt, li2023camel]. The dominant paradigm treats agents as organizational units: planners decompose tasks, managers delegate subtasks, and workers execute instructions under hierarchical supervision. This coordination overhead scales poorly with agent count and task complexity.
We demonstrate that implicit coordination through shared state outperforms explicit hierarchical controlâwithout coordinators, planners, or message passing. Across 1350 total trials on meeting room scheduling (270 per strategy), pressure-field coordination achieves over 30 $Ă$ higher solve rates than hierarchical control and 4 $Ă$ higher than conversation-based approaches [wu2023autogen] (all $p<0.001$ with large effect sizes). Sequential and random baselines achieve only 0.4%.
Our approach draws inspiration from natural coordination mechanismsâant colonies, immune systems, neural tissueâthat coordinate through environment modification rather than message passing. Agents observe local quality signals (pressure gradients), take locally-greedy actions, and coordination emerges from shared artifact state. The key insight is that local greedy decisions are effective for constraint satisfaction: when problems exhibit locality (fixing one region rarely breaks distant regions), decentralized greedy optimization outperforms centralized planning. Temporal decay prevents premature convergence by ensuring continued exploration.
Our contributions:
1. We formalize pressure-field coordination as a role-free, stigmergic alternative to organizational Multi-Agent System paradigms. Unlike Generalized Partial Global Planningâs hierarchical message-passing or SharedPlansâ intention alignment, pressure-field achieves $O(1)$ coordination overhead through shared artifact state. Foundation Models enable this approach: their broad pretraining allows quality-improving patches from local pressure signals without domain-specific coordination protocols.
1. We introduce temporal decay as a mechanism for preventing premature convergence. Ablation studies show a 10 percentage point improvement with decay enabled (96.7% vs 86.7%), directionally consistent with the theoretical prediction that decay helps escape local minima, though not statistically significant at $n=30$ .
1. We prove convergence guarantees for this coordination scheme under pressure alignment conditions.
1. We provide empirical evidence across 1350 total trials (270 per strategy) showing: (a) pressure-field dramatically outperforms all baselines by an order of magnitude or more, (b) all pairwise comparisons are highly significant ( $p<0.001$ ).
This work demonstrates that Foundation Model capabilities and Multi-Agent System coordination mechanisms are mutually enabling, not merely additive. Foundation Models solve a fundamental Multi-Agent System problem: traditional coordination requires explicit action space enumeration, but for open-ended artifact refinement the space of valid improvements is unbounded. Foundation Modelsâ broad pretraining provides implicit coverage of improvement strategies without domain-specific action representationsâa âuniversal actorâ capability. Conversely, Multi-Agent System coordination solves a fundamental Foundation Model problem: how to combine multiple model outputs coherently. Pressure gradients provide an objective criterion for output selection, replacing ad-hoc voting or ranking with principled quality-based filtering. This bidirectional synthesis explains why pressure-field coordination outperforms conversation-based alternatives that lack objective gradients for output combination.
## 2 Related Work
Our approach bridges four research traditions: multi-agent systems coordination theory provides the conceptual foundation; swarm intelligence provides the stigmergic mechanism; Large Language Model systems provide the application domain; and decentralized optimization provides theoretical guarantees. We survey each and position pressure-field coordination within this landscape.
### 2.1 MAS Coordination Theory
Pressure-field coordination occupies a unique position in the Multi-Agent System landscape: it eliminates roles (unlike organizational paradigms), messages (unlike Generalized Partial Global Planning), and intention reasoning (unlike SharedPlans) while providing formal convergence guarantees (unlike purely reactive systems). This section positions our contribution within four established coordination frameworks, showing how artifact refinement with measurable quality signals enables this architectural simplification. For this domain class, coordination complexity collapses from quadratic message-passing to constant-time state-sharing.
#### 2.1.1 Organizational Paradigms and Dependency Management
Pressure-field coordination achieves role-free coordination: any agent can address any high-pressure region without negotiating access rights or awaiting task assignment. This contrasts sharply with traditional organizational paradigms. Horling and Lesser [horling2004survey] surveyed nine such paradigmsâfrom rigid hierarchies to flexible marketsâfinding that all assign explicit roles constraining agent behavior. Dignum [dignum2009handbook] systematizes this tradition, defining organizational models through three dimensions: structure (roles and relationships), norms (behavioral constraints), and dynamics (how organizations adapt). These dimensions require explicit specification and maintenanceâdesigners must anticipate role interactions, encode coordination norms, and implement adaptation mechanisms.
Pressure-field coordination eliminates all three dimensions through gradient-based coordination. Roles dissolve: any agent may address any high-pressure region without negotiating access rights or awaiting task assignment. Norms become implicit: the pressure function encodes what âgoodâ behavior means, and agents that reduce pressure are by definition norm-compliant. Dynamics emerge naturally: temporal decay continuously destabilizes the pressure landscape, forcing ongoing adaptation without explicit organizational change protocols. Where Dignumâs organizational models require designers to specify âwho may do what with whom,â pressure-field coordination answers: âanyone may improve anywhere, and coordination emerges from shared perception of quality signals.â
Our approach instantiates Malone and Crowstonâs [malone1994coordination] coordination framework with a critical difference: the artifact itself is the shared resource, and pressure gradients serve as dependency signals. Malone and Crowston identify âshared resourceâ management as a fundamental coordination pattern requiring protocols for access control, conflict resolution, and priority assignment. Pressure-field coordination implements this pattern through a different mechanism: rather than assigning roles to manage resource access, agents share read access to the entire artifact and propose changes to high-pressure regions. Selection and validation phases resolve conflicts implicitlyâonly pressure-reducing patches are applied, and the highest-scoring patch wins when proposals conflict. Coordination emerges from pressure alignmentâagents reduce local pressure, which reduces global pressure through the artifactâs shared state.
#### 2.1.2 Distributed Problem Solving and Communication Overhead
Pressure-field coordination achieves $O(1)$ inter-agent communication overheadâagents exchange no messages. Coordination occurs entirely through shared artifact reads and writes, eliminating the message-passing bottleneck. This contrasts with the Generalized Partial Global Planning framework [decker1995gpgp], which reduces communication from $O(n^2)$ pairwise negotiation to $O(n\log n)$ hierarchical aggregation through summary information exchange. While Generalized Partial Global Planning represents significant progress, its explicit messagesâtask announcements, commitment exchanges, schedule updatesâstill introduce latency and failure points at scale.
The approaches target different domains. Pressure-field coordination specializes in artifact refinement tasks where quality decomposes into measurable regional signalsâa class including code quality improvement, document editing, and configuration management. Generalized Partial Global Planning generalizes to complex task networks with precedence constraints. For artifact refinement, however, pressure-fieldâs stigmergic coordination eliminates message-passing overhead entirely.
#### 2.1.3 Shared Intentions and Alignment Costs
Pressure-field coordination eliminates intention alignment through pressure alignment. Rather than reasoning about what other agents believe or intend, agents observe artifact state and pressure gradients. When agents greedily reduce local pressure under separable or bounded-coupling conditions, global pressure decreases. This is coordination without communication about intentionsâagents align through shared objective functions, not mutual beliefs.
This contrasts sharply with two foundational frameworks for joint activity. The SharedPlans framework [grosz1996sharedplans] formalizes collaboration through shared mental attitudes: mutual beliefs about goals, commitments, and action sequences. Cohen and Levesqueâs [cohen1991teamwork] Joint Intentions theory provides an even more stringent requirement: team members must hold mutual beliefs about the joint goal, individual commitments to the goal, and mutual beliefs about each memberâs commitment. Both frameworks capture human-like collaboration but require significant cognitive machineryâintention recognition, commitment protocols, belief revisionâall computationally expensive operations that scale poorly with agent count.
Pressure-field coordination eliminates the mutual belief formation that Joint Intentions requires. Where Cohen and Levesque demand that each agent believe that all teammates are committed to the joint goal (and believe that all teammates believe this, recursively), pressure-field agents need only observe local pressure gradients. The shared artifact is the mutual beliefâagents perceive the same pressure landscape without explicit belief exchange. This eliminates the infinite regress of âI believe that you believe that I believeâ that makes Joint Intentions computationally expensive at scale.
Our experiments validate this analysis: pressure-field coordination eliminates the overhead of explicit dialogue by coordinating through shared artifact state. The coordination overhead of belief negotiation in explicit dialogue systems can exceed its organizational benefit for constraint satisfaction tasks. The trade-off is transparency: SharedPlans and Joint Intentions support dialogue about why agents act and what teammates are committed to; pressure-field agents react to gradients without explaining reasoning or maintaining models of teammate intentions.
#### 2.1.4 Self-Organization and Emergent Coordination
Pressure-field coordination satisfies the self-organization criteria established by two complementary frameworks. De Wolf and Holvoet [dewolf2005engineering] characterize self-organizing systems through absence of external control, local interactions producing global patterns, and dynamic adaptation. They explicitly cite âgradient fieldsâ as a self-organization design patternâour approach instantiates this pattern with formal guarantees.
Serugendo et al. [serugendo2005selforganisation] provide a more fine-grained taxonomy, identifying four mechanisms through which self-organization emerges: (1) positive feedback amplifying beneficial behaviors, (2) negative feedback dampening harmful behaviors, (3) randomness enabling exploration, and (4) multiple interactions allowing local behaviors to propagate globally. Pressure-field coordination instantiates all four mechanisms:
- Positive feedback: Successful patches are stored as few-shot examples (âpositive pheromonesâ), increasing the probability of similar improvements in neighboring regions. This amplifies productive behaviors.
- Negative feedback: Temporal decay continuously erodes fitness, preventing any region from becoming permanently âsolved.â Inhibition further dampens over-activity in recently-patched regions.
- Randomness: Stochastic model sampling and band escalation (exploitation $â$ exploration) inject controlled randomness that prevents premature convergence to local optima.
- Multiple interactions: Each tick produces $K$ parallel patch proposals across agents, with selection applying the best improvements. Multiple interactions per timestep accelerate the propagation of successful strategies.
No external controller existsâagents observe and act autonomously based on local pressure signals. Coordination emerges from local decisions: agents reduce regional pressure through greedy actions, and global coordination arises from shared artifact state. Temporal decay provides dynamic adaptationâfitness erodes continuously, preventing premature convergence and enabling continued refinement.
The theoretical contribution formalizes this intuition through potential game theory. Theorem 5.1 establishes convergence guarantees for aligned pressure systems; the Basin Separation result (Theorem 5.3) explains why decay is necessary to escape suboptimal basins. This connects self-organization principles to formal coordination theory: Serugendoâs four mechanisms map to our formal modelâpositive feedback to pheromone reinforcement, negative feedback to decay and inhibition, randomness to stochastic sampling, and multiple interactions to parallel validation.
#### 2.1.5 Foundation Model Enablement
Foundation Models enable stigmergic coordination by providing capabilities that map directly to stigmergic requirements. This mapping explains why pressure-field coordination becomes practical with Foundation Models in ways it was not with prior agent architectures.
Broad pretraining enables domain-general patch proposal. Stigmergic coordination requires agents that can propose improvements based solely on local quality signals. Traditional agents require domain-specific action representationsâenumerated moves, parameterized operators, or learned policies. Foundation Modelsâ pretraining on diverse corpora (code, text, structured data) allows patch generation across artifact types without fine-tuning. An Large Language Model can propose meeting schedule adjustments, code refactorings, or configuration changes through the same interface: observe local context, receive pressure feedback, generate improvement. This âuniversal actorâ capability is what makes stigmergic coordination practical for open-ended artifact refinement.
Instruction-following replaces action space enumeration. Stigmergic agents need only local context to actâthey should not require global state, explicit goals, or complex action grammars. Foundation Modelsâ instruction-following capabilities allow operation from natural language pressure descriptions alone. Rather than encoding âreduce scheduling conflictsâ as a formal operator with preconditions and effects, we simply prompt: âThis time block has 3 double-bookings. Propose a schedule change to reduce conflicts.â The Foundation Model interprets this pressure signal and generates appropriate patches without requiring designers to enumerate valid actions.
Zero-shot reasoning interprets quality signals. Stigmergic coordination requires agents to recognize quality deficiencies locally. Foundation Models can identify constraint violations, inefficiencies, and improvement opportunities from examples alone, without explicit training on domain-specific quality metrics. When presented with a schedule showing âMeeting A and Meeting B both have Alice at 2pm,â the model recognizes the conflict and proposes resolutionsânot because it was trained on scheduling, but because conflict recognition transfers from pretraining.
In-context learning implements pheromone memory. Stigmergic systems reinforce successful strategies through positive feedback. Foundation Modelsâ in-context learning provides this mechanism naturally: successful patches become few-shot examples in subsequent prompts, increasing the probability of similar improvements. This âpositive pheromoneâ effect requires no external memory systemâthe prompt itself carries reinforcement signals.
Generative flexibility enables unbounded solution spaces. Traditional stigmergic systems (ant colony optimization, particle swarm) operate over discrete, enumerated solution spaces. Foundation Models generate from effectively continuous spaces, proposing patches that no designer anticipated. This generative flexibility is essential for open-ended artifact refinement where the space of valid improvements cannot be enumerated in advance.
These five capabilitiesâdomain-general patches, instruction-based operation, zero-shot quality recognition, in-context reinforcement, and generative flexibilityâcollectively enable stigmergic coordination for artifact refinement tasks that were previously intractable. The FM-MAS synthesis is not merely additive: Foundation Models solve the action enumeration problem that blocked stigmergic approaches, while stigmergic coordination solves the output combination problem that limits single-Foundation Model systems.
### 2.2 Multi-Agent LLM Systems
Recent work has explored multi-agent architectures for Large Language Model-based task solving. AutoGen [wu2023autogen] introduces a conversation-based framework where customizable agents interact through message passing, with support for human-in-the-loop workflows. MetaGPT [hong2023metagpt] encodes Standardized Operating Procedures into agent workflows, assigning specialized roles (architect, engineer, QA) in an assembly-line paradigm. CAMEL [li2023camel] proposes role-playing between AI assistant and AI user agents, using inception prompting to guide autonomous cooperation. CrewAI [crewai2024] similarly defines agents with roles, goals, and backstories that collaborate on complex tasks.
These frameworks share a common design pattern: explicit orchestration through message passing, role assignment, and hierarchical task decomposition. While effective for structured workflows, this approach faces scaling limitations. Central coordinators become bottlenecks, message-passing overhead grows with agent count, and failures in manager agents cascade to dependents. Our work takes a fundamentally different approach: coordination emerges from shared state rather than explicit communication.
Foundation Models enable pressure-field coordination through capabilities that prior agent architectures lacked. Their broad pretraining allows patches across diverse artifact typesâcode, text, configurationsâwithout domain-specific fine-tuning. Their instruction-following capabilities allow operation from pressure signals and quality feedback alone. Their zero-shot reasoning interprets constraint violations and proposes repairs without explicit protocol training. These properties make Foundation Models particularly suitable for stigmergic coordination: they require only local context and quality signals to generate productive actions, matching the locality constraints of pressure-field systems.
### 2.3 Swarm Intelligence and Stigmergy
The concept of stigmergyâindirect coordination through environment modificationâwas introduced by GrassĂ© [grasse1959stigmergie] to explain termite nest-building behavior. Termites deposit pheromone-infused material that attracts further deposits, leading to emergent construction without central planning. This directly instantiates Malone and Crowstonâs [malone1994coordination] shared resource coordination: pheromone trails encode dependency information about solution quality. Complex structures arise from simple local rules without any agent having global knowledge.
Dorigo and colleagues [dorigo1996ant, dorigo1997acs] formalized this insight into Ant Colony Optimization, where artificial pheromone trails guide search through solution spaces. Key mechanisms include positive feedback (reinforcing good paths), negative feedback (pheromone evaporation), and purely local decision-making. Ant Colony Optimization has achieved strong results on combinatorial optimization problems including Traveling Salesman Problem, vehicle routing, and scheduling.
Our pressure-field coordination directly inherits from stigmergic principles. The artifact serves as the shared environment; regional pressures are analogous to pheromone concentrations; decay corresponds to evaporation. However, we generalize beyond path-finding to arbitrary artifact refinement and provide formal convergence guarantees through the potential game framework.
### 2.4 Decentralized Optimization
Potential games, introduced by Monderer and Shapley [monderer1996potential], are games where individual incentives align with a global potential function. A key property is that any sequence of unilateral improvements converges to a Nash equilibriumâgreedy local play achieves global coordination. This provides the theoretical foundation for our convergence guarantees: under pressure alignment, the artifact pressure serves as a potential function.
Distributed gradient descent methods [nedic2009distributed, yuan2016convergence] address optimization when data or computation is distributed across nodes. The standard approach combines local gradient steps with consensus averaging. While these methods achieve convergence rates matching centralized alternatives, they typically require communication protocols and synchronization. Our approach avoids explicit communication entirely: agents coordinate only through the shared artifact, achieving $O(1)$ coordination overhead.
The connection between multi-agent learning and game theory has been extensively studied [shoham2008multiagent]. Our contribution is applying these insights to Large Language Model-based artifact refinement, where the âgameâ is defined by pressure functions over quality signals rather than explicit reward structures.
## 3 Problem Formulation
We formalize artifact refinement as a dynamical system over a pressure landscape rather than an optimization problem with a target state. The system evolves through local actions and continuous decay, settling into stable basins that represent acceptable artifact states.
### 3.1 State Space
An artifact consists of $n$ regions with content $c_iâC$ for $iâ\{1,âŠ,n\}$ , where $C$ is an arbitrary content space (strings, Abstract Syntax Tree nodes, etc.). Each region also carries auxiliary state $h_iâH$ representing confidence, fitness, and history. Regions are passive subdivisions of the artifact; agents are active proposers that observe regions and generate patches.
The full system state is:
$$
s=((c_1,h_1),âŠ,(c_n,h_n))â(CĂH)^n
$$
### 3.2 Pressure Landscape
A signal function $Ï:Cââ^d$ maps content to measurable features. Signals are local: $Ï(c_i)$ depends only on region $i$ .
A pressure function $Ï:â^dââ_â„ 0$ maps signals to scalar âbadness.â We consider $k$ pressure axes with weights $wââ^k_>0$ . The region pressure is:
$$
P_i(s)=â_j=1^kw_jÏ_j(Ï(c_i))
$$
The artifact pressure is:
$$
P(s)=â_i=1^nP_i(s)
$$
This defines a landscape over artifact states. Low-pressure regions are âvalleysâ where the artifact satisfies quality constraints.
### 3.3 System Dynamics
The system evolves in discrete time steps (ticks). Each tick consists of four phases:
Phase 1: Decay. Auxiliary state erodes toward a baseline. For fitness $f_i$ and confidence $Îł_i$ components of $h_i$ :
$$
f_i^t+1=f_i^t· e^-λ_f, γ_i^t+1=γ_i^t· e^-λ_γ
$$
where $λ_f,λ_γ>0$ are decay rates. Decay ensures that stability requires continuous reinforcement.
Phase 2: Proposal. For each region $i$ where pressure exceeds activation threshold ( $P_i>Ï_act$ ) and the region is not inhibited, each actor $a_k:CĂHĂâ^dâC$ proposes a content transformation in parallel. Each actor observes only local state $(c_i,h_i,Ï(c_i))$ âactors do not communicate or coordinate their proposals.
Phase 3: Validation. When multiple patches are proposed, each is validated on an independent fork of the artifact. Forks are created by cloning artifact state; validation proceeds in parallel across forks. This addresses a fundamental resource constraint: a single artifact cannot be used to test multiple patches simultaneously without cloning.
Phase 4: Reinforcement. Regions where actions were applied receive fitness and confidence boosts, and enter an inhibition period preventing immediate re-modification. Inhibition allows changes to propagate through the artifact and forces agents to address other high-pressure regions, preventing oscillation around local fixes.
$$
f_i^t+1=\min(f_i^t+Î_f,1), Îł_i^t+1=\min(Îł_i^t+Î_Îł,1)
$$
### 3.4 Stable Basins
**Definition 3.1 (Stability)**
*A state $s^*$ is stable if, under the system dynamics with no external perturbation:
1. All region pressures are below activation threshold: $P_i(s^*)<Ï_act$ for all $i$
1. Decay is balanced by residual fitness: the system remains in a neighborhood of $s^*$*
The central questions are:
1. Existence: Under what conditions do stable basins exist?
1. Quality: What is the pressure $P(s^*)$ of states in stable basins?
1. Convergence: From initial state $s_0$ , does the system reach a stable basin? How quickly?
1. Decentralization: Can stability be achieved with purely local decisions?
### 3.5 The Locality Constraint
The constraint distinguishing our setting from centralized optimization: agents observe only local state. An actor at region $i$ sees $(c_i,h_i,Ï(c_i))$ but not:
- Other regionsâ content $c_j$ for $jâ i$
- Global pressure $P(s)$
- Other agentsâ actions
This rules out coordinated planning. Stability must emerge from local incentives aligned with global pressure reduction.
## 4 Method
We now present a coordination mechanism that achieves stability through purely local decisions. Under appropriate conditions, the artifact pressure $P(s)$ acts as a potential function: local improvements by individual agents decrease global pressure, guaranteeing convergence without coordination.
### 4.1 Pressure Alignment
The locality constraint prohibits agents from observing global state. For decentralized coordination to succeed, we need local incentives to align with global pressure reduction.
**Definition 4.1 (Pressure Alignment)**
*A pressure system is aligned if for any region $i$ , state $s$ , and action $a_i$ that reduces local pressure:
$$
P_i(s^\prime)<P_i(s) \Longrightarrow P(s^\prime)<P(s)
$$
where $s^\prime=s[c_i⊠a_i(c_i)]$ is the state after applying $a_i$ .*
Alignment holds automatically when pressure functions are separable: each $P_i$ depends only on $c_i$ , so $P(s)=â_iP_i(s)$ and local improvement directly implies global improvement.
More generally, alignment holds when cross-region interactions are bounded:
**Definition 4.2 (Bounded Coupling)**
*A pressure system has $Δ$ -bounded coupling if for any action $a_i$ on region $i$ :
$$
|P_j(s^\prime)-P_j(s)|â€Î” â jâ i
$$
That is, modifying region $i$ changes other regionsâ pressures by at most $Δ$ .*
Under $Δ$ -bounded coupling with $n$ regions, if a local action reduces $P_i$ by $Ύ>(n-1)Δ$ , then global pressure decreases by at least $Ύ-(n-1)Δ>0$ .
### 4.2 Connection to Potential Games
The aligned pressure system forms a potential game where:
- Players are regions (or agents acting on regions)
- Strategies are content choices $c_iâC$
- The potential function is $Ί(s)=P(s)$
In potential games, any sequence of improving moves converges to a Nash equilibrium. In our setting, Nash equilibria correspond to stable basins: states where no local action can reduce pressure below the activation threshold.
This connection provides our convergence guarantee without requiring explicit coordination.
Note that this convergence result assumes finite action spaces. In practice, patches are drawn from a finite set of Large Language Model-generated proposals per region, satisfying this requirement. More fundamentally, the validation phase (Phase 2b) implicitly discretizes the action space: only patches that reduce pressure are accepted, so the effective action set at any state is the finite set of pressure-reducing proposals generated that tick. For infinite content spaces, convergence to approximate equilibria can be established under Lipschitz continuity conditions on pressure functions.
### 4.3 The Coordination Algorithm
The tick loop implements greedy local improvement with decay-driven exploration:
Algorithm 1: Pressure-Field Tick
Input: State $s^t$ , signal functions $\{Ï_j\}$ , pressure functions $\{Ï_j\}$ , actors $\{a_k\}$ , parameters $(Ï_act,λ_f,λ_Îł,Î_f,Î_Îł,Îș)$
Phase 1: Decay For each region $i$ : $f_iâ f_i· e^-λ_f, Îł_iâÎł_i· e^-λ_Îł$
Phase 2: Activation and Proposal $Pââ
$
For each region $i$ where $P_i(s)â„Ï_act$ and not inhibited:
$\boldsymbol{Ï}_iâÏ(c_i)$
For each actor $a_k$ :
$ÎŽâ a_k(c_i,h_i,\boldsymbol{Ï}_i)$
$PâPâȘ\{(i,ÎŽ,\hat{Î}(ÎŽ))\}$
Phase 3: Parallel Validation and Selection For each candidate patch $(i,ÎŽ,\hat{Î})âP$ :
Fork artifact: $(f_id,A_f)â A.fork()$
Apply $ÎŽ$ to fork $A_f$
Validate fork (run tests, check compilation)
Collect validation results $\{(i,ÎŽ,Î_actual,valid)\}$
Sort validated patches by $Î_actual$
Greedily select top- $Îș$ non-conflicting patches
Phase 4: Application and Reinforcement For each selected patch $(i,Ύ,·)$ :
$c_iâÎŽ(c_i)$
$f_iâ\min(f_i+Î_f,1)$ , $Îł_iâ\min(Îł_i+Î_Îł,1)$
Mark region $i$ inhibited for $Ï_inh$ ticks
Return updated state $s^t+1$
The algorithm has three key properties:
Locality. Each actor observes only $(c_i,h_i,Ï(c_i))$ . No global state is accessed.
Bounded parallelism. At most $Îș$ patches per tick prevents thrashing. Inhibition prevents repeated modification of the same region.
Decay-driven exploration. Even stable regions eventually decay below confidence thresholds, attracting re-evaluation. This prevents premature convergence to local minima.
### 4.4 Stability and Termination
The system reaches a stable basin when:
1. All region pressures satisfy $P_i(s)<Ï_act$
1. Decay is balanced: fitness remains above the threshold needed for stability
Termination is economic, not logical. The system stops acting when the cost of action (measured in pressure reduction per patch) falls below the benefit. This matches natural systems: activity ceases when gradients flatten, not when an external goal is declared achieved.
In practice, we also impose budget constraints (maximum ticks or patches) to bound computation.
## 5 Theoretical Analysis
We establish three main results: (1) convergence to stable basins under alignment, (2) bounds on stable basin quality, and (3) scaling properties relative to centralized alternatives.
### 5.1 Convergence Under Alignment
**Theorem 5.1 (Convergence)**
*Let the pressure system be aligned with $Δ$ -bounded coupling. Let $Ύ_\min>0$ be the minimum local pressure reduction $P_i(s)-P_i(s^\prime)$ from any applied patch, and assume $Ύ_\min>(n-1)Δ$ where $n$ is the number of regions. Then from any initial state $s_0$ with pressure $P_0=P(s_0)$ , the system reaches a stable basin within:
$$
Tâ€\frac{P_0}{ÎŽ_\min-(n-1)Δ}
$$
ticks, provided the fitness boost $Î_f$ from successful patches exceeds decay during inhibition: $Î_f>1-e^-λ_f·Ï_inh$ .*
Proof sketch. Under alignment with $Δ$ -bounded coupling, each applied patch reduces global pressure by at least $ÎŽ_\min-(n-1)Δ>0$ . Since $P(s)â„ 0$ and decreases by a fixed minimum per tick (when patches are applied), the system must reach a state where no region exceeds $Ï_act$ within the stated bound. The decay constraint ensures that stability is maintained once reached: fitness reinforcement from the final patches persists longer than the decay erodes it. $\square$
The bound is loose but establishes that convergence time scales with initial pressure, not with state space size or number of possible actions.
### 5.2 Basin Quality
**Theorem 5.2 (Basin Quality)**
*In any stable basin $s^*$ , the artifact pressure satisfies:
$$
P(s^*)<n·Ï_act
$$
where $n$ is the number of regions and $Ï_act$ is the activation threshold.*
Proof. By definition of stability, $P_i(s^*)<Ï_act$ for all $i$ . Summing over regions: $P(s^*)=â_iP_i(s^*)<n·Ï_act$ . $\square$
This bound is tight: adversarial initial conditions can place the system in a basin where each region has pressure just below threshold. However, in practice, actors typically reduce pressure well below $Ï_act$ , yielding much lower basin pressures.
**Theorem 5.3 (Basin Separation)**
*Under separable pressure (zero coupling), distinct stable basins are separated by pressure barriers of height at least $Ï_act$ .*
Proof sketch. Moving from one basin to another requires some region to exceed $Ï_act$ (otherwise no action is triggered). The minimum such exceedance defines the barrier height. $\square$
This explains why decay is necessary: without decay, the system can become trapped in suboptimal basins. Decay gradually erodes fitness, eventually allowing re-evaluation and potential escape to lower-pressure basins.
### 5.3 Scaling Properties
**Theorem 5.4 (Linear Scaling)**
*Let $m$ be the number of regions and $n$ be the number of parallel agents. The per-tick complexity is:
- Signal computation: $O(m· d)$ where $d$ is signal dimension
- Pressure computation: $O(m· k)$ where $k$ is the number of pressure axes
- Patch proposal: $O(m· a)$ where $a$ is the number of actors
- Selection: $O(m· a·\log(m· a))$ for sorting candidates
- Coordination overhead: $O(1)$ âno inter-agent communication (fork pool is $O(K)$ where $K$ is fixed)
Total: $O(m·(d+k+a·\log(ma)))$ , independent of agent count $n$ .*
Adding agents increases throughput (more patches proposed per tick) without increasing coordination cost. This contrasts with hierarchical schemes where coordination overhead grows with agent count.
**Theorem 5.5 (Parallel Convergence)**
*Under the same alignment conditions as Theorem 5.1, with $K$ patches validated in parallel per tick where patches affect disjoint regions, the system reaches a stable basin within:
$$
Tâ€\frac{P_0}{K·(ÎŽ_\min-(n-1)Δ)}
$$
This improves convergence time by factor $K$ while maintaining guarantees.*
Proof sketch. When $K$ non-conflicting patches are applied per tick, each reduces global pressure by at least $Ύ_\min-(n-1)Δ$ . The combined reduction is $K·(Ύ_\min-(n-1)Δ)$ per tick. The bound follows directly. Note that if patches conflict (target the same region), only one is selected per region, and effective speedup is reduced. $\square$
### 5.4 Comparison to Alternatives
We compare against three coordination paradigms:
Centralized planning. A global planner evaluates all $(m· a)$ possible actions, selects optimal subset. Per-step complexity: $O(m· a)$ evaluations, but requires global state access. Sequential bottleneck prevents parallelization.
Hierarchical delegation. Manager agents decompose tasks, delegate to workers. Communication complexity: $O(n\log n)$ for tree-structured delegation with $n$ agents. Latency scales with tree depth. Failure of manager blocks all descendants.
Message-passing coordination. Agents negotiate actions through pairwise communication. Convergence requires $O(n^2)$ messages in worst case for $n$ agents. Consensus protocols add latency.
$$
O(m· a) O(n\log n) O(n^2) O(1) \min(n,m,K) \tag{1}
$$
Table 1: Coordination overhead comparison. $K$ denotes the fork pool size for parallel validation.
Pressure-field coordination achieves $O(1)$ coordination overhead because agents share state only through the artifact itselfâa form of stigmergy. Agents can fail, join, or leave without protocol overhead.
## 6 Experiments
We evaluate pressure-field coordination on meeting room scheduling: assigning $N$ meetings to $R$ rooms over $D$ days to minimize gaps (unscheduled time), overlaps (attendee double-bookings), and maximize utilization balance. This domain provides continuous pressure gradients (rather than discrete violations), measurable success criteria, and scalable difficulty through problem size.
Key findings: Pressure-field coordination outperforms all baselines (§ 6.2). Temporal decay shows a beneficial trend, though statistical significance requires larger samples (§ 6.3). The approach maintains consistent performance from 1 to 4 agents (§ 6.4). Despite using more tokens per trial, pressure-field achieves 12% better token efficiency per successful solve (§ 6.9).
### 6.1 Setup
#### 6.1.1 Task: Meeting Room Scheduling
We generate scheduling problems with varying difficulty:
| Easy | 3 | 20 | 70% |
| --- | --- | --- | --- |
| Medium | 5 | 40 | 50% |
| Hard | 5 | 60 | 30% |
Table 2: Problem configurations. Pre-scheduled percentage indicates meetings already placed; remaining meetings must be scheduled by agents.
Each schedule spans 5 days with 30-minute time slots (8amâ4pm). Regions are 2-hour time blocks (4 blocks per day $Ă$ 5 days = 20 regions per schedule). A problem is âsolvedâ when all meetings are scheduled with zero attendee overlaps within 50 ticks.
Pressure function: $P=gaps· 1.0+overlaps· 2.0+util\_var· 0.5+unsched· 1.5$
where gaps measures empty slots as a fraction, overlaps counts attendee double-bookings, util_var measures room utilization variance, and unsched is the fraction of unscheduled meetings.
Alignment verification: The per-region pressure computation uses only gaps, overlaps, and util_var âall strictly local to each time block. The unsched component is added to total pressure only, not per-region. This makes the per-region pressure separable: modifying region $i$ has zero effect on region $j$ âs pressure for $jâ i$ , satisfying the alignment condition (Definition 2) with $Δ=0$ . While attendee constraints could theoretically create cross-region coupling (the same person attending meetings in different time blocks), our overlap sensor counts overlaps only within each time block, eliminating this coupling source. Empirical analysis (Appendix B) confirms that all observed pressure improvements are positive, consistent with separable pressure.
#### 6.1.2 Baselines
We compare five coordination strategies, all using identical Large Language Models (qwen2.5:0.5b/1.5b/3b via Ollama) to isolate coordination effects:
Pressure-field (ours): Full system with decay (fitness half-life 5s), inhibition (2s cooldown), greedy region selection (highest-pressure region per tick), and parallel validation. Includes band escalation (Exploitation $â$ Balanced $â$ Exploration) and model escalation (0.5b $â$ 1.5b $â$ 3b).
Conversation: AutoGen-style multi-agent dialogue where agents exchange messages to coordinate scheduling decisions. Agents discuss conflicts and propose solutions through explicit communication.
Hierarchical: Single agent selects the highest-pressure time block each tick, proposes a schedule change, and validates before applying (only accepts pressure-reducing patches). Uses identical prompts to pressure-field. The differences are: (1) greedy region selection always targets the hardest region, and (2) sequential execution processes one region per tick. This represents centralized, quality-gated control.
Sequential: Single agent iterates through time blocks in fixed order, proposing schedule changes one region at a time. No parallelism, pressure guidance, or patch validationâapplies any syntactically valid patch regardless of quality impact.
Random: Selects random time blocks and proposes schedule changes. No patch validationâapplies any syntactically valid patch regardless of quality impact.
Note on parallelism: Pressure-field validates multiple patches in parallel ( $K$ regions per tick), while hierarchical validates one patch sequentially. This asymmetry is inherent to the coordination paradigm, not an implementation choice: hierarchical control requires the manager to select a region, delegate to a worker, and validate the result before proceedingâdelegating to multiple workers simultaneously would require additional coordination protocols (work distribution, conflict resolution, result aggregation) that would transform it into a different architecture entirely. The sequential bottleneck is the cost of centralized control. When hierarchicalâs single patch is rejected, the tick produces no progress; when one of pressure-fieldâs parallel patches is rejected, others may still succeed.
Model choice rationale: We deliberately use small, minimally-capable models (0.5bâ3b parameters) rather than frontier models. This design choice strengthens our thesis: if the coordination mechanism can extract effective performance from weak models, the mechanism itself is valuableâindependent of model capability. Using identical model chains across all strategies isolates coordination effects from model effects. We hypothesize that frontier models (e.g., GPT-4, Claude) would raise absolute solve rates across all strategies while preserving relative rankings: the coordination advantage is orthogonal to model capability, so pressure-fieldâs $4Ă$ improvement over conversation baselines should persist even as the baseline rises.
#### 6.1.3 Metrics
- Solve rate: Percentage of schedules reaching all meetings placed with zero overlaps within 50 ticks.
- Ticks to solve: Convergence speed for solved cases
- Final pressure: Remaining gaps, overlaps, and unscheduled meetings for unsolved cases
- Token efficiency: Total prompt and completion tokens consumed per trial and per successful solve
#### 6.1.4 Implementation
Hardware: NVIDIA RTX 4070 8GB Graphics Processing Unit, AMD Ryzen 9 7940HS, 64GB RAM. Software: Rust implementation with Ollama. Trials: 30 per configuration. Full protocol in Appendix A.
Band escalation: When pressure velocity (rate of improvement) drops to zero for 7 consecutive ticks, sampling parameters escalate: Exploitation (T=0.2, p=0.85) $â$ Balanced (T=0.4, p=0.9) $â$ Exploration (T=0.7, p=0.95).
Model escalation: After exhausting all bands with zero progress (21 ticks total), the system escalates through the model chain: 0.5b $â$ 1.5b $â$ 3b, resetting to Exploitation band. Section 6.5 analyzes this mechanism.
### 6.2 Main Results
Across 1350 total trials spanning three difficulty levels (easy, medium, hard) and agent counts (1, 2, 4), we find that pressure-field coordination outperforms all baselines:
| Pressure-field Conversation Hierarchical | 131/270 30/270 4/270 | 48.5% 11.1% 1.5% | 42.6%â54.5% 7.9%â15.4% 0.6%â3.7% |
| --- | --- | --- | --- |
| Sequential | 1/270 | 0.4% | 0.1%â2.1% |
| Random | 1/270 | 0.4% | 0.1%â2.1% |
Table 3: Aggregate solve rates across all experiments (1350 total trials, 270 per strategy). Chi-square test across all five strategies: $Ï^2>200$ , $p<0.001$ .
The results show clear stratification:
Pressure-field dominates: Pressure-field achieves 48.5% solve rate, roughly 4 $Ă$ higher than the next-best baseline (conversation at 11.1%). The effect size is large: Cohenâs $h=1.16$ versus conversation on easy problems, and $h>1.97$ versus all other baselines.
Conversation provides intermediate performance: The AutoGen-style conversation baseline achieves 11.1% overall, significantly better than hierarchical ( $p<0.001$ ) but far below pressure-field. Notably, conversation solves only easy problems (33.3% on easy, 0% on medium and hard).
Hierarchical and sequential fail: Despite explicit coordination, hierarchical control achieves only 1.5% solve rateâcomparable to random (0.4%). Both strategies fail entirely on medium and hard problems.
This result contradicts the common assumption that explicit hierarchical coordination should outperform implicit coordination. The overhead of centralized control and message passing appears to harm rather than help performance on constraint satisfaction tasks.
<details>
<summary>x1.png Details</summary>

### Visual Description
## Dot Plot with Error Bars: Solve Rate by Method and Difficulty
### Overview
The image displays a three-panel dot plot comparing the "Solve Rate" of five different methods (P-Field, Conv, Hier, Seq, Rand) across three difficulty levels: Easy, Medium, and Hard. Each panel represents a difficulty level, and within each panel, the solve rate for each method is shown as a colored dot with vertical error bars indicating variability or confidence intervals.
### Components/Axes
* **Chart Type:** Three-panel dot plot with error bars.
* **Panel Titles (Top):** "Easy", "Medium", "Hard" (centered above each respective panel).
* **Y-Axis (Left):** Labeled "Solve Rate". The scale runs from 0% to 100%, with major tick marks at 0%, 25%, 50%, 75%, and 100%.
* **X-Axis (Bottom of each panel):** Lists the five methods. The labels are rotated approximately 45 degrees for readability. The order is consistent across panels: `P-Field`, `Conv`, `Hier`, `Seq`, `Rand`.
* **Data Series (Color-Coded):**
* **P-Field:** Blue dot and error bars.
* **Conv:** Purple dot and error bars.
* **Hier:** Green dot and error bars.
* **Seq:** Red dot and error bars.
* **Rand:** Gray dot and error bars.
* **Legend:** Not explicitly shown as a separate box. The method labels on the x-axis serve as the legend, with their associated colors consistently applied to the data points directly above them.
### Detailed Analysis
**Panel 1: Easy**
* **P-Field (Blue):** Highest solve rate. Dot is positioned at approximately 87%. Error bars extend from ~78% to ~95%.
* **Conv (Purple):** Second highest. Dot at ~33%. Error bars from ~25% to ~42%.
* **Hier (Green):** Very low. Dot at ~5%. Error bars from ~0% to ~12%.
* **Seq (Red):** Near zero. Dot at ~1%. Error bars from ~0% to ~5%.
* **Rand (Gray):** Near zero. Dot at ~0%. Error bars from ~0% to ~7%.
**Panel 2: Medium**
* **P-Field (Blue):** Still the highest, but reduced. Dot at ~43%. Error bars from ~33% to ~53%.
* **Conv (Purple):** Near zero. Dot at ~0%. Error bars from ~0% to ~3%.
* **Hier (Green):** Near zero. Dot at ~0%. Error bars from ~0% to ~2%.
* **Seq (Red):** Near zero. Dot at ~0%. Error bars from ~0% to ~2%.
* **Rand (Gray):** Near zero. Dot at ~0%. Error bars from ~0% to ~5%.
**Panel 3: Hard**
* **P-Field (Blue):** Lowest among the P-Field results, but still the highest in this panel. Dot at ~15%. Error bars from ~8% to ~23%.
* **Conv (Purple):** Near zero. Dot at ~0%. Error bars from ~0% to ~2%.
* **Hier (Green):** Near zero. Dot at ~0%. Error bars from ~0% to ~2%.
* **Seq (Red):** Near zero. Dot at ~0%. Error bars from ~0% to ~2%.
* **Rand (Gray):** Near zero. Dot at ~0%. Error bars from ~0% to ~5%.
### Key Observations
1. **Dominant Performance:** The `P-Field` method (blue) significantly outperforms all other methods across all difficulty levels.
2. **Clear Difficulty Trend:** The solve rate for `P-Field` decreases sharply as task difficulty increases (Easy: ~87% â Medium: ~43% â Hard: ~15%).
3. **Secondary Method:** `Conv` (purple) shows moderate performance only on the "Easy" task (~33%). Its performance collapses to near zero for "Medium" and "Hard" tasks.
4. **Ineffective Methods:** `Hier`, `Seq`, and `Rand` show solve rates at or near 0% across all three difficulty levels, with error bars indicating minimal to no successful solves.
5. **Variability:** The error bars for `P-Field` are substantial, especially in the "Easy" and "Medium" panels, indicating considerable variability in its performance. The error bars for other methods are small, consistent with their near-zero success rates.
### Interpretation
This chart demonstrates a clear hierarchy of method effectiveness for the given task. `P-Field` is the only method that achieves meaningful solve rates, though its efficacy is highly sensitive to problem difficulty. The `Conv` method has limited utility, only showing promise on the easiest problems. The remaining methods (`Hier`, `Seq`, `Rand`) appear to be ineffective baselines or failed approaches for this specific challenge.
The data suggests that the problem space has a structure that `P-Field` is uniquely able to exploit, but this advantage erodes as complexity increases. The near-zero performance of `Rand` (presumably a random baseline) confirms that solving the task requires more than chance. The stark drop-off for all methods from "Easy" to "Medium" indicates a significant step-change in problem complexity between these levels. The consistent, low variability (small error bars) for the non-`P-Field` methods at the bottom of the chart reinforces the conclusion that they are not just occasionally failing, but are fundamentally unsuited to the task.
</details>
Figure 1: Strategy comparison by difficulty level. Error bars show 95% Wilson Confidence Intervals. Pressure-field outperforms all baselines at every difficulty level. On medium and hard problems, only pressure-field achieves non-zero solve rates.
### 6.3 Ablations
#### 6.3.1 Effect of Temporal Decay
Decay appears beneficial, though the effect does not reach statistical significance with our sample size:
| Full (with decay) Without decay | 29/30 26/30 | 96.7% 86.7% | 83.3%â99.4% 70.3%â94.7% |
| --- | --- | --- | --- |
Table 4: Decay ablation on easy scheduling problems (30 trials each). Fisherâs exact test: $p=0.35$ , indicating the 10 percentage point difference is not statistically significant at $α=0.05$ .
The observed effectâa 10 percentage point reduction when decay is disabledâis directionally consistent with our theoretical predictions. The non-significant $p$ -value ( $p=0.35$ ) reflects both limited sample size and a ceiling effect: the high baseline solve rate on easy problems (96.7% with decay) leaves limited statistical room to detect improvement. The overlapping confidence intervals (83.3%â99.4% vs 70.3%â94.7%) reflect this uncertainty. Without decay, fitness saturates after initial patchesâregions that received early patches retain high fitness indefinitely, making them appear âstableâ even when they still contain unscheduled meetings. Since greedy selection prioritizes high-pressure regions, these prematurely-stabilized regions are never reconsidered. This mechanism is consistent with the Basin Separation result (Theorem 5.3): without decay, agents may remain trapped in the first stable basin they reach. Larger-scale ablation studies would be needed to establish the statistical significance of decayâs contribution.
#### 6.3.2 Effect of Inhibition and Examples
The ablation study tested combinations of decay, inhibition, and few-shot examples on easy scheduling problems:
| Full | â | â | â | 96.7% |
| --- | --- | --- | --- | --- |
| No Decay | $Ă$ | â | â | 86.7% |
| No Inhibition | â | $Ă$ | â | 96.7% |
| No Examples | â | â | $Ă$ | 90.0% |
| Baseline | $Ă$ | $Ă$ | $Ă$ | 90.0% |
Table 5: Ablation results (30 trials each configuration on easy difficulty).
Feature contributions:
- Decay: +10.0% (full 96.7% vs no_decay 86.7%)
- Inhibition: +0.0% (no detectable effect)
- Examples: +6.7% (full 96.7% vs no_examples 90.0%)
Decay shows the largest effect: configurations with decay achieve higher solve rates, though the differences do not reach statistical significance at $n=30$ . Inhibition shows no detectable effect in this domain, possibly because the 50-tick budget provides sufficient exploration without explicit cooldowns.
<details>
<summary>x2.png Details</summary>

### Visual Description
## Horizontal Bar Chart with Feature Matrix: Solve Rate by Condition
### Overview
The image displays a horizontal bar chart titled "Solve Rate" on the x-axis, comparing the performance of eight different experimental conditions. To the left of the chart is a small 8x3 grid (matrix) that acts as a visual legend, indicating which features (Decay, Inhibition, Examples) are active or inactive for each condition. The chart shows that conditions with more active features generally achieve higher solve rates.
### Components/Axes
* **Chart Type:** Horizontal Bar Chart.
* **X-Axis:** Labeled "Solve Rate". The scale runs from 0% to 100%, with major tick marks at 0%, 25%, 50%, 75%, and 100%.
* **Y-Axis (Categories):** Lists eight experimental conditions. From top to bottom:
1. No Inhibition
2. Full
3. No Decay + No Inhib
4. No Examples
5. Baseline
6. No Decay
7. No Decay + No Ex
8. No Inhib + No Ex
* **Legend/Feature Matrix:** Positioned to the left of the y-axis labels. It is a grid with 8 rows (matching the conditions) and 3 columns.
* **Column Headers (Bottom):** "Decay", "Inhib", "Ex" (likely short for Examples).
* **Cell Content:** A blue square indicates the feature is **active** for that condition. A light gray square indicates the feature is **inactive**.
* **Data Labels:** The precise solve rate percentage is printed to the right of each bar.
### Detailed Analysis
The following table reconstructs the data from the chart and the feature matrix. The matrix is read row-by-row, corresponding to the condition listed on the same row.
| Condition (Y-Axis Label) | Decay (Feature) | Inhib (Feature) | Ex (Feature) | Solve Rate (Bar Value) |
| :--- | :---: | :---: | :---: | :---: |
| No Inhibition | Active (Blue) | Inactive (Gray) | Active (Blue) | 96.7% |
| Full | Active (Blue) | Active (Blue) | Active (Blue) | 96.7% |
| No Decay + No Inhib | Inactive (Gray) | Inactive (Gray) | Active (Blue) | 93.3% |
| No Examples | Active (Blue) | Active (Blue) | Inactive (Gray) | 90.0% |
| Baseline | Inactive (Gray) | Inactive (Gray) | Inactive (Gray) | 90.0% |
| No Decay | Inactive (Gray) | Active (Blue) | Active (Blue) | 86.7% |
| No Decay + No Ex | Inactive (Gray) | Active (Blue) | Inactive (Gray) | 83.3% |
| No Inhib + No Ex | Active (Blue) | Inactive (Gray) | Inactive (Gray) | 80.0% |
**Trend Verification:** The visual trend of the bars slopes downward from top to bottom. The top two bars ("No Inhibition" and "Full") are the longest and equal. The bars progressively shorten, with the bottom bar ("No Inhib + No Ex") being the shortest. This visual trend is confirmed by the numerical data, which decreases from 96.7% to 80.0%.
### Key Observations
1. **Top Performance:** The highest solve rate (96.7%) is achieved by two conditions: "No Inhibition" (Decay + Examples active) and "Full" (all features active). This suggests that for peak performance, the "Inhibition" feature may be redundant if "Decay" and "Examples" are present.
2. **Baseline Performance:** The "Baseline" condition, where all three features are inactive, achieves a 90.0% solve rate. This is a surprisingly high baseline.
3. **Feature Impact Hierarchy:** Removing features generally reduces performance, but not uniformly.
* Removing "Examples" ("No Examples") drops the rate to 90.0%, matching the Baseline.
* Removing "Decay" ("No Decay") drops it further to 86.7%.
* The lowest performance (80.0%) occurs when both "Inhibition" and "Examples" are removed ("No Inhib + No Ex"), even while "Decay" is active.
4. **Non-Additive Effects:** The impact of features is not simply additive. For example, "No Decay + No Inhib" (only Examples active) scores 93.3%, which is higher than "No Examples" (90.0%) or "No Decay" (86.7%) alone. This indicates complex interactions between the features.
### Interpretation
This chart likely presents ablation study results from a machine learning or cognitive architecture experiment. The "Solve Rate" is the primary performance metric. The three featuresâDecay, Inhibition, and Examplesâare components of a system being tested.
The data suggests that the **Examples** feature is critical for maintaining high performance above the baseline, especially when other features are missing. The **Decay** feature also contributes positively, but its absence is less detrimental than the absence of Examples in most combinations. The **Inhibition** feature appears to have a nuanced role; its presence doesn't hurt performance ("Full" vs. "No Inhibition"), but its absence in combination with the absence of Examples leads to the worst outcome.
The key takeaway is that the system is robust, maintaining a 90% solve rate even with no features active (Baseline). However, to push performance to the mid-90s, the combination of Decay and Examples is essential, with Inhibition providing no additional benefit in that specific context. The lowest performance occurs when the system is stripped of both its memory/reference mechanism (Examples) and its regulatory mechanism (Inhibition), even while a decay process is active. This implies that Examples and Inhibition may serve complementary or stabilizing roles that become crucial when the system is otherwise simplified.
</details>
Figure 2: Ablation study results. Left: feature matrix showing which components are enabled. Right: solve rates for each configuration. Decay shows the largest observed effect (+10%), followed by examples (+6.7%), though neither reaches statistical significance at $n=30$ . Inhibition shows no detectable effect.
<details>
<summary>x3.png Details</summary>

### Visual Description
## Bar Chart: Full Configuration vs Single Feature Removed
### Overview
This is a vertical bar chart illustrating the impact on a "Solve Rate" when specific features are removed from a full system configuration. The chart quantifies the contribution of three distinct featuresâDecay, Examples, and Inhibitionâby showing the percentage change in the solve rate when each is individually removed.
### Components/Axes
* **Chart Title:** "Full Configuration vs Single Feature Removed" (top center).
* **Y-Axis:**
* **Label:** "Contribution to Solve Rate" (rotated vertically on the left).
* **Scale:** Linear scale with percentage markers at -5%, 0%, 5%, 10%, and 15%.
* **Baseline:** A solid black horizontal line at the 0% mark.
* **X-Axis:**
* **Categories:** Three categorical labels positioned below their respective bars: "Decay", "Examples", "Inhibition".
* **Data Series:** A single series represented by three blue vertical bars. Each bar has a data label directly above it stating its exact value.
* **Legend:** Not present. The categories are directly labeled on the x-axis.
### Detailed Analysis
The chart presents the following data points, from left to right:
1. **Decay:**
* **Visual Trend:** The tallest bar, extending significantly above the 0% baseline.
* **Value:** `+10.0%` (labeled above the bar).
* **Interpretation:** Removing the "Decay" feature from the full configuration results in a 10.0% *increase* in the solve rate.
2. **Examples:**
* **Visual Trend:** A bar of moderate height, shorter than the "Decay" bar but still clearly positive.
* **Value:** `+6.7%` (labeled above the bar).
* **Interpretation:** Removing the "Examples" feature results in a 6.7% *increase* in the solve rate.
3. **Inhibition:**
* **Visual Trend:** No visible bar height; the data point sits exactly on the 0% baseline.
* **Value:** `+0.0%` (labeled above the baseline).
* **Interpretation:** Removing the "Inhibition" feature has no measurable effect (0.0% change) on the solve rate.
### Key Observations
* **Positive Impact of Removal:** Both "Decay" and "Examples" show a positive contribution when removed, indicating their presence in the full configuration *lowers* the solve rate.
* **Magnitude of Impact:** The "Decay" feature has the most significant negative impact on performance (its removal yields the largest gain). "Examples" has a moderate negative impact.
* **Neutral Feature:** The "Inhibition" feature appears to be neutral or non-contributory in this context, as its removal causes no change.
* **Data Presentation:** Values are presented with one decimal place of precision. All changes are expressed as positive percentages relative to the full configuration's baseline.
### Interpretation
This chart is an **ablation study** result, a common technique in machine learning and systems engineering to understand the contribution of individual components. The data suggests the following:
1. **Counterintuitive Findings:** The features "Decay" and "Examples" are not helping the system solve the target problem; they are actively hindering it. Their removal improves performance. This could indicate they are poorly tuned, introduce noise, or are misaligned with the task objective.
2. **Feature Utility:** The "Inhibition" feature, in this specific configuration and evaluation, provides no benefit or detriment. It may be redundant, inactive, or its effects are perfectly balanced by other components.
3. **Optimization Direction:** To improve the system's solve rate, the primary focus should be on understanding why "Decay" and "Examples" are detrimental. The next step would be to either remove them entirely, re-engineer them, or investigate if they are interacting negatively with other parts of the system. The "Inhibition" feature could potentially be removed to simplify the model without affecting performance.
4. **Underlying Assumption:** The chart measures the *marginal contribution* of each feature when absent. The full configuration's solve rate is the implicit baseline (0% change). A positive value means the feature was a net negative in the complete system.
</details>
Figure 3: Individual feature contributions to solve rate. Decay contributes +10.0%, examples contribute +6.7%, and inhibition shows no measurable effect in this domain.
#### 6.3.3 Negative Pheromones
In addition to positive pheromones (successful patches stored for few-shot examples), we implement negative pheromones: tracking rejected patches that worsened pressure. When agents repeatedly propose ineffective patches (pressure stuck at maximum), the system accumulates rejection history and injects guidance into subsequent prompts.
Unlike the âAVOIDâ framing that small models (1.5B parameters) struggle to follow, we use positive language: rejected empty-room patches become âTIP: Schedule meetings in Room A (improves by X).â This reframes what not to do as what to try instead.
Negative pheromones decay at the same rate as positive examples ( $weightĂ 0.95$ per tick, evicted below 0.1), ensuring that old failures donât permanently block valid approaches. Up to 3 recent rejections per region are included in prompts as âHints for better scheduling.â
### 6.4 Scaling Experiments
Pressure-field maintains consistent performance from 1 to 4 agents on easy difficulty:
| 1 2 4 | 25/30 28/30 25/30 | 83.3% 93.3% 83.3% | 66.4%â92.7% 78.7%â98.2% 66.4%â92.7% |
| --- | --- | --- | --- |
Table 6: Pressure-field scaling from 1 to 4 agents (easy difficulty, 30 trials each). Performance remains stable across agent counts.
The result demonstrates robustness: pressure-field coordination maintains consistent solve rates despite 4 $Ă$ variation in agent count. This validates Theorem 5.4: coordination overhead remains $O(1)$ , enabling effective scaling. The slight peak at 2 agents (93.3%) is within Confidence Interval overlap of 1 and 4 agents, indicating no significant agent-count effect.
### 6.5 Band and Model Escalation: FM-MAS Symbiosis in Practice
The escalation mechanism demonstrates the FM-MAS symbiosis that pressure-field coordination enables. Traditional approaches face a dilemma: use a large, capable model everywhere (expensive) or use a small, efficient model everywhere (limited). Pressure-field coordination dissolves this dilemma through gradient-driven capability invocation: the coordination mechanism itself determines when greater Foundation Model capability is needed.
Ant colonies balance exploitation of known food sources against exploration for new ones. When a pheromone trail grows staleâindicating a depleted sourceâforagers abandon trail-following and resume random exploration, eventually discovering new paths that become the next generation of trails. This exploitation-exploration balance is fundamental to stigmergic systems: premature commitment to suboptimal solutions must be counteracted by mechanisms that restore exploratory behavior.
Our escalation mechanism implements this principle through two complementary dynamics. Band escalation governs the exploitation-exploration trade-off within a single model: when pressure velocity drops to zero (the âtrail goes coldâ), sampling parameters shift from exploitation (low temperature, focused proposals) through balanced to exploration (high temperature, diverse proposals). This mirrors the antâs behavioral switch from trail-following to random wandering when pheromone signals weaken.
| Exploitation Balanced Exploration | 0.15â0.35 0.35â0.55 0.55â0.85 | 0.80â0.90 0.85â0.95 0.90â0.98 |
| --- | --- | --- |
Table 7: Sampling parameter ranges per band. Temperature and top-p are randomly sampled within range for diversity. Escalation proceeds Exploitation $â$ Balanced $â$ Exploration as pressure velocity stalls.
Model escalation addresses a different failure mode: when exploration within a modelâs capability envelope fails to discover productive paths, the system recruits more capable Foundation Models. This is where FM-MAS symbiosis becomes concrete: the Multi-Agent System coordination mechanism (pressure-field) adaptively invokes higher-capability Foundation Models based on pressure signals alone. No explicit task decomposition is neededâthe pressure gradient indicates that current capabilities are insufficient, triggering escalation. Model escalation (0.5b $â$ 1.5b $â$ 3b) reserves greater reasoning capacity for regions that resist simpler approaches. Each model upgrade resets to exploitation band, giving the more capable model opportunity to exploit solutions invisible to its predecessor before resorting to exploration.
This architecture instantiates the bidirectional relationship between Foundation Models and Multi-Agent Systems:
- FM contribution: Each model tier provides broad solution coverage without requiring explicit action enumeration. The 3b model can propose patches invisible to the 0.5b model, but both operate through the same interfaceâobserve pressure, propose patch.
- MAS contribution: The pressure gradient provides the objective criterion for when to escalate. No heuristics about âproblem difficultyâ are needed; stagnant pressure velocity is sufficient signal. The coordination mechanism manages heterogeneous Foundation Model capabilities without explicit capability models.
This two-level mechanismâbehavioral adaptation within agents (band escalation) and capability escalation across agents (model escalation)âmaintains the stigmergic principle: coordination emerges from environment signals (pressure gradients) rather than explicit planning. The system does not âdecideâ to explore or escalate; it reacts to pressure stagnation, just as ants react to pheromone decay. The result is adaptive Foundation Model utilization: cheap models handle easy regions, expensive models are reserved for regions where cheap models stall.
### 6.6 Difficulty Scaling
Performance varies across difficulty levels, revealing the unique strength of pressure-field coordination:
| Difficulty Easy Medium | Pressure-field 86.7% (78/90) 43.3% (39/90) | Conversation 33.3% (30/90) 0.0% (0/90) | Hierarchical 4.4% (4/90) 0.0% (0/90) | Sequential/Random 1.1% (1/90) 0.0% (0/90) |
| --- | --- | --- | --- | --- |
| Hard | 15.6% (14/90) | 0.0% (0/90) | 0.0% (0/90) | 0.0% (0/90) |
| Total | 48.5% (131/270) | 11.1% (30/270) | 1.5% (4/270) | 0.4% (1/270) |
Table 8: Solve rate by difficulty level (90 trials each per difficulty, 270 per strategy total). Only pressure-field solves medium and hard problems. See Table 3 for confidence intervals.
The difficulty scaling reveals critical insights:
1. Pressure-field is the only strategy that scales: While all strategies degrade on harder problems, pressure-field maintains meaningful solve rates (43.3% medium, 15.6% hard) where all baselines achieve 0%.
1. The gap widens with difficulty: On easy problems, pressure-field leads by 53.4 percentage points over conversation (86.7% vs 33.3%). On medium and hard problems, the gap becomes absoluteâpressure-field solves problems that no baseline can solve.
1. Effect sizes are large: All pairwise comparisons on easy problems exceed Cohenâs âlarge effectâ threshold ( $h>0.8$ ); see Figure 4 for the full breakdown.
<details>
<summary>x4.png Details</summary>

### Visual Description
## Forest Plot: Pressure Field vs Baselines (Easy Difficulty)
### Overview
This image is a forest plot (effect size plot) comparing the performance of a method called "Pressure Field" against four different baseline methods in an "Easy Difficulty" task. The plot visualizes the effect size (Cohen's h) for each comparison, including point estimates and confidence intervals. All comparisons show a positive effect size favoring Pressure Field.
### Components/Axes
* **Chart Title:** "Pressure Field vs Baselines (Easy Difficulty)"
* **X-Axis:**
* **Label:** "Cohen's h (Effect Size)"
* **Scale:** Linear scale from 0.0 to 2.5.
* **Markers:** Major tick marks at 0.0, 0.5, 1.0, 1.5, 2.0, 2.5.
* **Y-Axis (Categories):** Lists the four baseline comparisons. From top to bottom:
1. "vs Conversation"
2. "vs Hierarchical"
3. "vs Sequential"
4. "vs Random"
* **Reference Lines:**
* A vertical **gray dashed line** at x = 0.0, representing no effect.
* A vertical **red dotted line** at x = 0.8, annotated with the text "Large effect threshold (0.8)".
* **Data Series:** Each comparison is represented by a colored point (the point estimate) with horizontal error bars (the confidence interval).
* **vs Conversation:** Purple point and error bars.
* **vs Hierarchical:** Green point and error bars.
* **vs Sequential:** Red point and error bars.
* **vs Random:** Gray point and error bars.
### Detailed Analysis
The plot presents the following effect size estimates (Cohen's h) for Pressure Field versus each baseline. Values are approximate based on visual inspection of the chart.
1. **vs Conversation (Purple):**
* **Trend:** The point estimate is clearly to the right of the large effect threshold (0.8).
* **Point Estimate:** ~1.2
* **Confidence Interval:** Spans from approximately 0.9 to 1.5.
2. **vs Hierarchical (Green):**
* **Trend:** Shows a very large effect size, well beyond the threshold.
* **Point Estimate:** ~2.0
* **Confidence Interval:** Spans from approximately 1.7 to 2.3.
3. **vs Sequential (Red):**
* **Trend:** Exhibits the largest point estimate among the four comparisons.
* **Point Estimate:** ~2.2
* **Confidence Interval:** Spans from approximately 1.9 to 2.5.
4. **vs Random (Gray):**
* **Trend:** Also shows a very large effect size.
* **Point Estimate:** ~2.1
* **Confidence Interval:** Spans from approximately 1.8 to 2.4.
### Key Observations
* **Universal Large Effect:** All four point estimates and their entire confidence intervals lie to the right of the "Large effect threshold (0.8)" line. This indicates that Pressure Field demonstrates a statistically and practically significant improvement over every baseline method in the easy difficulty setting.
* **Magnitude of Effect:** The effect sizes are substantial, ranging from approximately 1.2 to 2.2. In the context of Cohen's h, values above 0.8 are considered large, and values around 2.0 are very large.
* **Relative Performance:** The effect size is smallest when comparing to the "Conversation" baseline (~1.2) and largest when comparing to the "Sequential" baseline (~2.2). The effects against "Hierarchical" and "Random" are very similar and large (~2.0-2.1).
* **Precision:** The confidence intervals for "vs Conversation" appear slightly narrower than the others, suggesting a more precise estimate for that comparison. All intervals are of moderate width, indicating reasonable certainty in the estimated effect sizes.
### Interpretation
This forest plot provides strong, quantitative evidence that the "Pressure Field" method significantly outperforms a range of alternative approaches (Conversation, Hierarchical, Sequential, and Random) on an easy-difficulty task. The data suggests:
1. **Robust Superiority:** Pressure Field's advantage is not limited to a single type of baseline; it is consistently and substantially better across different comparison points.
2. **Practical Significance:** The effect sizes are not just statistically significant (as implied by confidence intervals not crossing zero) but are also large in a practical sense, exceeding the conventional threshold for a "large effect." This implies the performance difference is meaningful and likely observable in real-world applications of the task.
3. **Insight into Baselines:** The varying effect sizes hint at the relative difficulty of the baselines. The smallest effect against "Conversation" might suggest it is the strongest or most similar baseline to Pressure Field among those tested. The very large effects against "Sequential" and "Random" indicate these are much weaker approaches for this specific easy task.
4. **Conclusion:** For the "Easy Difficulty" scenario, adopting the Pressure Field method over any of the presented baselines would be expected to yield a major improvement in performance. The chart effectively communicates the strength and consistency of this advantage.
</details>
Figure 4: Effect sizes (Cohenâs $h$ ) for pressure-field versus each baseline on easy problems. The dashed line indicates the âlarge effectâ threshold ( $h=0.8$ ). All comparisons exceed this threshold, with effects ranging from $h=1.16$ (vs conversation) to $h=2.18$ (vs sequential/random).
### 6.7 Convergence Speed
For solved cases, pressure-field converges faster than baselines on easy problems and maintains consistent convergence speed across difficulty levels:
| Pressure-field Conversation Hierarchical | 17.8 (n=78) 29.4 (n=30) 40.0 (n=4) | 34.6 (n=39) â â | 32.3 (n=14) â â |
| --- | --- | --- | --- |
Table 9: Average ticks to solution by difficulty (solved cases only). Only pressure-field solves medium and hard problems. Dashes indicate no solved cases.
On easy problems, pressure-field solves 1.65 $Ă$ faster than conversation and 2.2 $Ă$ faster than hierarchical. Notably, pressure-fieldâs convergence speed on hard problems (32.3 ticks) is comparable to medium problems (34.6 ticks)âthe hard problems that do get solved converge at similar rates, suggesting that solvability rather than convergence speed is the limiting factor on difficult problems. This bimodal patternâfast convergence when solvable, complete failure otherwiseâsuggests that model capability rather than search time is the limiting factor on hard problems.
<details>
<summary>x5.png Details</summary>

### Visual Description
## Bar Chart: Mean Ticks to Solution by Method (Easy Difficulty)
### Overview
This is a vertical bar chart titled "Easy Difficulty (Solved Trials Only)". It compares the performance of three different methods or conditionsâPressure Field, Conversation, and Hierarchicalâbased on the average number of "ticks" required to reach a solution. The chart includes error bars for each data point, indicating variability or uncertainty in the measurements.
### Components/Axes
* **Chart Title:** "Easy Difficulty (Solved Trials Only)" (located at the top-left).
* **Y-Axis:**
* **Label:** "Mean Ticks to Solution" (rotated vertically on the left side).
* **Scale:** Linear scale from 0 to 50, with major gridlines and numerical markers at intervals of 10 (0, 10, 20, 30, 40, 50).
* **X-Axis:**
* **Categories (from left to right):** "Pressure Field", "Conversation", "Hierarchical". The labels are rotated approximately 45 degrees for readability.
* **Data Series:** Three distinct bars, each representing a category. The legend is implicit, with color directly associated with the x-axis label.
* **Pressure Field:** Blue bar.
* **Conversation:** Purple bar.
* **Hierarchical:** Green bar.
* **Error Bars:** Each bar has a black, vertical error bar (likely representing standard error or confidence interval) extending above and below the top of the bar.
### Detailed Analysis
The chart presents the following specific data points and visual trends:
1. **Pressure Field (Blue Bar, Leftmost):**
* **Mean Value:** 17.8 (explicitly labeled above the bar).
* **Visual Trend:** This is the shortest bar, indicating the lowest mean ticks to solution.
* **Error Bar:** Extends approximately from 16.0 to 19.5 on the y-axis. The range is relatively narrow.
2. **Conversation (Purple Bar, Center):**
* **Mean Value:** 29.4 (explicitly labeled above the bar).
* **Visual Trend:** This bar is taller than Pressure Field but shorter than Hierarchical, placing it in the middle.
* **Error Bar:** Extends approximately from 27.5 to 31.5 on the y-axis. The range is moderate.
3. **Hierarchical (Green Bar, Rightmost):**
* **Mean Value:** 40.0 (explicitly labeled above the bar).
* **Visual Trend:** This is the tallest bar, indicating the highest mean ticks to solution.
* **Error Bar:** Extends approximately from 34.0 to 46.0 on the y-axis. This is the widest error bar, suggesting the greatest variability in performance for this method.
### Key Observations
* **Clear Performance Hierarchy:** There is a distinct, stepwise increase in the mean ticks to solution from Pressure Field (lowest) to Conversation to Hierarchical (highest).
* **Increasing Variability:** The size of the error bars increases from left to right. The Hierarchical method not only has the highest mean but also the largest spread of data, indicating less consistent performance compared to the other two methods.
* **No Overlap in Central Tendency:** The mean values (17.8, 29.4, 40.0) are well-separated. The error bars for Pressure Field and Hierarchical do not overlap at all, and the error bar for Conversation only slightly overlaps with the upper range of Pressure Field's error bar.
### Interpretation
The data suggests a strong relationship between the method used and efficiency in solving "easy difficulty" tasks, as measured by ticks (a unit of time or computational steps).
* **Efficiency vs. Complexity:** The "Pressure Field" method is the most efficient, requiring roughly 55% fewer ticks than the "Conversation" method and less than half the ticks of the "Hierarchical" method. This could imply that for simple problems, a more direct or physics-based approach (Pressure Field) outperforms methods involving interaction (Conversation) or structured organization (Hierarchical).
* **Trade-off with Consistency:** The increasing size of the error bars introduces a critical nuance. While the Hierarchical method is the slowest on average, its performance is also the most variable. This might mean it occasionally solves problems very quickly or very slowly, whereas the Pressure Field method is more reliably fast.
* **Implication for System Design:** If the goal is to minimize average solution time for easy tasks, the Pressure Field method is the clear choice. However, if understanding the range of possible outcomes is important, the high variability of the Hierarchical method would be a significant drawback. The Conversation method represents a middle ground in both average speed and consistency.
**Language Note:** All text in the image is in English.
</details>
Figure 5: Mean ticks to solution on easy problems (solved cases only). Error bars show standard error. Pressure-field converges fastest (17.8 ticks), followed by conversation (29.4 ticks) and hierarchical (40.0 ticks).
### 6.8 Final Pressure Analysis
For both solved and unsolved cases, final pressure reveals solution quality:
| Pressure-field Conversation Hierarchical | 0.13 26.5 28.6 | 1.02 59.3 59.4 | 2.48 92.2 88.0 |
| --- | --- | --- | --- |
Table 10: Average final pressure by difficulty (lower is better).
Pressure-field achieves 200 $Ă$ lower final pressure on easy, 57 $Ă$ lower on medium, and 35 $Ă$ lower on hard problems compared to conversation and hierarchical baselines. Even when pressure-field does not fully solve a problem, it achieves much better partial solutions.
### 6.9 Token Efficiency
A natural question is whether pressure-fieldâs superior solve rate comes at the cost of increased Large Language Model usage. We tracked token consumption across all strategies:
| Pressure-field Conversation Hierarchical | 532,106 154,845 32,662 | 85,165 6,626 5,424 | 617,271 161,472 38,087 |
| --- | --- | --- | --- |
| Sequential | 40,667 | 5,239 | 45,906 |
| Random | 41,227 | 5,253 | 46,480 |
Table 11: Average token usage per trial. Pressure-field uses more tokens per trial due to parallel multi-agent execution.
At first glance, pressure-field appears expensive: 617K tokens per trial versus conversationâs 161Kâroughly 4 $Ă$ higher. However, this raw comparison ignores success rate. The relevant metric for practical deployment is tokens per successful solve:
| Pressure-field | 166.7M | 131 | 1.27M |
| --- | --- | --- | --- |
| Conversation | 43.6M | 30 | 1.45M |
Table 12: Token efficiency per successful solve (270 trials each). Despite higher per-trial cost, pressure-field achieves lower cost per solve due to its superior success rate.
Pressure-field requires 1.27M tokens per solve versus conversationâs 1.45Mâ 12% more efficient per successful outcome. The apparent cost disadvantage inverts when accounting for success rate.
Token usage also varies by outcome:
| Pressure-field Conversation | 318K (n=131) 78K (n=30) | 900K (n=139) 172K (n=240) |
| --- | --- | --- |
Table 13: Average tokens by outcome. Both strategies use fewer tokens on solved trials (early termination), but pressure-fieldâs unsolved trials are more expensive due to sustained parallel exploration.
Solved trials terminate early, using fewer tokens. For pressure-field, the 2.8 $Ă$ gap between solved (318K) and unsolved (900K) reflects the cost of exhaustive exploration when a problem proves intractable. Conversationâs smaller gap (2.2 $Ă$ ) indicates less intensive searchâwhich may explain its lower solve rate on difficult problems.
Cost control through escalation. The band and model escalation mechanism (Section 6.5) serves as an implicit cost-control mechanism. The system begins with cheap exploration: small models (0.5b) and exploitation-focused sampling. Tokens are spent on larger models and broader exploration only when pressure stagnates âthat is, when the problem actually requires more capable search. Easy problems that solve quickly never trigger escalation, consuming only baseline tokens. The 4 $Ă$ per-trial cost difference reflects the average across all difficulty levels; on easy problems where pressure-field solves 86.7% of instances, most trials terminate before expensive escalation occurs.
## 7 Discussion
### 7.1 Why Does Pressure-Field Dominate?
Our results contradict the intuition that explicit coordination should outperform implicit coordination. We identify three factors explaining pressure-fieldâs dominance:
Coordination overhead harms performance. Hierarchical systems spend computational budget on coordination rather than problem-solving. The manager-worker protocol requires multiple Large Language Model calls per patch (planning, delegation, execution), while pressure-field requires only one (patch proposal). This overhead compounds: hierarchical attempts fewer patches per tick, reducing exploration.
Local greedy decisions are effective for constraint satisfaction. Meeting room scheduling exhibits locality: fixing a conflict in one time block rarely creates conflicts in distant blocks. This matches pressure-fieldâs locality assumption, making greedy local optimization effective. Hierarchical coordinationâs global planning provides no benefit for locally-decomposable problems.
Parallel validation amplifies pressure-fieldâs advantage. Pressure-field validates patches for multiple regions simultaneously, applying the highest-scoring patch per region that reduces pressure. Hierarchical validates one patch at a time, requiring multiple ticks to explore alternatives. On problems with many valid solutions, parallel exploration finds solutions faster.
### 7.2 Failure Analysis: Why Hierarchical Collapses
The 30 $Ă$ performance gap (48.5% vs 1.5%) demands deeper investigation. Analysis of hierarchical trials reveals a catastrophic failure pattern: the rejection loop.
The rejection loop mechanism. Hierarchical always selects the highest-pressure region for improvement. But the highest-pressure regions are high-pressure precisely because they are difficult to improve. When the Large Language Model proposes a patch that fails validation (does not reduce pressure), the region remains highest-pressure and is selected again next tick. This creates a self-reinforcing cycle: 66.7% of hierarchical runs (180/270) applied zero patches across all 50 ticks, stuck targeting the same intractable region repeatedly.
Rejection rates confirm the pattern. Across all hierarchical trials, only 173 patches were accepted out of 13,460 proposedâa 98.7% rejection rate. By contrast, pressure-fieldâs parallel exploration means that even if one agentâs patch is rejected, other agents make progress on different regions. The non-blocking architecture prevents any single difficult region from stalling the entire system.
The architectural lesson. Hierarchicalâs design embodies a reasonable intuition: focus intelligent effort on the worst problems. But this creates a trap when combined with strict validation: the hardest problems resist improvement, causing repeated rejection, which blocks progress everywhere. Pressure-field avoids this trap through distributed explorationâprogress happens where it can, not where a central planner dictates it must.
### 7.3 Limitations
Our experiments reveal several important limitations:
Absolute solve rates are modest on hard problems. Even pressure-field achieves only 15.6% on hard problems. Meeting room scheduling with tight constraints (5 rooms, 60 meetings, 30% pre-scheduled) remains challenging for small models (0.5bâ3b parameters); larger models may achieve higher absolute solve rates.
Domain specificity. Results on meeting room scheduling may not generalize to domains lacking measurable pressure gradients or locality properties. Tasks requiring global planning or long-horizon reasoning may favor hierarchical approaches.
However, meeting scheduling is representative of a broader class of resource allocation problems with soft constraints: cloud compute scheduling (VMs to hosts with affinity/anti-affinity rules), logistics optimization (packages to routes with capacity and time windows), and workforce rostering (employees to shifts with skill and availability constraints). These domains share the structural properties that enable pressure-field coordination: (1) decomposable quality metrics (utilization, constraint violations, balance), (2) locality (fixing one allocation rarely breaks distant allocations), and (3) multiple valid solutions (no single optimum, many acceptable configurations). We selected meeting scheduling as the evaluation domain because it is intuitive, reproducible, and exhibits these properties clearly. The coordination mechanism itselfâpressure gradients, temporal decay, parallel validationâis domain-agnostic; only the pressure function requires domain-specific design.
Additional practical limitations:
- Requires well-designed pressure functions (not learned from data)
- Decay rates $λ_f,λ_γ$ and inhibition period require task-specific tuning
- May not suit tasks requiring long-horizon global planning
- Goodhartâs Law: agents may game poorly-designed metrics
- Resource cost of parallel validation: testing $K$ patches requires $O(K·|A|)$ memory where $|A|$ is artifact size
#### 7.3.1 Hallucination Filtering and Emergent Trajectories
The validation phase (Phase 2b) serves as a filter for individual Foundation Model hallucinations: patches that increase pressure or violate syntactic constraints are rejected before application. This provides strong guarantees at the individual-action levelâno single hallucination can directly harm artifact quality.
However, system-level behavior emerges from sequences of validated patches, and trajectory-level risks remain. Consider a hypothetical scenario: an agent proposes a patch that reduces pressure in region $R_1$ by making a change that creates a subtle dependency on region $R_2$ . The patch is validated (pressure decreased), but subsequent patches to $R_2$ now have unpredictable effects on $R_1$ . The individual patches are all pressure-reducing, but the emergent trajectory creates fragile coupling that was invisible to local validation.
More generally, validation filters based on pressure reduction cannot detect:
- Coherence drift: Individual improvements that collectively shift the artifact toward an inconsistent state
- Emergent gaming: Patches that exploit pressure function weaknesses in ways only apparent over multiple steps
- Dependency accumulation: Gradual introduction of hidden couplings that reduce future improvability
Mitigating trajectory-level risks requires mechanisms beyond local validation: periodic global coherence checks, trajectory logging for post-hoc analysis, and pressure functions that penalize not just local quality but also coupling metrics. These remain areas for future work.
#### 7.3.2 Decay Miscalibration Failure Modes
Temporal decay is critical for escaping local optima (Theorem 5.3), but miscalibrated decay rates introduce distinct failure modes:
Too-fast decay prevents stability. When fitness decays faster than agents can reinforce successful regions, the system cannot maintain any stable configuration. Solved regions immediately become high-pressure again, triggering unnecessary rework. In the limit, excessively fast decay produces perpetual oscillation: agents patch, decay erases, agents re-patch, indefinitely.
Too-slow decay traps in local minima. When fitness decays slower than the exploration timescale, agents cannot escape suboptimal basins. Early patches establish fitness peaks that persist indefinitely, preventing reconsideration even when better solutions exist. This is the failure mode our ablation study suggests: without decay, we observed a 10 percentage point reduction in solve rate, consistent with agents remaining trapped in initial basins (though the effect did not reach statistical significance at $n=30$ ).
The optimal decay rate depends on problem characteristics: harder problems with deeper local minima require faster decay to enable escape, while problems with fragile solutions require slower decay to maintain stability. Our experiments use fixed decay rates ( $λ_f=0.1$ , fitness half-life 5 seconds), which may be suboptimal for some problem instances. Adaptive decayâadjusting rates based on pressure velocityâis a promising direction for future work.
### 7.4 When to Choose Each Approach
Our results suggest the following guidance:
Pressure-field coordination is preferable when:
1. Performance matters. Pressure-field achieves 3â30 $Ă$ higher solve rates than alternatives.
1. Simplicity is valued. No coordinator agent needed; coordination emerges from shared state.
1. Fault tolerance matters. No single point of failure; agents can join/leave without protocol overhead.
1. Pressure signals are available. The domain provides measurable quality gradients.
1. Problems are locally decomposable. Local fixes improve global quality without cascading conflicts.
Hierarchical coordination may be appropriate when:
1. Explicit control is required. Some domains require deterministic task assignment for regulatory or safety reasons.
1. Interpretability is critical. Hierarchical task assignment provides clear audit trails.
1. Global planning is essential. Tasks with strong non-local dependencies may benefit from centralized reasoning.
### 7.5 Band and Model Escalation as Adaptive Capability
All experiments use a two-level escalation mechanism. Band escalation cycles through sampling strategies (Exploitation $â$ Balanced $â$ Exploration, 7 ticks each) before model escalation progresses through model sizes (0.5b $â$ 1.5b $â$ 3b parameters). Model escalation triggers when regions remain high-pressure for 21 consecutive ticks.
This mechanism proves beneficial for pressure-field: on hard problems, pressure-field achieves 15.6% with escalation enabled. The escalation mechanism works because larger models have broader solution coverage and different sampling bands explore different regions of solution space.
### 7.6 Future Work
- Learned pressure functions: Current sensors are hand-designed. Can we learn pressure functions from solution traces?
- Adversarial robustness: Can malicious agents exploit pressure gradients to degrade system performance?
- Multi-artifact coordination: Extension to coupled artifacts where patches in one affect pressure in another
- Larger-scale experiments: Testing on schedules with more rooms and longer time horizons to characterize scaling limits
- Alternative domains: Applying pressure-field coordination to code refactoring, configuration management, and other artifact refinement tasks
### 7.7 FM-MAS Reciprocity: Bidirectional Problem Solving
The intersection of Foundation Models and Multi-Agent Systems is not merely additiveâeach paradigm solves fundamental problems that the other cannot address alone. This reciprocity suggests that pressure-field coordination represents a genuine synthesis rather than a simple combination.
#### 7.7.1 Foundation Models Solve a MAS Coordination Problem
Traditional Multi-Agent System coordination requires explicit action space enumeration. Generalized Partial Global Planning coordinates tasks through declared capabilities and commitments; SharedPlans reasons about action sequences and their preconditions; organizational models assign roles with specified behaviors. All these approaches assume that designers can enumerate what agents can do and under what conditions.
For open-ended artifact refinementâimproving code quality, refining documents, optimizing configurationsâthis enumeration is intractable. The space of possible improvements is unbounded; no finite action language captures all valid patches. Traditional Multi-Agent System approaches require either (a) restricting the problem to a tractable subset with enumerable actions, or (b) developing domain-specific action representations for each artifact type.
Foundation Models eliminate this enumeration requirement. Their broad pretraining provides implicit coverage of improvement strategies across diverse domains without explicit action specification. An Large Language Model actor can propose patches for code, natural language, structured data, or configuration files using the same interface: observe local context, receive pressure feedback, generate improvement proposals. This âuniversal actorâ capability enables stigmergic coordination on problem classes where traditional Multi-Agent System approaches would require extensive domain engineering.
#### 7.7.2 Multi-Agent Systems Solve a FM Problem
Conversely, Foundation Models face a fundamental problem that Multi-Agent System coordination solves: how to combine multiple outputs coherently. A single Foundation Model produces one response per query; scaling to complex artifacts requires orchestrating multiple generations. Current approaches use ad-hoc combination strategies: voting, ranking, chain-of-thought aggregation, or human-in-the-loop selection.
Pressure-field coordination provides a principled framework for combining Foundation Model outputs. Rather than voting on which response is âbestâ or ranking outputs by heuristic scores, the pressure gradient defines an objective criterion: accept patches that reduce pressure, reject those that do not. Multiple Foundation Model outputs compete not through popularity or arbitrary ranking, but through their effect on a well-defined quality function.
This framing clarifies why pressure-field coordination outperforms conversation baselines. AutoGen-style multi-agent dialogue is an ad-hoc combination strategyâagents exchange messages and reach conclusions through emergent consensus. Pressure-field coordination replaces emergent consensus with objective gradients: no negotiation is needed when the artifact state adjudicates quality.
#### 7.7.3 Implications for FM-MAS Integration
This reciprocity suggests design principles for future FM-MAS systems:
1. Leverage FM coverage, constrain via MAS gradients. Use Foundation Models for their broad solution coverage, but use Multi-Agent System coordination mechanisms to filter and combine outputs objectively.
1. Prefer stigmergic over explicit coordination. When Foundation Models serve as actors, stigmergic coordination avoids the overhead of explicit protocols that FMs may not reliably follow.
1. Design pressure functions as FM-MAS interfaces. The pressure function is the critical interface: it must capture human intent precisely enough that Foundation Model outputs reducing pressure are genuinely improvements.
Our experimental resultsâpressure-field achieving $4Ă$ higher solve rates than conversation baselinesâprovide empirical support for these principles. The coordination overhead of explicit dialogue exceeds its organizational benefit when objective gradients are available.
### 7.8 Societal Implications
Pressure-field coordination raises societal concerns that extend beyond technical performance. We identify three critical issuesâaccountability attribution, metric gaming through Goodhartâs Law, and explainability challengesâthat require deliberate design choices in deployment.
#### 7.8.1 Accountability and Attribution
When coordination emerges from shared pressure gradients rather than explicit delegation, attributing outcomes to individual agents becomes challenging. In hierarchical systems, task assignment creates clear accountability chains. In pressure-field coordination, multiple agents may contribute to a region through independent pressure-reducing actions, with no record of which agent âownedâ the outcome.
This accountability diffusion has both benefits and risks. The benefit is fault tolerance: agent failures degrade performance gracefully rather than catastrophically. The risk is opacity in failure analysis: identifying which agent proposed a problematic patchâand what pressure signal motivated itârequires detailed logging that the minimal coordination mechanism does not inherently provide.
For deployment in regulated domains, this suggests an augmentation requirement: pressure-field systems must maintain audit logs recording patch provenance, pressure signals at proposal time, and validation outcomes. The coordination mechanism remains simpleâagents coordinate through shared stateâbut operational deployment adds logging infrastructure preserving accountability.
#### 7.8.2 Goodhartâs Law and Metric Gaming
Goodhartâs Law states: âWhen a measure becomes a target, it ceases to be a good measure.â Pressure-field coordination is vulnerable to this dynamic because agents are optimized to reduce pressure as defined by designer-specified functions. If those functions imperfectly capture true qualityâand they inevitably doâagents will discover and exploit the mismatch.
Consider code quality pressure functions penalizing complexity metrics. An agent might reduce complexity by splitting functions excessively, harming readability while improving the metric. The mitigation is not abandoning pressure functions but designing them defensively: use multiple orthogonal pressure axes, include adversarial sensors detecting gaming strategies, and audit whether pressure reduction correlates with human quality judgments. Pressure functions should evolve as agents discover exploits.
Foundation models introduce second-order gaming concerns: Large Language Models trained on internet-scale text may have implicit knowledge of how to game specific benchmarks. This suggests pressure functions for Large Language Model-based systems should favor domain-specific quality signals harder to optimize without genuine improvement.
#### 7.8.3 Explainability Challenges
In hierarchical systems, explanations follow delegation chains: âManager X assigned task Y to Worker Z because condition C held.â In pressure-field coordination, the explanation is: âRegion R had high pressure, agent A proposed patch $Î$ reducing pressure by $ÎŽ$ .â This is mechanistically transparent but causally opaqueâit describes what happened without explaining why that particular patch was chosen.
This is the explainability trade-off inherent to emergent coordination: simplicity in mechanism comes at the cost of legibility in rationale. For many domainsâcode formatting, resource optimization, routine maintenanceâthe trade-off is acceptable: outcomes are verifiable even if reasoning is opaque. For high-stakes domains requiring human oversight, opacity is unacceptable.
The design implication is domain-dependent deployment: pressure-field coordination suits domains where outcome verification is cheap even if reasoning transparency is limited. For domains requiring justification to human stakeholders, hierarchical coordination remains necessary despite overhead costs.
#### 7.8.4 Design Implications
These concerns suggest three requirements for responsible deployment: comprehensive audit logging preserving patch provenance and pressure signals, defensive pressure function design with multiple orthogonal axes, and domain-appropriate verification matching coordination opacity with outcome verifiability. The coordination mechanism remains simpleâbut responsible deployment requires surrounding infrastructure addressing accountability, gaming, and explainability.
## 8 Conclusion
We presented pressure-field coordination, a decentralized approach to Multi-Agent Systems that achieves coordination through shared state and local pressure gradients rather than explicit orchestration.
Our theoretical analysis establishes convergence guarantees under pressure alignment conditions, with coordination overhead independent of agent count. Empirically, on meeting room scheduling across 1350 total trials (270 per strategy), we find:
1. Implicit coordination outperforms explicit coordination. Pressure-field achieves 48.5% aggregate solve rateânearly half of all problems solved through local pressure-following alone, with no coordinator, no message passing, and no explicit task delegation. The gap versus hierarchical control ( $30Ă$ ) and conversation-based coordination ( $4Ă$ ) is both large and highly significant (all $p<0.001$ ).
1. Pressure-field is the only strategy that scales to harder problems. On medium and hard problems, pressure-field achieves 43.3% and 15.6% solve rates respectively, while all baselines achieve 0%.
1. Temporal decay shows beneficial effects. Ablation studies observe a 10 percentage point improvement with decay enabled, consistent with theoretical predictions about escaping local minima, though larger samples would be needed to establish statistical significance.
This work demonstrates that implicit coordination can outperform explicit coordination for constraint satisfaction tasks. Pressure-field achieves this with simpler architecture: no coordinator agent, no explicit message passing, just shared state and local pressure gradients.
Foundation Modelsâ zero-shot capabilities eliminate the need for domain-specific action representations; pressure-field coordination eliminates the need for complex multi-agent protocols; together they enable simple Multi-Agent Systems.
These results suggest that for domains with measurable quality signals and locally-decomposable structure, implicit coordination through shared state offers not just a simpler but a more effective alternative to explicit hierarchical control.
Code Availability
Code is available at https://github.com/Govcraft/pressure-field-experiment.
Acknowledgments
The author thanks the anonymous reviewers for their constructive feedback.
## Declarations
Funding
No funding was received for conducting this study.
Competing Interests
The author has no relevant financial or non-financial interests to disclose.
Data Availability
All experimental data and analysis scripts are available in the code repository.
Use of AI Tools
The author used Claude (Anthropic) to provide feedback on draft versions of this manuscript. All analytical decisions, experimental design, and final content were made by the author, who takes full responsibility for the work.
## Appendix A Experimental Protocol
This appendix provides complete reproducibility information for all experiments.
### A.1 Hardware and Software
Hardware: NVIDIA RTX 4070 8GB Graphics Processing Unit, AMD Ryzen 9 7940HS, 64GB RAM
Software:
- Rust 1.75+ (edition 2024)
- Ollama (local Large Language Model inference server)
- Models: qwen2.5:0.5b, qwen2.5:1.5b, qwen2.5:3b
### A.2 Model Configuration
Models are served via Ollama with a system prompt configured for schedule optimization:
âŹ
You optimize meeting room schedules. Given a schedule with gaps or conflicts,
propose ONE change: move, swap, or reschedule a meeting to reduce gaps,
overlaps, and utilization variance. Return ONLY your proposed change
in the format: MOVE meeting_id TO room day start_time
For multi-model setups (model escalation), models share a single Ollama instance with automatic routing based on model name.
### A.3 Sampling Diversity
The experiment framework overrides default sampling parameters with three exploration bands per Large Language Model call:
| Exploitation Balanced Exploration | 0.15â0.35 0.35â0.55 0.55â0.85 | 0.80â0.90 0.85â0.95 0.90â0.98 |
| --- | --- | --- |
Table 14: Sampling parameter ranges. Each LLM call randomly samples from one band.
This diversity prevents convergence to local optima and enables exploration of the solution space.
### A.4 Baseline Implementation Details
To ensure fair comparison, all strategies use identical Large Language Model prompts and parsing logic. The only differences are region selection and patch validation:
Prompt (identical across all strategies):
âŹ
You are a meeting room scheduler. Output schedules in the exact format requested.
Meeting Room Schedule Optimization.
Goal: Schedule meetings to minimize gaps and avoid conflicts.
Time Block: [time range]
Rooms:
Room A: capacity 10
Room B: capacity 8
...
Current assignments:
[current schedule for this time block]
Unscheduled meetings that could fit in this block:
Meeting 5: 60 min, 4 attendees
Meeting 12: 30 min, 2 attendees
...
Constraints:
- No attendee can be in multiple meetings at the same time
- Room capacity must fit attendees
Output the schedule for this time block.
Region selection per strategy:
- Pressure-field: Multiple agents select regions in parallel based on pressure gradients
- Hierarchical: Always selects the single highest-pressure region
- Sequential: Round-robin through regions in fixed order
- Random: Uniform random region selection
Patch validation per strategy:
- Pressure-field: Validates all patches; applies only those reducing pressure
- Hierarchical: Validates single patch; applies only if pressure reduces
- Sequential/Random: No validation; applies any syntactically valid patch
Why âone patch per tickâ for hierarchical? This asymmetry is inherent to the coordination paradigm, not an implementation choice. Hierarchical control requires the manager to: (1) observe global state, (2) select a target region, (3) delegate to a worker, (4) receive the proposed patch, and (5) validate before applying. Delegating to $K$ workers simultaneously would require additional coordination protocolsâwork distribution to avoid conflicts, result aggregation, tie-breaking when multiple patches target overlapping stateâthat would transform hierarchical control into a fundamentally different architecture. The sequential bottleneck is the cost of centralized decision-making. Pressure-field avoids this overhead through stigmergic coordination: agents observe shared state independently and propose patches without central delegation, enabling natural parallelism.
Conversation baseline implementation The conversation baseline implements AutoGen-style multi-agent dialogue with three roles: a Coordinator that selects which region to address, a Proposer that suggests schedule changes, and a Validator that critiques proposals. Each tick proceeds as follows: (1) the Coordinator identifies a high-pressure region (1 Large Language Model call), (2) the Proposer and Validator engage in up to 5 dialogue turns (2 calls per turn), (3) if the Validator approves, the patch is applied. This yields 4â12 Large Language Model calls per tick for a single region, compared to pressure-fieldâs $K$ parallel calls across $K$ regions. The conversation terminates when the Validator approves or the turn limit is reached. We use fixed sampling parameters (T=0.3, top-p=0.9) for structured dialogue, lower than pressure-fieldâs exploration bands. Note that this represents one instantiation of conversation-based coordination inspired by AutoGenâs design principles; actual AutoGen deployments may use different role configurations, turn protocols, or termination conditions.
Fairness guarantees for baseline comparison: To ensure the conversation baseline had equal opportunity to succeed:
- Same models: All strategies use identical model chains (qwen2.5:0.5b $â$ 1.5b $â$ 3b) with the same escalation triggers
- Same problem access: All strategies observe identical schedule state, room capacities, meeting requirements, and constraint information
- Retry logic: The conversation baseline includes up to 5 dialogue turns per region, allowing the Proposer to refine proposals based on Validator feedbackâthis is more opportunity for refinement than pressure-fieldâs single-shot proposals
- Same tick budget: All strategies receive the same 50-tick limit to solve each problem
- Output parsing: All strategies use identical response parsing and schedule extraction logic
The only intentional differences are the coordination mechanisms themselves: how regions are selected, how proposals are generated, and how patches are validated.
### A.5 Problem Generation and Seeding
Fair strategy comparison requires identical problem instances: each strategy must face the same scheduling challenge within a trial. We achieve this through deterministic seeding.
Each trial generates its problem from a seed:
$$
seed=trialĂ 1000+agent\_count
$$
Trial 5 with 2 agents yields seed 5002, producing identical meeting configurations whether evaluated with pressure-field, conversation, or hierarchical coordination.
The seed governs all stochastic generation:
- Meeting durations (1â4 time slots)
- Attendee assignments (2â5 participants)
- Room preferences and capacity requirements
- Pre-scheduled vs. unassigned meeting distribution
- Time slot availability patterns
### A.6 Experiment Commands
Main Grid (Strategy Comparison):
âŹ
schedule - experiment -- host http:// localhost:11434 \
grid -- trials 30 \
-- strategies pressure_field, sequential, random, hierarchical, conversation \
-- agents 1,2,4 -- difficulties easy, medium, hard \
-- max - ticks 50
Ablation Study:
âŹ
schedule - experiment -- host http:// localhost:11434 \
ablation -- trials 30 -- agents 2 -- difficulty easy -- max - ticks 50
Scaling Analysis:
âŹ
schedule - experiment -- host http:// localhost:11434 \
grid -- trials 30 \
-- strategies pressure_field \
-- agents 1,2,4 -- difficulties easy \
-- max - ticks 50
Band and Model Escalation:
âŹ
# Full escalation chain enabled by default
# Band escalation: Exploitation -> Balanced -> Exploration (7 ticks each)
# Model escalation: 0.5 b -> 1.5 b -> 3 b (after 21 ticks at high pressure)
schedule - experiment -- host http:// localhost:11434 \
grid -- trials 30 -- agents 2 -- difficulties medium \
-- max - ticks 100
Difficulty Scaling:
âŹ
# Easy: 3 rooms, 20 meetings, 70% pre - scheduled
schedule - experiment -- host http:// localhost:11434 \
grid -- trials 30 -- agents 2 -- difficulties easy -- max - ticks 50
# Medium: 5 rooms, 40 meetings, 50% pre - scheduled
schedule - experiment -- host http:// localhost:11434 \
grid -- trials 30 -- agents 2 -- difficulties medium -- max - ticks 50
# Hard: 5 rooms, 60 meetings, 30% pre - scheduled
schedule - experiment -- host http:// localhost:11434 \
grid -- trials 30 -- agents 2 -- difficulties hard -- max - ticks 100
### A.7 Metrics Collected
Each experiment records:
- solved: Boolean indicating all meetings scheduled with zero overlaps
- total_ticks: Iterations to solve (or max if unsolved)
- pressure_history: Pressure value at each tick (gaps + overlaps + util_var + unscheduled)
- band_escalation_events: Sampling band changes (tick, from_band, to_band)
- model_escalation_events: Model tier changes (tick, from_model, to_model)
- final_model: Which model tier solved the schedule
- token_usage: Prompt and completion tokens consumed per trial
### A.8 Replication Notes
Each configuration runs 30 independent trials with different random seeds to ensure reliability. Results report mean solve rates and tick counts across trials.
### A.9 Estimated Runtime
| Main Grid | 45 | 30 | 5 hours |
| --- | --- | --- | --- |
| Ablation | 5 | 30 | 1 hour |
| Scaling | 3 | 30 | 45 min |
| Difficulty | 15 | 30 | 2.5 hours |
| Total | | | $âŒ$ 9 hours |
Table 15: Estimated runtime for all experiments on NVIDIA RTX 4070 8GB GPU.
## Appendix B Pressure Alignment Verification
This appendix verifies that the meeting scheduling domain satisfies the alignment condition (Definition 2) required for convergence (Theorem 1).
### B.1 Per-Region Pressure Components
The per-region pressure function includes three components:
- gap_ratio: Fraction of empty slots in this time block. Strictly local âdepends only on meetings scheduled within this region.
- overlap_count: Number of attendee double-bookings within this time block. Strictly local âthe overlap sensor counts attendees with multiple meetings in this block only, not across blocks.
- utilization_variance: Variance in room utilization within this time block. Strictly local âmeasures balance across rooms for meetings in this region.
The unscheduled_count component is added to total pressure only, not to per-region pressure. This design choice ensures separability.
### B.2 Coupling Analysis
For the alignment condition, we need $|P_j(s^\prime)-P_j(s)|â€Î”$ for all $jâ i$ when modifying region $i$ .
Result: $Δ=0$ (per-region pressure is separable).
Rationale: All three per-region components depend only on the regionâs own content:
1. Modifying region $i$ âs schedule changes which meetings occupy region $i$ âs slots.
1. This affects region $i$ âs gap_ratio, overlap_count, and utilization_variance.
1. It has zero effect on any other region $j$ âs pressure, since $j$ âs components depend only on meetings within $j$ âs time slots.
Note on attendee constraints: While the same attendee may have meetings in multiple time blocks, our overlap sensor counts overlaps within each block independently. Moving a meeting from region $i$ to region $j$ may create or resolve overlaps in region $j$ , but this is a local effect on $j$ âs pressure, not a coupling effect where modifying region $i$ affects region $j$ through some indirect mechanism.
### B.3 Empirical Validation
Analysis of 270 pressure-field trials confirms separability:
- Total tick-to-tick transitions analyzed: 9,873
- Transitions with pressure improvement: 1,160 (11.7%)
- Transitions with pressure degradation: 0 (0.0%)
- Transitions with no change: 8,713 (88.3%)
The complete absence of pressure degradation is consistent with separable pressure: each accepted patch reduces local pressure, which directly reduces global pressure without adverse effects on other regions. (Patches that would increase local pressure are rejected by validation.)
Mean improvement magnitude: 2.67 pressure units. For context, initial problem pressure averages 12.7 units (range 0â42), so each successful patch reduces pressure by approximately 21% of the typical starting value. This substantially exceeds any plausible coupling bound, confirming that local improvements reliably translate to global improvements.
Verification of $ÎŽ_\min>0$ : Theorem 5.1 requires $ÎŽ_\min>(n-1)Δ$ . With $Δ=0$ (separable pressure), this reduces to $ÎŽ_\min>0$ . Empirically, the minimum observed pressure reduction across all 1,160 accepted patches is 1.0 pressure units, confirming $ÎŽ_\minâ„ 1>0$ . The theoremâs convergence guarantee therefore applies to this domain.
## Appendix C Theorem Proofs
This appendix provides complete proofs for the theoretical results in Section 4. Theorem 5.2 (Basin Quality) is omitted here as its proof is trivial and appears in full in the main text.
### C.1 Proof of Theorem 5.1 (Convergence)
* Proof*
We show that the system reaches a stable basin within the stated bound. Setup. Let $P^t=P(s^t)$ denote global pressure at tick $t$ . Under $Δ$ -bounded coupling, when a patch reduces local pressure $P_i$ by $Ύ_i$ , global pressure changes by:
$$
P^t+1-P^t=-ÎŽ_i+â_jâ i(P_j(s^t+1)-P_j(s^t))
$$
By bounded coupling, $|P_j(s^t+1)-P_j(s^t)|â€Î”$ for each $jâ i$ . With $n$ regions:
$$
P^t+1-P^tâ€-ÎŽ_i+(n-1)Δ
$$ Progress guarantee. If $ÎŽ_iâ„ÎŽ_\min$ and $ÎŽ_\min>(n-1)Δ$ , then:
$$
P^t+1-P^tâ€-ÎŽ_\min+(n-1)Δ<0
$$
Each tick with an applied patch reduces global pressure by at least $ÎŽ_\min-(n-1)Δ>0$ . Termination. Since $P(s)â„ 0$ for all states and pressure decreases by a fixed minimum $Î=ÎŽ_\min-(n-1)Δ$ per tick (when patches are applied), the system must reach a state where no region exceeds $Ï_act$ within:
$$
Tâ€\frac{P_0}{Î}=\frac{P_0}{ÎŽ_\min-(n-1)Δ}
$$
ticks. At this point, no region activates, and the system has reached a stable basin. Stability maintenance. The decay constraint $Î_f>1-e^-λ_f·Ï_inh$ ensures that fitness reinforcement from the final patches exceeds decay during the inhibition period. This prevents immediate re-activation after reaching stability. â
### C.2 Proof of Theorem 5.3 (Basin Separation)
* Proof*
We show that distinct stable basins are separated by pressure barriers of height at least $Ï_act$ . Basin definition. A stable basin $B$ is a connected region of state space where all states satisfy $P_i(s)<Ï_act$ for all regions $i$ . Within a basin, no agent takes action (pressure below activation threshold). Transition requirement. To move from basin $B_1$ to a distinct basin $B_2$ , the system must pass through states not in either basin. At the boundary of $B_1$ , at least one region $i$ must have $P_i(s)â„Ï_act$ (otherwise, by continuity under separable pressure, the state would still be in $B_1$ ). Barrier height. The minimum pressure exceedance required to exit a basin is $Ï_act$ . This defines the barrier height separating basins. Implication for decay. Without decay, once the system enters a basin, it remains there indefinitelyâfitness remains high, preventing re-activation. Decay erodes fitness over time, eventually allowing pressure to exceed $Ï_act$ and enabling transition to potentially lower-pressure basins. â
### C.3 Proof of Theorem 5.4 (Linear Scaling)
* Proof*
We analyze the per-tick computational complexity. Signal computation. For each of $m$ regions, computing $d$ -dimensional signals requires $O(d)$ operations (reading local content, computing quality metrics). Total: $O(m· d)$ . Pressure computation. For each region, computing pressure from $k$ signal axes requires $O(k)$ operations (weighted sum). Total: $O(m· k)$ . Patch proposal. Each of $a$ actors proposes patches for activated regions. In the worst case (all regions activated), this is $O(m· a)$ proposals. Each proposal involves an Large Language Model call, which dominates wall-clock time but is constant per call. Selection. Sorting $O(m· a)$ candidate patches by quality requires $O(m· a·\log(m· a))$ comparisons. Greedy selection of non-conflicting patches is $O(m· a)$ with appropriate data structures. Coordination overhead. Critically, agents share no messages. Each agent reads shared artifact state (read-only) and writes proposed patches to a central queue. No inter-agent communication occurs. The fork pool for parallel validation has fixed size $K$ , contributing $O(K)$ overhead independent of agent count. Total. Summing components: $O(m· d+m· k+m· a+m· a·\log(ma)+K)=O(m·(d+k+a·\log(ma)))$ , independent of agent count $n$ beyond its contribution to $a$ . â
### C.4 Proof of Theorem 5.5 (Parallel Convergence)
* Proof*
We extend Theorem 5.1 to the parallel setting. Setup. Suppose $K$ patches are validated in parallel per tick, targeting disjoint regions $i_1,âŠ,i_K$ . Combined pressure reduction. Under $Δ$ -bounded coupling, each patch $k$ targeting region $i_k$ reduces local pressure by at least $ÎŽ_\min$ . The global pressure change from patch $k$ is:
$$
Î P_kâ€-ÎŽ_\min+(n-1)Δ
$$ When patches target disjoint regions, their effects on global pressure are additive (no double-counting of coupling effects within the modified regions). The combined reduction is:
$$
P^t+1-P^tâ€â_k=1^KÎ P_k†K·(-ÎŽ_\min+(n-1)Δ)
$$ Convergence bound. With combined reduction $K·(Ύ_\min-(n-1)Δ)$ per tick:
$$
Tâ€\frac{P_0}{K·(ÎŽ_\min-(n-1)Δ)}
$$ Conflict handling. When multiple patches target the same region, only one is selected (highest quality). This reduces effective parallelism but maintains correctnessâthe selected patch still provides at least $ÎŽ_\min-(n-1)Δ$ reduction. â
## References