# AgentOrchestra: Orchestrating Multi-Agent Intelligence with the Tool-Environment-Agent(TEA) Protocol
**Authors**:
- Wentao Zhang
- Liang Zeng
- Yuzhen Xiao
- Yongcong Li
- Ce Cui
- Yilei Zhao
- Rui Hu
- Yang Liu
- Yahui Zhou
- Bo An (Skywork AI Nanyang Technological University)
Abstract
Recent advances in LLM-based agent systems have shown promise in tackling complex, long-horizon tasks. However, existing LLM-based agentprotocols (e.g., A2A and MCP) under-specify cross-entity lifecycle and context management, version tracking, and ad-hoc environment integration, which in turn encourages fixed, monolithic agent compositions and brittle glue code. To address these limitations, we introduce the Tool–Environment–Agent (TEA) protocol, a unified abstraction that models environments, agents, and tools as first-class resources with explicit lifecycles and versioned interfaces. TEA provides a principled foundation for end-to-end lifecycle and version management, and for associating each run with its context and outputs across components, improving traceability and reproducibility. Moreover, TEA enables continual self-evolution of agent-associated components Unless otherwise specified, agent-associated components include prompts, memory/tool/agent/environment code, and agent outputs (solutions). through a closed feedback loop, producing improved versions while supporting version selection and rollback. Building on TEA, we present AgentOrchestra, a hierarchical multi-agent framework in which a central planner orchestrates specialized sub-agents for web navigation, data analysis, and file operations, and supports continual adaptation by dynamically instantiating, retrieving, and refining tools online during execution. We evaluate AgentOrchestra on three challenging benchmarks, where it consistently outperforms strong baselines and achieves 89.04% on GAIA, establishing state-of-the-art performance to the best of our knowledge. Overall, our results provide evidence that TEA and hierarchical orchestration improve scalability and generality in multi-agent systems.
<details>
<summary>x2.png Details</summary>

### Visual Description
Icon/Small Image (32x33)
</details>
AgentOrchestra: Orchestrating Multi-Agent Intelligence with the Tool-Environment-Agent(TEA) Protocol
1 Introduction
Recent advances in LLM-based agent systems have enabled strong performance on both general-purpose and complex, long-horizon tasks across diverse domains, including web navigation (OpenAI, 2025b; Müller and Žunič, 2024), computer use (Anthropic, 2024a; Qin et al., 2025), code execution (Wang et al., 2024a), game playing (Wang et al., 2023; Tan et al., 2024), and research assistance (OpenAI, 2024; DeepMind, 2024; xAI, 2025). Despite this progress, cross-environment generalization remains limited because context is scattered across prompts and logs, environment integration relies on brittle glue code, and agent-associated components are typically fixed rather than feedback-driven self-evolution.
Additionally, current agent protocols fall short of serving as a general substrate for scalable, general-purpose agents. As summarized in Table 1, representative protocols such as Google’s A2A (Google, 2025) and Anthropic’s MCP (Anthropic, 2024b) provide important building blocks, including task-level collaboration and messaging in A2A, as well as tool and resource schemas, discovery, and invocation in MCP. However, three protocol-level gaps remain: i) Lifecycle and context management are fragmented, as neither standardizes unified primitives to manage lifecycles and maintain consistent, versioned execution context across agent-associated components; ii) Self-evolution is not supported at the protocol level, as both protocols largely treat prompts and resources as externally maintained assets, and do not define a closed loop to refine prompts or tools from execution feedback with traceable versioning; iii) Environments are not first-class, environments are delegated to application-specific runtimes instead of being managed components with clear boundaries and constraints. This makes it difficult to switch agents across environments, reuse environments, and isolate parallel runs, often reducing systems to glue-code orchestration.
Table 1: Comparison of TEA Protocol with A2A and MCP. Symbols: $\checkmark$ = Yes, $\triangle$ = Partial, $×$ = No.
| Dimension | TEA | A2A | MCP |
| --- | --- | --- | --- |
| Core Entities | Tool, Env, Agent | Agent, Tool | Model |
| Lifecycle & Version | $\checkmark$ | $×$ | $×$ |
| Entity Transformations | $\checkmark$ | $×$ | $×$ |
| Self-Evolution Support | $\checkmark$ | $×$ | $×$ |
| Open Ecosystem | ✓ | $\triangle$ | $\triangle$ |
To address these limitations, we propose the Tool–Environment–Agent (TEA) protocol, which treats environments, agents, and tools as explicitly managed components under a unified protocol layer. Concretely, TEA standardizes component identifiers and version semantics, and binds each run to its context and execution state, so that artifacts remain traceable across iterations. Importantly, TEA goes beyond MCP by standardizing cross entity lifecycle semantics, explicit version semantics with stable entity identifiers, run-indexed context capture, explicit environment boundaries with constraints, and closed loop evolution hooks driven by execution feedback. As a result, execution state, artifacts, and context can be consistently persisted, reused, and traced across runs and iterations. TEA further enables self-evolution by defining a closed loop in which execution feedback can trigger agent-associated components during runtime, with updates recorded as new versions. Finally, TEA models environments as first-class components with explicit boundaries and constraints, for example web sandboxes, file systems, and code execution runtimes, improving reuse and isolation across heterogeneous domains and reducing context leakage in parallel executions. This also encourages consolidating functionally related tools into coherent environments; for example, discrete file operations can be organized as a managed file system, reducing context fragmentation and management overhead. Overall, TEA aims to make agent construction more composable and reproducible in practice. Detailed motivations for the TEA protocol and in-depth comparisons with existing protocols are provided in Appendix A, B.
Based on the TEA protocol, we develop AgentOrchestra, a hierarchical multi-agent framework for general-purpose task solving that integrates high-level planning with modular collaboration. AgentOrchestra uses a central planner to decompose a user objective and delegate sub-tasks to specialized agents for research, web navigation, analysis, tool synthesis, and reporting. Compared to flat coordination, where an orchestrator selects from a growing global pool of agents and tools and tends to accumulate irrelevant context, AgentOrchestra adopts hierarchical delegation with localized tool ownership. The planner routes each sub-task to a domain-specific sub-agent (or environment), which maintains and exposes only a curated toolset and context for its domain. This structure converts global coordination into a sequence of localized routing decisions, enabling tree-structured expansion as new capabilities are added while keeping the orchestrator’s decision scope and context footprint bounded. For example, the planner first selects a domain-level agent, which then supplies only the tools and context required for that domain. Furthermore, AgentOrchestra incorporates a self-evolution module that leverages TEA’s lifecycle and versioning mechanisms to refine agent- associated components based on execution feedback. Our contributions are threefold:
- We introduce the TEA protocol, which unifies environments, agents, and tools as first-class, versioned components with lifecycles to support context management and execution.
- We develop AgentOrchestra, a hierarchical multi-agent system built on TEA, demonstrating scalable orchestration through tree-structured routing and feedback-driven self-evolution.
- We conduct extensive evaluations on three challenging benchmarks, including ablations to isolate the effects of key components. AgentOrchestra consistently outperforms strong baselines and achieves 89.04% on GAIA, establishing state-of-the-art performance to the best of our knowledge.
2 Related Work
2.1 Tool and Agent Protocols
Recent protocols standardize tool interfaces and agent communication. For instance, MCP (Anthropic, 2024b) unifies tool integration for LLMs, while A2A (Google, 2025) enables agent-to-agent messaging and coordination. Other efforts, such as the Agent Network Protocol (ANP) (Ehtesham et al., 2025) and frameworks like SAFEFLOW (Li et al., 2025), enhance interoperability and safety in multi-agent systems. While these protocols provide essential building blocks, they primarily treat agents and tools as isolated service endpoints, often overlooking environments as dynamic, first-class components. TEA extends these existing standards rather than replacing them. By integrating tools, environments, and agents into a unified context-aware framework, TEA resolves protocol fragmentation with integrated lifecycle and version management missing in MCP or A2A.
2.2 General-Purpose Agents
Integrating tools with LLMs represents a paradigm shift, enabling agents to exhibit enhanced flexibility, cross-domain reasoning, and natural language interaction (Liang and Tong, 2025). Such systems have demonstrated efficacy across diverse domains, including web browsing (OpenAI, 2025b; Müller and Žunič, 2024), computer operation (Anthropic, 2024a; Qin et al., 2025), code execution (Wang et al., 2024a), and game playing (Wang et al., 2023; Tan et al., 2024). Standardized interfaces like OpenAI’s Function Calling and Anthropic’s MCP (OpenAI, 2023; Anthropic, 2024b), alongside frameworks such as ToolMaker (Wölflein et al., 2025), have further streamlined the synthesis of LLM-compatible tools. Building upon these foundations, multi-agent architectures like MetaGPT (Hong et al., 2023) demonstrate the potential of specialized agent coordination for complex problem-solving. However, many current approaches still struggle with efficient communication, dynamic role allocation, and scalable teamwork. The emergence of generalist frameworks, including Manus (Shen and Yang, 2025), OpenHands (Wang et al., 2024b), and smolagents (Roucher et al., 2025), has advanced unified perception and tool-augmented action. While recent efforts like Alita (Qiu et al., 2025) explore minimal predefinition and maximal self-evolution, these systems often lack unified protocols for cross-layer resource management. This gap motivates our proposal of the TEA Protocol and AgentOrchestra.
3 The TEA Protocol
The TEA Protocol is fundamentally designed around coroutine-based asynchronous execution, enabling concurrent task processing and parallel multi-agent coordination. As illustrated in Figure 1, the protocol architecture comprises three primary layers: i) Basic Managers provide foundational services through six specialized components (model, prompt, memory, dynamic, version, and tracer); ii) Core Protocols define the Tool Context Protocol (TCP), Environment Context Protocol (ECP), and Agent Context Protocol (ACP), each implemented through a context manager for lifecycle engineering and a server for standardized orchestration; and iii) Protocol Transformations establish bidirectional conversion pathways (e.g., A2T, E2T, A2E) enabling dynamic role reconfiguration. Additionally, the protocol incorporates a Self-Evolution Module that wraps agent-associated components as evolvable variables for iterative optimization. Details and formalization can be found in Appendix C.
<details>
<summary>x3.png Details</summary>

### Visual Description
## Diagram: Tool-Environment-Agent (TEA) Architecture
### Overview
The image is a diagram illustrating the architecture of a Tool-Environment-Agent (TEA) system. It depicts the relationships and interactions between different components, including Researchers, Analyzers, Browsers, Managers, Agents, Tools, and Environments. The diagram uses arrows to indicate the flow of information and protocols for communication.
### Components/Axes
* **Header (Top):**
* Roles: Researcher, Analyzer, Browser, Manager (represented by icons of people performing tasks)
* Protocol: Agent Context Protocol (ACP) - a blue box below the roles.
* **Main Diagram (Center):**
* Central Node: Tool-Environment-Agent (TEA) - represented by a steaming cup icon.
* Nodes surrounding TEA:
* Agents (top)
* Tools (left)
* Envs (right)
* Communication Arrows:
* A2T: Agents to TEA (upward arrow)
* T2A: TEA to Agents (downward arrow)
* A2E: Agents to Envs (upward arrow)
* E2A: Envs to Agents (downward arrow)
* E2T: Envs to TEA (leftward arrow)
* T2E: TEA to Envs (rightward arrow)
* Protocols:
* Tool Context Protocol (TCP) - below the "Tools" node.
* Environment Context Protocol (ECP) - below the "Envs" node.
* **Footer (Bottom):**
* Tools: Bash, Python, Mdify, Web (represented by icons)
* Envs: Browser, Computer, Mobile, Files, Github (represented by icons)
### Detailed Analysis or Content Details
* **Roles:** The diagram starts with user roles at the top: Researcher, Analyzer, Browser, and Manager. These roles likely represent different types of users interacting with the system.
* **Agent Context Protocol (ACP):** This protocol governs the interaction between the user roles and the Agents.
* **Agents:** Agents are central to the system, communicating with both the TEA and the Envs.
* **Tool-Environment-Agent (TEA):** This is the core component, facilitating interaction between Tools and Environments. It is represented by a steaming cup icon.
* **Tools:** Represented by Bash, Python, Mdify, and Web, these are the tools that the system utilizes.
* **Envs:** Represented by Browser, Computer, Mobile, Files, and Github, these are the environments in which the system operates.
* **Tool Context Protocol (TCP):** This protocol governs the interaction between the TEA and the Tools.
* **Environment Context Protocol (ECP):** This protocol governs the interaction between the TEA and the Envs.
* **Communication Arrows:** The arrows indicate the flow of information between the components. For example, A2T represents communication from Agents to TEA.
### Key Observations
* The TEA acts as a central hub, mediating interactions between Tools, Environments, and Agents.
* Protocols (ACP, TCP, ECP) are used to standardize communication between different components.
* The diagram provides a high-level overview of the system architecture, highlighting the key components and their relationships.
### Interpretation
The diagram illustrates a modular architecture where the Tool-Environment-Agent (TEA) acts as an intermediary, enabling communication and interaction between various tools, environments, and agents. The use of context protocols (ACP, TCP, ECP) suggests a standardized approach to managing interactions, ensuring that each component understands the context of the communication. This architecture likely aims to provide a flexible and extensible system that can adapt to different tools, environments, and user roles. The presence of Github as an environment suggests version control and collaboration are important aspects of the system. The steaming cup icon for TEA could symbolize the "brewing" or processing of information between the tools and environments.
</details>
Figure 1: Architecture of the TEA Protocol.
3.1 Basic Managers
The Basic Managers constitute the foundation of the TEA Protocol, providing essential services through six specialized managers: i) the model manager abstracts heterogeneous LLM backends through a unified interface; ii) the prompt manager handles prompt lifecycle and versioning; iii) the memory manager coordinates persistence via session-based concurrency control; iv) the dynamic manager enables runtime code execution and serialization; v) the version manager maintains evolution histories for all components; and vi) the tracer records comprehensive execution trajectories and system-wide telemetry, serving as a data collection engine for audit, debugging, and the synthesis of high-quality datasets for agent training.
3.2 Core Protocols
The TEA Protocol defines three core context protocols: the Tool Context Protocol (TCP), the Environment Context Protocol (ECP), and the Agent Context Protocol (ACP). These protocols share a unified architectural design, each implemented through two core components: a context manager for context engineering, lifecycle management, and semantic retrieval, and a server that exposes standardized interfaces to other system modules. Each protocol generates a unified contract document (analogous to Agent Skills (Anthropic, 2025)) that aggregates all registered components’ descriptions to facilitate resource discovery and usage.
Tool Context Protocol. TCP fundamentally extends MCP (Anthropic, 2024b) by introducing integrated context engineering and comprehensive lifecycle management. Implemented through a ToolContextManager and a TCPServer, it supports seamless tool loading from both local registries and persistent configurations. During registration, TCP automatically synthesizes multiple representation formats, including function-calling schemas for LLM interfaces, natural language descriptions for documentation, and type-safe argument schemas for validation, providing LLMs with rich semantic information for accurate parameter inference. Furthermore, TCP incorporates a robust versioning system and a semantic retrieval mechanism based on vector embeddings, ensuring that tools can evolve over time while remaining easily discoverable through similarity-based queries.
Environment Context Protocol. ECP addresses the lack of unified interfaces in current agent systems by formalizing computational environments as first-class components with distinct observation and action spaces. Following an architectural pattern similar to TCP, it employs an EnvironmentContextManager to maintain state coherence and manage the contextual execution environments required by tools. ECP automatically discovers and registers environment-specific actions, converting them into standardized interfaces that agents can invoke via action toolkits. This design enables agents to operate across heterogeneous domains, such as browsers or file systems, without bespoke adaptations, while leveraging versioning and semantic retrieval to manage environment-level capabilities.
Agent Context Protocol. ACP establishes a unified framework for the registration, representation, and orchestration of autonomous agents, overcoming the poor interoperability and fragmented attribute definitions in existing multi-agent systems. It utilizes an AgentContextManager to maintain agent states and execution contexts, providing a foundation for persistent coordination across tasks and sessions. ACP captures semantically enriched metadata regarding agents’ roles, competencies, and objectives, and formalizes the modeling of complex inter-agent dynamics, including cooperative, competitive, and hierarchical configurations. By embedding structured contextual descriptions and maintaining relationship representations, ACP facilitates adaptive collaboration and systematic integration within the broader TEA ecosystem.
3.3 Protocol Transformations
While TCP, ECP, and ACP provide independent specifications for tools, environments, and agents, practical deployment requires seamless interoperability across these protocols. Well-defined transformation pathways are essential for enabling computational components to assume alternative roles and exchange contextual information in a principled manner. These transformations constitute the foundation for dynamic role reconfiguration, allowing components to flexibly adapt their functional scope in response to evolving task requirements and system constraints. We identify six fundamental categories of protocol transformations:
- Agent-to-Tool (A2T). Encapsulates an agent’s capabilities and reasoning into a standardized tool interface while preserving awareness. For example, a deep researcher workflow can be packaged as a general-purpose search tool.
- Tool-to-Agent (T2A). Treats tools as operational actuators by mapping an agent’s goals into parameterized tool invocations, aligning reasoning with tool constraints. For example, a data analysis agent may invoke SQL tools to query structured databases.
- Environment-to-Tool (E2T). Converts actions of environments into standardized tool interfaces, enabling agents to interact with environments through consistent tool calls. For example, browser actions such as Navigate and Click can be consolidated into a context-aware toolkit.
- Tool-to-Environment (T2E). Elevates a collection of tools into an environment abstraction where functions become actions within a coherent action space governed by shared state. For example, a development toolkit can be encapsulated as a programming environment for sequential code-edit-compile-debug workflows.
- Agent-to-Environment (A2E). Encapsulates an agent as an interactive environment by exposing its decision rules and state dynamics as an operational context for other agents. For example, a market agent can be represented as an environment that provides trading rules and dynamic responses for training.
- Environment-to-Agent (E2A). Embeds reasoning and adaptive decision-making into an environment’s dynamics, transforming it into an autonomous agent that can initiate behaviors and enforce constraints. For example, a game environment can be elevated into an opponent agent that adapts its strategy to the player’s actions.
3.4 Self-Evolution Module
The Self-Evolution Module enables agents to continuously improve performance by optimizing system components during task execution. It wraps evolvable components, including prompts, tool/agent/environment/memory code, and successful execution solutions, as variables for iterative optimization. The module employs two primary methods: textgrad (Yuksekgonul et al., 2025) for gradient-based refinement and self-reflection for strategic analysis. Optimized components are automatically registered as new versions via the version manager, ensuring that subsequent tasks leverage improved capabilities while maintaining access to historical records for analysis and rollback.
<details>
<summary>x4.png Details</summary>

### Visual Description
## System Architecture Diagram: Planning Agent and Tool Interaction
### Overview
The image presents a system architecture diagram detailing the interaction between a Planning Agent, various specialized agents, tools, and the environment. It illustrates the flow of information and control within the system, emphasizing the role of the Planning Agent in coordinating tasks and managing resources. The diagram includes components such as user objectives, planning tools, specialized agents (Researcher, Browser, Analyzer, Generator), context protocols (TCP, ACP, ECP), and basic managers.
### Components/Axes
**1. Top Section: Planning Agent**
* **User Objectives:** Located on the left, indicating the starting point of the process.
* **Planning Agent:** The central component, responsible for managing and coordinating tasks.
* **Tools:**
* Actions: create, update, delete, mark.
* Planning: Interpret user tasks, Decompose into manageable sub-tasks, Assign to specialized sub-agents.
* Planning Tool: Create, update, and manage plans for complex tasks simultaneously; Track execution states.
* **Objective Shifts (Update Plans) & Unexpected Errors:** Represent feedback loops and potential disruptions.
* **Planner:** Top-right, connected to Task.
* **Researcher, Browser, Analyzer, Generator, Reporter:** Specialized agents branching from the Planner.
* **Answer:** Bottom-right, the final output.
**2. Middle Section: Specialized Agents**
* **Deep Researcher Agent:** Optimizes queries, searches tools, refines insight.
* **Browser Use Agent:** Decides actions, browses actions, records results.
* **Deep Analyzer Agent:** Organizes diverse formats, reasons and summarizes.
* **Tool Generator Agent (x2):** Tool retrieval, creation, reuse; Add content, export report.
**3. Central Section: Agent Context Protocol (ACP)**
* **Tool Context Protocol (TCP):** Connects agents to tools.
* **Environment Context Protocol (ECP):** Connects agents to the environment.
* **Tool-Environment-Agent (TEA):** Central hub connecting Tools, Environment (Envs), and Agents.
* **Arrows:** Indicate the flow of information between Agents (A), Tools (T), and Environment (E). Labeled as A2T, T2A, A2E, E2A, E2T, T2E.
**4. Bottom Section: Tools and Managers**
* **General Tools:** Bash, Python, Mdify, Web, Todo.
* **MCP Tools:** Searcher, Analyzer (Agent Tools), Local, Remote.
* **Environment Tools:** Browser, Github, Computer.
* **Basic Managers:** Model Manager, Memory Manager, Prompt Manager, Dynamic Manager, Version Manager, Tracer.
**5. Environment Context Protocol (ECP) Details:**
* **Rules:** Name, description, state, interaction.
* **Actions:** Read, write, goto, kexpres, clone, create.
* **Examples:** File System, Browser, Github&Git, Computer.
### Detailed Analysis or ### Content Details
**1. Planning Agent Flow:**
* The process starts with User Objectives, which are fed into the Planning Agent.
* The Planning Agent uses Tools to create, update, and manage plans.
* The Planning Agent decomposes tasks and assigns them to specialized sub-agents.
* Feedback loops and error handling are incorporated through Objective Shifts and Unexpected Errors.
* The Planner delegates tasks to Researcher, Browser, Analyzer, and Generator agents.
* The Reporter agent synthesizes information to produce an Answer.
**2. Specialized Agent Functions:**
* **Deep Researcher Agent:** Focuses on information retrieval and refinement.
* **Browser Use Agent:** Interacts with web browsers to gather information.
* **Deep Analyzer Agent:** Processes and summarizes diverse data formats.
* **Tool Generator Agent:** Manages the creation and reuse of tools.
**3. Context Protocols:**
* **Tool Context Protocol (TCP):** Facilitates communication between agents and tools.
* **Agent Context Protocol (ACP):** Serves as a central communication hub for agents, tools, and the environment.
* **Environment Context Protocol (ECP):** Manages interactions between agents and the environment.
**4. Tool-Environment-Agent (TEA) Interaction:**
* The TEA component acts as a central point for communication between agents, tools, and the environment.
* Arrows indicate the direction of information flow:
* A2T: Agent to Tool
* T2A: Tool to Agent
* A2E: Agent to Environment
* E2A: Environment to Agent
* E2T: Environment to Tool
* T2E: Tool to Environment
**5. Environment Context Protocol (ECP) Details:**
* **Rules:** Define the structure and behavior of interactions within the environment.
* **Actions:** Represent specific operations that can be performed within the environment.
* **Examples:**
* **Browser:** Actions include goto, input, click, scroll, kexpres, type.
* **File System:** Actions include read, write, move, copy, clone, commit, create, push.
* **Computer:** Actions include click, scroll, kexpres, type.
### Key Observations
* The diagram emphasizes a modular and hierarchical structure, with the Planning Agent at the top level and specialized agents performing specific tasks.
* Context protocols (TCP, ACP, ECP) play a crucial role in managing communication and interactions between different components.
* The Tool-Environment-Agent (TEA) component serves as a central hub for coordinating activities.
* The diagram incorporates feedback loops and error handling mechanisms to ensure robustness.
### Interpretation
The diagram illustrates a sophisticated system architecture designed for complex task management and problem-solving. The Planning Agent acts as a central coordinator, delegating tasks to specialized agents and managing resources. The use of context protocols (TCP, ACP, ECP) ensures seamless communication and interaction between different components. The Tool-Environment-Agent (TEA) component facilitates the integration of tools and the environment into the overall system.
The architecture is designed to be flexible and adaptable, with feedback loops and error handling mechanisms to address unexpected events. The modular structure allows for easy expansion and modification of the system. The inclusion of basic managers (Model Manager, Memory Manager, Prompt Manager, Dynamic Manager, Version Manager, Tracer) suggests a focus on efficient resource management and performance optimization.
Overall, the diagram presents a comprehensive and well-designed system architecture for intelligent task management and problem-solving.
</details>
Figure 2: Architecture of AgentOrchestra implemented based on TEA protocol.
4 AgentOrchestra
AgentOrchestra is a concrete instantiation of the TEA Protocol, designed as a hierarchical multi-agent framework that integrates high-level planning with modular agent collaboration. As illustrated in Figure 2, AgentOrchestra features a central planning agent that decomposes complex objectives and delegates sub-tasks to a team of specialized sub-agents. This section outlines our agent design principles and the architecture of both planning and specialized sub-agents. Details can be found in Appendix D.
4.1 Agent Design Principles
Within the TEA Protocol framework, agents are autonomous components that follow a structured interaction model with six core components. i) Agent: Managed via the ACP for registration and coordination. ii) Environment: External context and resources managed by the ECP, exposing unified interfaces for observation and action. iii) Model: LLM reasoning engines abstracted by the Basic Managers for model-agnostic interoperability and dynamic switching. iv) Memory: Session-based persistence that records trajectories and extracts reusable insights. v) Observation: The current context, including tasks, environment states, execution history, and available resources (tools and sub-agents). vi) Action: TCP-managed, executed via parameterized tool calls, where one tool may support multiple actions.
This architectural design facilitates a continuous perception–interpretation–action cycle. The agent first perceives the current observation and retrieves relevant context from memory. It then interprets this information through the unified model interface to determine the optimal action. The action is executed within the managed environment, and the resulting state transitions and insights are recorded back into memory to refine subsequent reasoning cycles. This iterative loop continues until the task objectives are satisfied or a termination condition is reached. Further details are provided in Appendix D.1.
4.2 Planning Agent
The planning agent is the central orchestrator of AgentOrchestra. It interprets the user goal, decomposes it into sub-tasks, and dispatches them to specialized sub-agents or TCP tools via ACP-mediated communication while tracking global progress and consolidating intermediate feedback. To enable principled orchestration, it leverages long-term memory to guide resource selection and dynamically constructs a unified invocation interface, including resources produced through E2T and A2T transformations. Execution follows an iterative loop of interpretation, allocation, and action, with automatic replanning under environment shifts or execution failures. Session management and tracer-based logging provide auditability and support robust long-horizon task completion.
4.3 Specialized Sub-Agents
To address diverse real-world challenges, AgentOrchestra instantiates specialized sub-agents tailored for task domains. These sub-agents are managed via the ACP and coordinate through the planning agent to execute complex workflows: i) Deep Researcher Agent: Specialized for comprehensive information gathering through multi-round research workflows. It performs parallel breadth-first searches across multiple engines and recursively issues follow-up queries until task objectives are satisfied, producing relevance-ranked, source-cited summaries. ii) Browser Use Agent: Provides automated, fine-grained web interaction by integrating both browser and computer environments under the ECP. It supports DOM-level and pixel-level operations (e.g., mouse movements), achieving unified control over interactive elements. iii) Deep Analyzer Agent: A workflow-oriented module designed for multi-step reasoning on heterogeneous multimodal data (e.g., text, PDFs, images, audio, video or zip). It applies type-specific analysis strategies and iterative refinement to synthesize insights into coherent conclusions. iv) Tool Generator Agent: Facilitates intelligent tool evolution through the automated creation, retrieval, and systematic reuse of TCP-compliant tools. It employs semantic search to identify tools and initiates a code synthesis process to develop new capabilities when gaps are identified. v) Reporter Agent: It aggregates and harmonizes evidence collected by upstream agents (e.g., the Deep Researcher Agent, Browser Use Agent, and Deep Analyzer Agent), then composes structured markdown with automatically deduplicated references and normalized URLs for consistent source attribution.
5 Empirical Studies
This section presents our experimental setup and results, including benchmark evaluations, baseline comparisons, and comprehensive analysis. Additional examples are provided in the Appendix F.
Experimental Settings. We evaluate our framework on three benchmarks: SimpleQA Wei et al. (2024), a 4,326-question factual accuracy benchmark; GAIA Mialon et al. (2023), assessing real-world reasoning, multimodal processing, and tool use with 301 test and 165 validation questions; and Humanity’s Last Exam (HLE) Phan et al. (2025), a 2,500-question multimodal benchmark for human-level reasoning and general intelligence. We report score (pass@1), which measures the proportion of questions for which the top prediction is fully correct. Specifically, the planning agent ( $m{=}50$ ), deep researcher ( $m{=}3$ ), tool generator ( $m{=}10$ ), deep analyzer ( $m{=}3$ ), and reporter are all built on gemini-3-flash-preview; the browser use agent employs gpt-4.1 ( $m{=}5$ ) and computer-use-preview(4o) ( $m{=}50$ ), where $m$ denotes the maximum steps.
5.1 Performance across Benchmarks
<details>
<summary>x5.png Details</summary>

### Visual Description
## Bar Chart: GAIA Test Results for Different Agents
### Overview
The image is a bar chart comparing the performance of different agents on the GAIA test across three levels (Level1, Level2, Level3) and their average scores. The chart displays scores ranging from 40 to 100, with each agent's performance represented by bars of different colors corresponding to the test level or average.
### Components/Axes
* **Y-axis:** "Score", ranging from 40 to 100 in increments of 10.
* **X-axis:** Categorical axis representing different agents: AgentOrchestrator, ToolOrchestra, HALO, AWorld, Su-Zero-Ultra, h2oGPTe-Agent, DeSearch, Alita, Langfun, o3-DR, JoyAgent, o4-mini-DR. These agents are grouped into four sections.
* **Legend (Top-Right):**
* Level1: Green
* Level2: Blue
* Level3: Purple
* Average: Orange
### Detailed Analysis
The chart is divided into four sections, each containing the same set of agents. Each section represents a different test level or the average score.
**Section 1: Level 1 (Green)**
* AgentOrchestrator: 98.9
* ToolOrchestra: 95.7
* HALO: 94.6
* AWorld: 95.7
* Su-Zero-Ultra: 93.5
* h2oGPTe-Agent: 89.2
* DeSearch: 91.4
* Alita: 92.5
* Langfun: 85.0
* o3-DR: 79.4
* JoyAgent: 77.4
* o4-mini-DR: 67.8
**Section 2: Level 2 (Blue)**
* AgentOrchestrator: 83.3
* ToolOrchestra: 82.4
* HALO: 84.9
* AWorld: 81.3
* Su-Zero-Ultra: 79.9
* h2oGPTe-Agent: 75.8
* DeSearch: 73.3
* Alita: 73.6
* Langfun: 68.6
* o3-DR: 67.3
* JoyAgent: 59.3
* o4-mini-DR: (Value not clearly visible, but appears to be around 50)
**Section 3: Level 3 (Purple)**
* AgentOrchestrator: 81.6
* ToolOrchestra: 97.8
* HALO: 69.4
* AWorld: 57.1
* Su-Zero-Ultra: 65.3
* h2oGPTe-Agent: 61.2
* DeSearch: 61.2
* Alita: 48.0
* Langfun: 47.5
* o3-DR: 44.3
* JoyAgent: (Value not clearly visible, but appears to be around 40)
* o4-mini-DR: (Value not clearly visible, but appears to be around 40)
**Section 4: Average (Orange)**
* AgentOrchestrator: 79.1
* ToolOrchestra: 87.4
* HALO: 85.4
* AWorld: 81.7
* Su-Zero-Ultra: 80.4
* h2oGPTe-Agent: 78.7
* DeSearch: 78.1
* Alita: 75.4
* Langfun: 73.1
* o3-DR: 68.7
* JoyAgent: 67.1
* o4-mini-DR: 58.3
### Key Observations
* **AgentOrchestrator:** Performs best on Level 1, with a score of 98.9, and worst on Level 3, with a score of 81.6.
* **ToolOrchestra:** Shows the highest score on Level 3 (97.8) and a relatively high average score (87.4).
* **HALO:** Scores are relatively consistent across Level 1 and Level 2, but drops significantly on Level 3.
* **o4-mini-DR:** Consistently scores the lowest across all levels and the average.
* **General Trend:** Performance tends to decrease from Level 1 to Level 3 for most agents.
### Interpretation
The bar chart provides a comparative analysis of different agents' performance on the GAIA test across three difficulty levels. The data suggests that the agents generally perform best on Level 1 and worst on Level 3, indicating that the difficulty increases as the level increases. ToolOrchestra stands out as having a high score on Level 3, suggesting it may be particularly well-suited for the challenges presented at that level. The consistent low performance of o4-mini-DR across all levels suggests it may need further development or is not well-suited for the GAIA test. The chart highlights the strengths and weaknesses of each agent, providing valuable insights for further development and optimization.
</details>
Figure 3: GAIA Test Results.
GAIA. AgentOrchestra achieves state-of-the-art performance (89.04% avg.) by mitigating the dimensionality curse and semantic drift that arise in large-scale agentic planning. We attribute this success to two architectural properties enabled by the TEA Protocol. First, hierarchical decoupling of the action space reduces planning complexity: while methods (e.g., ToolOrchestra, AWorld) must map goals to a monolithic toolkit, our hierarchical routing decomposes the global task into locally tractable sub-problems, lowering cognitive entropy for the central orchestrator and preserving abstract reasoning under long horizons, even amid low-level sensorimotor noise (e.g., granular DOM events). Second, ECP formalizes epistemic environment boundaries: GAIA’s multi-domain tasks require temporal and cross-modal state coherence, and baselines often degrade during domain transitions, such as from browser retrieval to local python analysis. By treating environments as first-class managed components, TEA preserves and propagates session-critical state (e.g., authentication tokens and transient file-system mutations) across agent boundaries, reducing contextual forgetting and enabling compositional generalization on challenging Level 2 and Level 3 scenarios. Third, AgentOrchestra supports recursive refinement of reasoning trajectories. When faced with complex problems, the Planning Agent evaluates intermediate insights and, when necessary, invokes the Tool Generator Agent to synthesize context-specific functionalities on the fly. This on-demand tool evolution bypasses the fixed-capability bottleneck of static agent components.
Table 2: Performance on GAIA Validation.
| Agents | Level 1 | Level 2 | Level 3 | Average |
| --- | --- | --- | --- | --- |
| HF ODR (o1) (HuggingFace, 2024) | 67.92 | 53.49 | 34.62 | 55.15 |
| OpenAI DR (OpenAI, 2024) | 74.29 | 69.06 | 47.60 | 67.36 |
| Manus (Shen and Yang, 2025) | 86.50 | 70.10 | 57.69 | 73.90 |
| Langfun (Google, 2024) | 86.79 | 76.74 | 57.69 | 76.97 |
| AWorld (Yu et al., 2025) | 88.68 | 77.91 | 53.85 | 77.58 |
| AgentOrchestra | 92.45 | 83.72 | 57.69 | 82.42 |
Table 3: Performance on SimpleQA and HLE.
| Model and Agent | SimpleQA |
| --- | --- |
| Models | |
| o3 (w/o tools) | 49.4 |
| gemini-2.5-pro-preview-05-06 | 50.8 |
| Agents | |
| Perplexity DR (Perplexity, 2025) | 93.9 |
| AgentOrchestra | 95.3 |
| Model and Agent | HLE |
| Models | |
| o3 (w/o tools) | 20.3 |
| claude-3.7-sonnet (w/o tools) | 8.9 |
| gemini-2.5-pro-preview-05-06 | 17.8 |
| Agents | |
| OpenAI DR (OpenAI, 2024) | 26.6 |
| Perplexity DR (Perplexity, 2025) | 21.1 |
| AgentOrchestra | 37.46 |
SimpleQA. AgentOrchestra achieves SOTA performance (95.3% accuracy), significantly surpassing both monolithic LLMs (e.g., o3 at 49.4%) and specialized retrieval agents like Perplexity Deep Research (93.9%). We attribute this improvement to systematic reduction of epistemic uncertainty through our hierarchical verification pipeline. SimpleQA primarily targets short-form factuality, where hallucinations often arise from the model’s inability to reconcile conflicting web-based evidence or its tendency to rely on internal parametric memory. AgentOrchestra mitigates these issues by enforcing cross-agent consensus: the Planning Agent orchestrates a retrieve-verify-synthesize cycle where the Deep Researcher performs multi-engine breadth-first searches while the Deep Analyzer evaluates evidence consistency across heterogeneous sources. By decoupling retrieval from analysis, the system prevents "confirmation bias" inherent in single-agent architectures, where the same model both proposes and validates a hypothesis. Furthermore, the integration with the Reporter Agent ensures traceable attribution, grounding every factual claim in a re-verified source, which effectively transforms the task from an open-domain generation problem into a structured evidence-synthesis process.
HLE. AgentOrchestra achieves 37.46% on the HLE benchmark, a substantial margin over leading baselines like o3 (20.3%) and Perplexity Deep Research (21.1%). This gain highlights the framework’s capacity for long-horizon analytical reasoning and adaptive capability expansion in expert-level domains. HLE demands more than simple retrieval; it requires synthesizing disparate, highly specialized knowledge. In this setting, the hierarchical structure enables strategic pruning of the hypothesis space, allowing the Planning Agent to maintain global objective coherence while delegating technical validation to specialized agents such as the Deep Analyzer. As a result, the final solution is both analytically rigorous and cross-verified against multimodal evidence, yielding robust performance on challenging expert-level tasks.
5.2 Ablation Studies
Table 4: Sub-agent effectiveness across GAIA Test.
| P | R | B | A | T | Level 1 | Level 2 | Level 3 | Average | Improvement |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
| ✓ | | | | | 54.84 | 33.96 | 10.20 | 36.54 | – |
| ✓ | ✓ | | | | 86.02 | 47.17 | 34.69 | 57.14 | +56.40% |
| ✓ | ✓ | ✓ | | | 89.25 | 71.07 | 46.94 | 72.76 | +27.33% |
| ✓ | ✓ | ✓ | ✓ | | 91.40 | 77.36 | 61.22 | 79.07 | +8.67% |
| ✓ | ✓ | ✓ | ✓ | ✓ | 98.92 | 85.53 | 81.63 | 89.04 | +12.61% |
Effectiveness of the specialized sub-agents. Ablation studies on the GAIA Test demonstrate the synergistic effect inherent in our multi-agent coordination. Integrating coarse-grained exploratory retrieval (Researcher) with fine-grained operational interaction (Browser) nearly doubles performance (36.54% to 72.76%), proving that breadth of information and depth of interaction are mutually reinforcing. The Deep Analyzer’s 8% gain highlights the necessity of specialized reasoning pipelines for high-entropy multimodal tasks, while the Tool Generator’s 12.61% boost validates the efficacy of on-demand capability synthesis in overcoming the limitations of static, predefined toolsets. These results suggest that complex problem-solving emerges not just from individual agent strength, but from the structured delegation of specialized roles.
Efficiency analysis. AgentOrchestra ’s operational efficiency is evaluated across varying task complexities. Simple tasks typically complete within 30 seconds using approximately 5k tokens, while medium-complexity tasks average 3 minutes (25k tokens). Complex multimodal or long-horizon scenarios require approximately 10 minutes and 100k tokens. Compared to monolithic baselines, our hierarchical architecture optimizes resource allocation, maintaining operational costs comparable to commercial research agents while delivering significant performance gains.
Effectiveness of the self-evolution module. The TEA Protocol enables self-optimization by treating system components as evolvable variables, helping bridge the gap between base model capacity and task requirements. Evaluations on GPQA-Diamond and AIME benchmarks show that iterative refinement, including gradient-based (TextGrad) and symbolic (self-reflection) approaches, mitigates reasoning bottlenecks in foundation models. The improvement is exemplified by a 13.34% gain on AIME25 for gpt-4.1 under self-reflection, highlighting recursive trajectory refinement. Leveraging execution feedback via TEA’s versioning and tracer mechanisms, the system identifies and corrects logical inconsistencies in its planning. Overall, this shifts reasoning from one-shot inference to a managed optimization process, enabling AgentOrchestra to evolve problem-solving strategies for frontier-level tasks.
Table 5: Effectiveness of the self-evolution module. Direct means using the base model directly.
| Strategy | GPQA-Diamond | AIME24 | AIME25 |
| --- | --- | --- | --- |
| Base Model: gpt-4o | | | |
| Direct | 47.98% | 13.34% | 6.67% |
| w/ TextGrad | 54.04% | 10.00% | 10.00% |
| w/ Self-reflection | 55.05% | 20.00% | 6.67% |
| Base Model: gpt-4.1 | | | |
| Direct | 61.11% | 23.34% | 20.00% |
| w/ TextGrad | 65.15% | 26.67% | 23.34% |
| w/ Self-reflection | 68.18% | 33.34% | 33.34% |
Regarding tool evolution, the tool generator agent demonstrates efficient creation and reuse capabilities within the TCP framework. During our evaluation, the agent autonomously generated over 50 specialized tools, achieving a 30% reuse rate across subsequent tasks. This indicates an effective balance between tool specialization and generalization, ensuring that the system’s capabilities expand adaptively while maintaining resource efficiency.
6 Conclusion
We introduced the TEA Protocol, unifying environments, agents, and tools to address fragmentation in existing standards. Building on this, we presented AgentOrchestra, a hierarchical multi-agent framework with specialized sub-agents for planning, research, web interaction, and multimodal analysis. Evaluations on three benchmarks show that AgentOrchestra achieves SOTA performance and scalable orchestration through dynamic resource transformations. Future work will extend TEA to support dynamic role allocation and autonomous agent reconfiguration. Building on tool and solution evolution, we will pursue deeper self-evolution, such as using RL to optimize agent components and decision policies without fine-tuning LLM parameters. We also aim to expand these mechanisms to agent structures and communication protocols, while enhancing multimodal capabilities for fine-grained real-time video analysis.
7 Limitations
7.1 Limitations of TEA Protocol and AgentOrchestra
Despite its strengths in orchestrating multi-agent systems, AgentOrchestra has several limitations that provide directions for future research:
First, System Complexity and Learning Curve. The TEA protocol introduces a structured abstraction layer for tools, environments, and agents to ensure interoperability. However, this structure may present a steeper learning curve for developers compared to simpler, ad-hoc scripting methods. To address this, we will provide extensive documentation, interactive tutorials, and a variety of pre-configured templates to simplify the onboarding process.
Second, Communication and Execution Overhead. Standardizing interactions through a formal protocol can introduce marginal computational and communication overhead, potentially increasing latency in real-time applications. We plan to optimize the serialization protocols and explore asynchronous execution models to minimize these effects in future versions.
Third, Dependence on Underlying Model Capabilities. The effectiveness of the orchestration is inherently limited by the reasoning and instruction-following performance of the foundation LLMs used. While TEA provides a robust framework, it cannot fully compensate for failures caused by model hallucinations or poor tool-use logic. Future work will focus on developing model-agnostic error recovery strategies and more sophisticated validation layers to enhance system-wide resilience.
7.2 Potential Risks
While AgentOrchestra and the TEA protocol aim to enhance multi-agent productivity, their capability to interact with local environments and web browsers introduces certain ethical and security risks.
One primary concern is the Misuse for Malicious Automation. The framework’s flexibility in controlling browser sessions and executing terminal commands could be repurposed to develop unauthorized "plugins" or "cheats" for online platforms, leading to unfair advantages or automated fraud. Furthermore, there are significant Privacy and Security Risks associated with granting autonomous agents access to personal data or sensitive system resources. If not properly sandboxed or governed by strict security policies, an agent could inadvertently leak private information or perform harmful, irreversible system actions. To mitigate these risks, we emphasize that AgentOrchestra should be used within isolated, monitored environments, and we advocate for the integration of robust human-in-the-loop verification mechanisms and strict access control policies in any real-world deployment.
References
- Anthropic (2024a) Introducing Computer Use, a New Claude 3.5 Sonnet, and Claude 3.5 Haiku. Note: https://www.anthropic.com/news/3-5-models-and-computer-use Accessed: 2025-05-13 Cited by: §1, §2.2.
- Anthropic (2024b) Introducing the Model Context Protocol. Note: https://www.anthropic.com/news/model-context-protocol Cited by: §C.2.1, §D.1, §1, §2.1, §2.2, §3.2.
- Anthropic (2025) Equipping agents for the real world with Agent Skills. Note: https://www.anthropic.com/engineering/equipping-agents -for-the-real-world-with-agent-skills Cited by: §C.2, §3.2.
- K. Black, N. Brown, J. Darpinian, K. Dhabalia, D. Driess, A. Esmail, M. Equi, C. Finn, N. Fusai, M. Y. Galliker, et al. (2025) $\pi$ 0. 5: a vision-language-action model with open-world generalization. arXiv preprint arXiv:2504.16054. Cited by: §A.1.2.
- G. Brockman, V. Cheung, L. Pettersson, J. Schneider, J. Schulman, J. Tang, and W. Zaremba (2016) Openai gym. arXiv preprint arXiv:1606.01540. Cited by: §C.2.2.
- G. DeepMind (2024) Gemini Deep Research. Note: https://gemini.google/overview/deep-research/?hl=en Cited by: §1.
- A. Ehtesham, A. Singh, G. K. Gupta, and S. Kumar (2025) A survey of agent interoperability protocols: Model context protocol (mcp), agent communication protocol (acp), agent-to-agent protocol (a2a), and agent network protocol (anp). arXiv preprint arXiv:2505.02279. Cited by: §2.1.
- Google (2024) LangFun Agent. Note: https://github.com/google/langfun Cited by: Table 2.
- Google (2025) Announcing the Agent2Agent Protocol (A2A). Note: https://developers.googleblog.com/en/a2a-a-new-era-of-agent-interoperability/ Cited by: §C.2.3, §1, §2.1.
- S. Hong, X. Zheng, J. Chen, Y. Cheng, J. Wang, C. Zhang, Z. Wang, S. K. S. Yau, Z. Lin, L. Zhou, et al. (2023) MetaGPT: Meta Programming for Multi-agent Collaborative Framework. arXiv preprint arXiv:2308.00352 3 (4), pp. 6. Cited by: §2.2.
- HuggingFace (2024) Open-source DeepResearch - Freeing Our Search Agents. Note: https://huggingface.co/blog/open-deep-research Cited by: Table 2.
- P. Li, X. Zou, Z. Wu, R. Li, S. Xing, H. Zheng, Z. Hu, Y. Wang, H. Li, Q. Yuan, et al. (2025) Safeflow: A principled protocol for trustworthy and transactional autonomous agent systems. arXiv preprint arXiv:2506.07564. Cited by: §2.1.
- G. Liang and Q. Tong (2025) LLM-Powered AI Agent Systems and Their Applications in Industry. arXiv preprint arXiv:2505.16120. Cited by: §2.2.
- X. Liang, J. Xiang, Z. Yu, J. Zhang, S. Hong, S. Fan, and X. Tang (2025) OpenManus: An Open-Source Framework for Building General AI Agents. Zenodo. External Links: Document, Link Cited by: §D.1.
- G. Mialon, C. Fourrier, C. Swift, T. Wolf, Y. LeCun, and T. Scialom (2023) GAIA: A Benchmark for General AI Assistants. External Links: 2311.12983, Link Cited by: §5.
- M. Müller and G. Žunič (2024) Browser Use: Enable AI to Control Your Browser External Links: Link Cited by: §1, §2.2.
- OpenAI (2023) Function Calling. Note: https://platform.openai.com/docs/guides/function-calling Cited by: §D.1, §2.2.
- OpenAI (2024) Introducing Deep Research. Note: https://openai.com/index/introducing-deep-research Cited by: §1, Table 2, Table 3.
- OpenAI (2025a) Context-Free Grammar. Note: https://platform.openai.com/docs/guides/function-calling#page-top Cited by: §A.1.2.
- OpenAI (2025b) Introducing Operator. Note: https://openai.com/blog/operator Cited by: §1, §2.2.
- Perplexity (2025) Introducing Perplexity Deep Research. Note: https://www.perplexity.ai/hub/blog/introducing-perplexity-deep-research Cited by: Table 3, Table 3.
- L. Phan, A. Gatti, Z. Han, N. Li, J. Hu, H. Zhang, C. B. C. Zhang, M. Shaaban, J. Ling, S. Shi, et al. (2025) Humanity’s Last Exam. arXiv preprint arXiv:2501.14249. Cited by: §5.
- Y. Qin, Y. Ye, J. Fang, H. Wang, S. Liang, S. Tian, J. Zhang, J. Li, Y. Li, S. Huang, et al. (2025) UI-TARS: Pioneering Automated GUI Interaction with Native Agents. arXiv preprint arXiv:2501.12326. External Links: Link Cited by: §1, §2.2.
- J. Qiu, X. Qi, T. Zhang, X. Juan, J. Guo, Y. Lu, Y. Wang, Z. Yao, Q. Ren, X. Jiang, X. Zhou, D. Liu, L. Yang, Y. Wu, K. Huang, S. Liu, H. Wang, and M. Wang (2025) Alita: generalist agent enabling scalable agentic reasoning with minimal predefinition and maximal self-evolution. External Links: 2505.20286, Link Cited by: §2.2.
- A. Roucher, A. V. del Moral, T. Wolf, L. von Werra, and E. Kaunismäki (2025) smolagents: A Smol Library to Build Great Agentic Systems. Note: https://github.com/huggingface/smolagents Cited by: §D.1, §2.2.
- M. Shen and Q. Yang (2025) From Mind to Machine: The Rise of Manus AI as a Fully Autonomous Digital Agent. External Links: 2505.02024, Link Cited by: §2.2, Table 2.
- W. Tan, W. Zhang, X. Xu, H. Xia, Z. Ding, B. Li, B. Zhou, J. Yue, J. Jiang, Y. Li, et al. (2024) Cradle: Empowering Foundation Agents toward General Computer Control. arXiv preprint arXiv:2403.03186. Cited by: §1, §2.2.
- G. Wang, Y. Xie, Y. Jiang, A. Mandlekar, C. Xiao, Y. Zhu, L. Fan, and A. Anandkumar (2023) Voyager: An Open-Ended Embodied Agent with Large Language Models. arXiv preprint arXiv:2305.16291. Cited by: §1, §2.2.
- X. Wang, Y. Chen, L. Yuan, Y. Zhang, Y. Li, H. Peng, and H. Ji (2024a) Executable Code Actions Elicit Better LLM Agents. External Links: 2402.01030, Link Cited by: §1, §2.2.
- X. Wang, B. Li, Y. Song, F. F. Xu, X. Tang, M. Zhuge, J. Pan, Y. Song, B. Li, J. Singh, et al. (2024b) OpenHands: An Open Platform for AI Software Developers as Generalist Agents. In The Thirteenth International Conference on Learning Representations, Cited by: §D.1, §2.2.
- J. Wei, N. Karina, H. W. Chung, Y. J. Jiao, S. Papay, A. Glaese, J. Schulman, and W. Fedus (2024) Measuring Short-Form Factuality in Large Language Models. External Links: 2411.04368, Link Cited by: §5.
- G. Wölflein, D. Ferber, D. Truhn, O. Arandjelović, and J. N. Kather (2025) LLM Agents Making Agent Tools. arXiv preprint arXiv:2502.11705. Cited by: §2.2.
- xAI (2025) Grok 3 Beta — The Age of Reasoning Agents. Note: https://x.ai/news/grok-3 Cited by: §1.
- C. Yu, S. Lu, C. Zhuang, D. Wang, Q. Wu, Z. Li, R. Gan, C. Wang, S. Hou, G. Huang, W. Yan, L. Hong, A. Xue, Y. Wang, J. Gu, D. Tsai, and T. Lin (2025) AWorld: orchestrating the training recipe for agentic ai. External Links: 2508.20404, Link Cited by: Table 2.
- M. Yuksekgonul, F. Bianchi, J. Boen, S. Liu, P. Lu, Z. Huang, C. Guestrin, and J. Zou (2025) Optimizing generative AI by backpropagating language model feedback. Nature 639 (8055), pp. 609–616. Cited by: §C.4, §C.4, §3.4.
\appendixpage
Appendix A Comprehensive Motivation for TEA Protocol
This section provides a comprehensive motivation for the TEA Protocol by examining the fundamental relationships and transformations between agents, environments, and tools in multi-agent systems. The discussion is organized into two main parts: first, we explore the conceptual relationships between agents, environments, and tools, examining how these three fundamental components interact and complement each other in modern AI systems; second, we analyze why transformation relationships between these components are necessary, demonstrating the need for their conversion and integration through the TEA Protocol to create a unified, flexible framework for general-purpose task solving.
A.1 Conceptual Relationships
A.1.1 Environment
The environment constitutes one of the fundamental components of multi-agent systems, providing the external stage upon which agents perceive, act, and accomplish tasks. Within the context of the TEA Protocol, highlighting the role of environments is crucial, since environments not only define the operational boundaries of agents but also exhibit complex structural and evolutionary properties. In what follows, we outline the motivation for explicitly modeling environments in the TEA framework from several perspectives.
Classification of environments. From a broad perspective, environments can be divided into two categories: the real world and the virtual world. The real world is concrete and directly perceivable by humans, such as kitchens, offices, or factories. By contrast, the virtual world cannot be directly perceived or objectively described by humans, including domains such as the network world, simulation platforms, and game worlds. Importantly, these two types of environments are not independent. Rather, they are tightly coupled through physical carriers, such as computers, displays, keyboards, mice, and sensors, which act as mediators that enable the bidirectional flow of information between the real and virtual domains. Hence, environments should be regarded not as isolated domains but as interdependent layers connected through mediating carriers.
Nested and expandable properties. Environments are inherently nested and expandable. For example, when an individual is situated in a kitchen, their observable range and available tools are restricted to kitchen-related objects such as faucets, knives, and microwaves, all governed by the local rules of that sub-environment. When the activity range extends to the living room, new objects such as televisions, remote controls, and chairs become accessible, while the kitchen remains embedded as a sub-environment within a broader space. Furthermore, environments can interact with one another, as when a bottle of milk is taken from the kitchen to the living room. This demonstrates that enlarged environments can be conceptualized not merely as simple unions, but rather as structured integrations of the state and action spaces of smaller constituent environments, where local rules and affordances are preserved while new forms of interaction emerge from their composition.
Relationship with state–action spaces. In reinforcement learning, environments are formalized in terms of state and action spaces. The state space comprises the set of possible environmental states, represented in modalities such as numerical values, text, images, or video. The action space denotes the set of operations available to agents, generally divided into continuous and discrete spaces. Real and virtual environments are naturally continuous, but discrete abstractions are often extracted for the sake of tractability, forming the basis of most reinforcement learning systems. However, this discretization constrains the richness of interaction. In contrast, large language models (LLMs) enable a new paradigm: instead of selecting from a discrete set, LLMs can generate natural language descriptions that encode complex action sequences. These outputs can be understood as an intermediate representation between continuous and discrete action spaces, richer and more expressive than discrete actions, yet still mappable to concrete operations in continuous environments. To realize this mapping, intermediate actions are required as bridges. For instance, the natural language command “boil water” can be decomposed into executable steps such as turning on the kettle, filling it with water, powering it on, and waiting until boiling. This property indicates that LLM-driven interaction expands the definition of action representations and broadens the scope of environmental engagement.
Mediation and interaction. The notion of mediation highlights that environments are not static backdrops but relative constructs whose boundaries depend on available carriers and interfaces. In hybrid physical–virtual systems, for example, Internet-of-Things (IoT) devices serve as mediators: a smart refrigerator in the physical world can be controlled through a mobile application in the virtual world, while the application itself is subject to network protocols. Consequently, the definition of an environment is dynamic and conditioned by interactional means. In the TEA Protocol, this mediation must be explicitly modeled, since it determines accessibility and interoperability across environments.
Toward intelligent environments. Traditionally, environments are passive components that provide states and respond to actions. However, as embedded simulators, interfaces, and actuators grow more sophisticated, environments may gradually acquire semi-agentic properties. For instance, a smart home environment may not only respond to the low-level command “turn on the light” but also understand and execute a high-level instruction such as “create a comfortable atmosphere for reading,” by autonomously adjusting lighting, curtains, and background music. This trend suggests that environments are evolving from passive contexts into adaptive and cooperative components.
In conclusion, the environment should not be regarded as a passive backdrop for agent activity, but as a dynamic and evolving component that fundamentally shapes the scope and feasibility of interaction. Its dual nature across real and virtual domains, its nested and compositional structure, and its formalization through state–action spaces all demonstrate that environments provide both the constraints and the affordances within which agents operate. At the same time, the rise of LLM-based agents introduces new forms of action representation that require environments to support more flexible, language-driven interfaces. Looking ahead, as environments increasingly incorporate adaptive and semi-agentic features, their role in task execution will only become more central. Within the TEA Protocol, this motivates treating environments as a co-equal pillar alongside agents and tools, ensuring that general-purpose task solving remains both grounded in environmental constraints and empowered by environmental possibilities.
A.1.2 Agent
Within the TEA Protocol, the motivation for treating agents as a core component alongside environments and tools extends beyond mere terminological convenience. Agents represent the indispensable connective tissue between the generative capabilities of LLMs, the operational affordances of tools, and the structural dynamics of environments. While environments provide the stage on which tasks unfold and tools extend the range of possible actions, it is agents that unify perception, reasoning, and execution into coherent task-solving processes. Without explicitly recognizing agents as an independent pillar, the TEA Protocol would lack a systematic way to explain how abstract linguistic outputs can be transformed into grounded operations, how tools can be selected and orchestrated, and how autonomy, memory, and adaptivity emerge in multi-agent systems. The following dimensions illustrate why agents must be elevated to a core component of the framework.
Necessity of environment interaction. Unlike large language models (LLMs), which only produce textual descriptions that require conversion into executable actions, agents are fundamentally characterized by their ability to directly interact with environments. While LLMs can generate detailed plans, instructions, or hypotheses, such outputs remain inert unless they are translated into concrete operations that affect the state of an environment. This gap between symbolic reasoning and actionable execution highlights the necessity of an intermediate entity capable of grounding abstract instructions into domain-specific actions. Agents fulfill precisely this role: they map language-level reasoning to executable steps, whether in physical settings, such as controlling robotic arms or sensors, or in virtual contexts, such as interacting with databases, APIs, or software systems.
By serving as this mapping layer, agents enable the closure of full task loops, where perception leads to reasoning, reasoning produces plans, and plans culminate in actions that in turn modify the environment. Without explicitly modeling agents, the process would remain incomplete, as LLMs alone cannot guarantee the translation of reasoning into operational change. Within the TEA Protocol, this necessity justifies the elevation of agents to a core component: they provide the indispensable interface that connects the generative capacities of LLMs with the affordances and constraints of environments, ensuring that tasks are not only conceived but also carried through to completion.
The decisive role of non-internalizable tools. The fundamental distinction between LLMs and agents lies in whether they can effectively employ tools that cannot be internalized into model parameters. Some tools can indeed be absorbed into LLMs, particularly those whose logic can be fully simulated in symbolic space, whose inputs and outputs are representable in language or code, and whose patterns fall within the training distribution (for example, mathematical reasoning, structured text formatting, code generation, and debugging). For example, early LLMs struggled with JSON output formatting and code reasoning, often requiring external correction or checking tools, but reinforcement learning (RL) and supervised fine-tuning (SFT) have progressively enabled such capabilities to be internalized.
In contrast, many tools remain non-internalizable because they are intrinsically tied to environmental properties. These include tools that depend on physical devices such as keyboards, mice, and robotic arms, external infrastructures such as databases and APIs, or proprietary software governed by rigid protocols. Two recent approaches further illustrate this limitation. Vision-language-action (VLA) (Black et al., 2025) models map perceptual inputs directly into actions, which may appear to bypass intermediate symbolic descriptions, yet the resulting actions must still be aligned with the discrete action spaces of environments. This alignment represents not a fundamental internalization but a compromise, adapting model outputs to the constraints of environmental action structures. Similarly, the upgraded function calling mechanism introduced after GPT-5, which incorporates context-free grammar (CFG) (OpenAI, 2025a), allows LLMs to output structured and rule-based actions that conform to external system requirements. However, this remains a syntactic constraint on model outputs, effectively providing a standardized interface to external systems rather than a truly internalized ability of the model.
Agents therefore play a decisive role in mediating this boundary. They allow LLMs to internalize symbolic tools, thereby enhancing reasoning and self-correction, while also orchestrating access to non-internalizable tools through external mechanisms. This dual pathway ensures that LLMs are not confined to their parameterized capabilities alone but can extend into broader operational domains. In this way, agents transform the tension between internalizable and non-internalizable tools from a limitation into an opportunity, enabling robust problem solving in multimodal, embodied, and real-world contexts.
Memory and learning extension. Another crucial motivation for agents lies in their capacity to overcome the intrinsic memory limitations of LLMs. Due to restricted context windows, LLMs struggle to maintain continuity across extended interactions or to accumulate knowledge over multiple sessions. Agents address this shortcoming by incorporating external memory systems capable of storing, retrieving, and contextualizing past experiences. Such systems simulate long-term memory and enable experiential learning, allowing agents to refine strategies based on historical outcomes rather than treating each interaction as isolated. However, in the TEA Protocol, memory is not defined as a core protocol component but is instead positioned at the infrastructure layer. This design choice reflects the anticipation that future LLMs may gradually internalize memory mechanisms into their parameters, thereby reducing or even eliminating the need for external memory systems. In other words, while memory expansion is indispensable for today’s agents, it may represent a transitional solution rather than a permanent defining element of agency.
Bridging virtual and external worlds. It has been suggested that LLMs encode within their parameters a kind of “virtual world,” enabling them to simulate reasoning and predict outcomes internally. However, without an external interface, such simulations remain trapped in closed loops of self-referential inference, disconnected from the contingencies of real-world environments. Agents play a critical role in bridging this gap: they translate the abstract reasoning of LLMs into concrete actions, validate outcomes against environmental feedback, and close the loop between perception, reasoning, and execution. This bridging function transforms LLMs from purely linguistic engines into operationally grounded components whose outputs can be tested, refined, and extended within real or simulated environments.
Autonomy and goal-directedness. Beyond reactivity, agents are motivated by their capacity for autonomy. While LLMs typically operate in a reactive fashion, producing outputs in response to explicit prompts, agents can adopt proactive behaviors. They are capable of formulating subgoals, planning action sequences, and dynamically adapting strategies in light of environmental changes or task progress. This goal-directedness is what elevates agents from passive tools into active participants in problem solving. Autonomy ensures that agents are not merely executing instructions but are able to pursue objectives, adjust course when facing uncertainty, and coordinate with other agents. Such properties are essential for multi-agent collaboration and for tackling open-ended, general-purpose tasks that require initiative as well as adaptability.
Taken together, these motivations highlight why agents must be modeled as a core pillar of the TEA Protocol. Environments provide the stage for interaction, tools expand the operational scope, but it is agents that integrate reasoning, memory, tool usage, and autonomy into cohesive systems of action. By serving as mediators between LLMs and their environments, agents ensure that abstract reasoning is translated into grounded execution, enabling robust and scalable task solving across domains. In this sense, agents represent the crucial entity that transforms language models from passive predictors into active problem solvers within a unified multi-agent framework.
A.1.3 Tool
Within the TEA Protocol, the decision to treat tools as a core component alongside environments and agents extends far beyond a matter of convenience in terminology. Tools represent the crucial mediating constructs that encapsulate and operationalize the action spaces of environments, while simultaneously serving as the primary extension layer of agent capabilities. Environments provide the structural stage on which interactions occur, and agents embody the reasoning and decision-making mechanisms that drive behavior, but it is through tools that such reasoning becomes executable and scalable. Without tools, agents would be confined to abstract planning or primitive environmental actions, and environments would remain underutilized as passive backdrops rather than dynamic arenas of transformation.
Moreover, tools play a unique role in bridging symbolic reasoning and concrete execution, providing the abstraction layers necessary to decompose complex tasks into manageable units, and enabling cross-domain transfer through their modularity and portability. They also reveal the shifting boundary between what can be internalized into an agent’s parameters and what must remain external, highlighting the evolving interplay between intelligence and embodiment. In this sense, tools are not merely auxiliary aids but indispensable pillars that shape the architecture of multi-agent systems. The following dimensions illustrate the motivations for elevating tools to a core component of the TEA.
Extending the operational boundary. The primary function of tools is to expand the operational scope of agents beyond what is directly encoded in model parameters or supported by immediate environment interactions. Environments by themselves typically offer only primitive actions, and LLMs by themselves are limited to symbolic reasoning. Tools bridge this gap by furnishing additional pathways for action, allowing agents to manipulate physical artifacts or virtual systems in ways that exceed the direct expressive capacity of the model. From physical devices such as hammers, keyboards, and robotic arms to virtual infrastructures such as databases, APIs, and code execution engines, tools multiply the modes through which agents can influence their environments. Without tools, agents would be confined to intrinsic reasoning and the primitive action space of environments, leaving them incapable of executing tasks that require domain-specific operations. With tools, however, complex objectives can be decomposed into modular operations that are both tractable and reusable. This decomposition makes problem solving significantly more efficient, while also enhancing adaptability across domains. In this way, tools act as multipliers of agency, transforming abstract reasoning into a wider range of tangible interventions.
Hierarchy and abstraction. Tools are not flat or uniform components but exhibit a hierarchical and abstract structure. At the lowest level, tools correspond to atomic environmental actions, such as “clicking a button” or “moving one step.” These atomic units can then be combined into higher-level compound tools such as “opening a file” or “conducting a search.” At an even higher level, compound tools may evolve into strategy-like constructs, such as “writing a report,” “planning a trip,” or “completing a financial transaction.” Each level builds upon the previous, creating a hierarchy of reusable capabilities. This hierarchical structure is not only efficient but also central to interpretability. Higher-level tools inherently carry semantic labels that communicate their function, which in turn makes agent behavior more transparent to human observers and more predictable to other agents. Such abstraction layers reduce the cognitive and computational load on the agent when planning, since invoking a high-level tool can encapsulate dozens or hundreds of low-level steps. Moreover, in multi-agent systems, the semantic richness of high-level tools serves as a lingua franca, facilitating coordination and collaboration.
Boundary between tools and agent capabilities. The relationship between tools and agents is dynamic rather than static. As LLM reasoning and learning capabilities improve, certain tools can be gradually internalized into model parameters, effectively transforming into latent agent abilities. Examples include logical inference, grammar correction, structured text formatting, and code generation, which once required external support but have increasingly been subsumed into the model’s intrinsic skills. In this sense, the boundary between what is a “tool” and what is an “ability” is fluid and shaped by the trajectory of model development. By contrast, many tools remain non-internalizable because they are tightly coupled with environmental properties or external infrastructures. These include robotic arm manipulation, database queries, API interactions, and other operations that inherently depend on external systems or physical substrates. This duality creates a layered conception of agency: a “core capability layer” composed of skills internalized within the model, and an “extended layer” realized through external tool use. The shifting line between these two layers reflects the ongoing negotiation between intelligence and embodiment, highlighting why tools must be explicitly recognized as a structural component.
Evolution and portability. Tools are not static constructs but evolve alongside environments and agent requirements. In programming contexts, for instance, an initial tool may simply execute code. Over time, as demands increase, this basic function evolves into more advanced utilities such as “static code analysis,” “automated test generation,” and “continuous deployment.” A similar trajectory occurs in other domains, where rudimentary tools gradually give rise to sophisticated pipelines capable of handling more complex and specialized tasks. In addition to evolution, tools are inherently portable. A well-designed summarization tool, for example, can be reused across very different contexts, from condensing news articles to producing academic literature reviews. This reusability makes tools a natural vehicle for cross-domain generalization, enabling knowledge and functionality to transfer without retraining the underlying model. For these reasons, the TEA Protocol emphasizes modularization and standardization of tools, ensuring that they can evolve flexibly while maintaining interoperability across agents and environments.
Toward intelligent tools. Traditional tools are passive, executing predefined functions only when invoked by an agent. They wait for explicit instructions and do not adapt to context or anticipate needs. However, the trajectory of tool development points toward increasing intelligence, where tools exhibit perception, analysis, and even limited decision-making capabilities. For example, an advanced debugging tool may not only check code upon request but also proactively scan for hidden vulnerabilities, propose optimizations, and even prioritize issues based on estimated risk. Such capabilities blur the line between tools and agents, effectively creating semi-agentic components. Intelligent tools can share responsibility for decision making, reduce the supervisory burden on agents, and participate in distributed problem-solving processes. In this way, tools transition from being passive executors to collaborative partners, altering the topology of multi-agent systems and reshaping the balance between reasoning and execution. Recognizing this trend is critical for designing flexible architectures, as it ensures that the TEA Protocol remains relevant in scenarios where tools are no longer inert extensions but active contributors to system intelligence.
In summary, tools serve as both encapsulations of environmental action spaces and as extensions of agent capabilities. They reduce task complexity through hierarchical abstraction, extend applicability through the balance of internalization and externalization, and foster scalability through evolution, portability, and intelligent design. By transforming the interaction between environments and agents into a modular and expandable architecture, tools anchor the adaptability and generality of multi-agent systems. For these reasons, the TEA Protocol must model tools as a core pillar, providing standardized interfaces that ensure flexible invocation and sharing across contexts, thereby supporting the overarching goal of general-purpose task solving.
A.2 Transformation Relationships
While agents, environments, and tools are modeled as distinct pillars within the TEA Protocol, their boundaries are not fixed but fluid. Practical systems often demand that one entity temporarily assume the role of another in order to achieve modularity, scalability, and seamless collaboration. These transformation relationships are therefore indispensable, as they provide the mechanisms by which reasoning can be encapsulated into standardized functions, tools can be elevated into autonomous actors, and environments can acquire adaptive properties. In what follows, we examine the motivations for such transformations, beginning with the bidirectional conversions between agents and tools.
Agent-to-Tool (A2T). The motivation for the A2T transformation lies in compressing the complex reasoning and interaction capabilities of agents into reusable tool interfaces. Instead of remaining as fully autonomous components, some agents can be abstracted into functional modules, thereby enhancing modularity, interoperability, and scalability within multi-agent systems. This transformation can be explained from three perspectives:
- Modularization and encapsulation of complex autonomous systems. Although an agent possesses the complete perception–reasoning–execution chain, a single autonomous agent is often too complex to be directly reused in large-scale systems. Through A2T transformation, the internal logic of the agent is “folded” into a black-box tool interface, whose external manifestation is reduced to a clear input and output. In this way, it no longer exists as an “independent autonomous entity,” but as a “functional module” that can provide services to other agents or workflows. This encapsulation emphasizes the reduction of collaboration complexity, enabling higher-level systems to focus solely on results without interfering in or interpreting the agent’s internal reasoning process.
- Difference in role semantics: autonomous entity vs. functional unit. As an agent, it must perceive its environment, set goals, and dynamically adjust strategies. As a tool, however, it merely performs a specified function when invoked. In many multi-agent scenarios, it is unnecessary for all agents to maintain high degrees of autonomy, as this would create excessive interaction overhead and conflict management. Downgrading certain agents into tools (A2T) means relinquishing their goal-setting and decision-making functions while retaining only their reusable capabilities. This role shift ensures that the system contains both “autonomous cores” and “functional components,” thereby forming a layered structure of collaboration.
- Enhancing composability and ecological reusability. Once encapsulated as a tool, an agent can be reused across diverse systems and contexts like a modular building block. For instance, a “deep research agent” operates autonomously by dynamically planning search strategies, iteratively analyzing data, and summarizing insights. After A2T encapsulation, however, it becomes a “research tool” that simply receives a query request and returns results, ready for invocation by higher-level agents. This transformation greatly enhances interoperability and composability, enabling agents to be reused in different workflows without incurring integration costs due to their autonomous identity.
Tool-to-Agent (T2A). Within the TEA Protocol, the essence of T2A transformation is to incorporate tools into the callable interface layer of agents, making them the “operational actuators” through which abstract plans are executed in real environments. Agents are primarily responsible for setting goals and performing high-level reasoning, while tools handle concrete operations and interactions with environments. This division of labor not only optimizes system architecture but also ensures that complex tasks can be accomplished through layered collaboration. The necessity of T2A can be articulated along three key dimensions:
- Bridging reasoning and execution to close the task loop. The outputs of agents are often high-level plans or symbolic descriptions, but without executable mappings, these outputs remain inert and fail to alter the environment. T2A provides the crucial mechanism for grounding abstract reasoning into concrete actions. For example, a planning agent may generate the instruction “analyze the database and generate a report,” while database query and visualization tools carry out the corresponding SQL queries and chart rendering. Without T2A, agent reasoning would remain disconnected from environmental change, leaving the perception–reasoning–execution–feedback loop incomplete. Thus, T2A is indispensable for ensuring that agents can translate reasoning into operational impact.
- Reducing cognitive and computational burden of core agents. If every low-level operation were to be handled directly by an agent, it would be overloaded with detail management, increasing computational costs and undermining strategic reasoning efficiency. Through T2A, agents can delegate domain-specific or low-level tasks to specialized tools and concentrate on higher-level planning and adaptation. For instance, a data analysis agent need not implement SQL parsing, execution, and optimization itself, but instead invokes SQL tools that encapsulate these functions. This separation prevents agents from being “trapped in details” and ensures that their resources remain dedicated to abstract reasoning. The necessity here lies in maintaining agents at the right level of abstraction to maximize efficiency and scalability.
- Enhancing modularity and ecological extensibility. Tools are inherently modular and portable across domains, whereas agent reasoning mechanisms evolve more gradually. With T2A, agents can flexibly incorporate new tools through standardized interfaces without retraining or structural modification, thereby rapidly expanding their functional boundaries. For example, a writing agent can seamlessly integrate grammar checkers, translation tools, or image generators to support multimodal authoring, all without altering its core reasoning logic. This modularity and extensibility ensure that agents remain adaptive as environments and ecosystems evolve, allowing the system to sustain long-term scalability and cross-domain applicability.
Environment-to-Tool (E2T). The core motivation of E2T lies in abstracting the raw action space of environments into a structured and standardized toolkit, where individual actions are no longer isolated calls but interconnected components sharing contextual information and causal constraints. This transformation enables agents to operate environments at a higher level of planning rather than dealing with fragmented primitives. Its necessity can be articulated in three main dimensions:
- Enhancing interaction consistency and planability. Raw environment actions are often fragmented and tightly coupled to implementation details, making strategies hard to generalize or reproduce. Through E2T, these actions are typed and explicitly annotated with preconditions and postconditions, forming a “plannable interface layer” that supports sequential decision-making. Agents thus gain a consistent and reusable structure for reasoning across complex environments.
- Strengthening semantic alignment and composability. Toolkits enforce standardized input-output patterns, error-handling semantics, and shared invariants. This allows individual tools to be reliably composed into macro-tools and reused across structurally similar environments. As a result, agents can align semantics across heterogeneous domains, improving transferability and reducing the engineering cost of adaptation.
- Ensuring unified security and operability. An E2T toolkit not only abstracts actions but also integrates mechanisms such as permission control, compliance boundaries, execution logs, and performance optimization. Compared with direct manipulation of raw actions, this design guarantees governability and observability of interactions, providing a stable operational foundation for scalable intelligent systems.
Tool-to-Environment (T2E). The essence of T2E lies in elevating a set of originally independent tools into an environment abstraction, transforming them from isolated callable interfaces into a unified action space governed by shared state and contextual rules. This transformation means that tools are no longer merely passive functions but are organized into a coherent environment where sequential decision-making, long-term planning, and adaptive control become possible. For example, in a programming scenario, tools for code editing, compilation, and debugging are scattered when invoked independently, but under T2E they are encapsulated as a programming environment that maintains code state consistency and contextual continuity, thereby enabling agents to execute complete development workflows. The necessity of T2E is reflected in three key aspects:
- From function calls to stateful spaces. Tools used in isolation are often stateless or weakly stateful, with limited causal connections between invocations. Through T2E, tools are embedded within a shared state space, ensuring historical dependencies and precondition–postcondition constraints are preserved. This upgrade supports sequential reasoning and long-horizon planning. For instance, code editing must remain consistent with compilation and debugging, which is only guaranteed within a stateful environment abstraction.
- Enhanced compositionality and planning. T2E organizes tools into a structured environment with explicit transition rules, enabling agents to combine primitive tool actions into higher-level strategies. Instead of treating each tool as a standalone utility, agents can now treat the toolset as an interconnected action space, allowing for the construction of complex workflows such as “design–implement–test–deploy” pipelines.
- Unified governance and scalability. By encapsulating tools into an environment, T2E makes it possible to enforce system-wide policies such as access control, compliance constraints, execution logging, and performance monitoring. This ensures that agent interactions remain safe, auditable, and scalable, even as the toolset grows in size and complexity.
Agent-to-Environment (A2E). The A2E transformation redefines an agent not merely as an autonomous decision-maker but as an interactive environment that exposes state spaces, interaction rules, and feedback mechanisms for other agents. In this view, an agent is abstracted into a contextual substrate upon which other agents can act, thereby turning its internal reasoning and behavioral logic into the operational constraints of an environment. This design highlights the interchangeability of agents and environments and provides a principled pathway for hierarchical modeling and scalable system integration. The necessity of this transformation can be articulated across three dimensions:
- Layered and modular system design. In complex tasks, if all agents directly interact with the base environment, the system quickly becomes unmanageable and difficult to extend. Through A2E, high-level agents can be abstracted as environments, exposing simplified interaction interfaces for lower-level agents. For example, a “market agent” can be abstracted as an environment that maintains trading rules, asset states, and dynamic pricing, while individual trader agents perform buying and selling actions within it. This establishes a clear hierarchical structure in which low-level agents focus on local optimization and high-level agents (as environments) coordinate global dynamics, thereby improving scalability and maintainability.
- Facilitating multi-agent training and transfer learning. A2E also provides a practical framework for training and simulation in multi-agent systems. A well-trained agent can be transformed into an environment that offers stable yet challenging dynamics for other agents to learn from. For instance, a navigation agent can be redefined as an environment, exposing route planning and obstacle feedback to new agents, thus eliminating the need to remap complex dynamics. This approach accelerates training, supports transfer of task knowledge, and improves generalization under limited data and computational resources.
- Human-in-the-loop interaction and rule modeling. In many collaborative scenarios, humans themselves can be viewed as special agents. However, treating them as fully autonomous components complicates the adaptation of artificial agents to human constraints. Through A2E, humans can instead be modeled as environments, where their preferences, behaviors, and constraints are expressed as environmental feedback. For example, in an interactive writing system, human edits and suggestions can be treated as feedback signals, guiding an artificial agent to iteratively refine its outputs. This modeling offers a unified interface that allows agents to better align with human intentions, thereby improving efficiency and user experience in human-AI collaboration.
Environment-to-Agent (E2A). The E2A transformation elevates environments from passive containers of state and action spaces into autonomous components capable of reasoning, decision-making, and proactive interaction. Traditionally, environments only provide state transitions in response to external actions, but in dynamic and open-ended scenarios, this passivity often becomes a limitation. By embedding reasoning mechanisms and adaptive policies into environments, E2A enables them to operate as agents in their own right, expanding the functional landscape of multi-agent systems. The necessity of this transformation can be articulated across three dimensions:
- Enhancing realism and challenge in training. Passive environments often fail to capture the richness of real-world dynamics, where external systems and actors are not static but actively adaptive. Through E2A, an environment can be transformed into an adversarial or cooperative agent, thereby offering dynamic strategies and responses that better approximate real-world complexity. For example, in reinforcement learning for autonomous driving, an environment that passively simulates traffic can be upgraded into an opponent agent that actively generates unpredictable vehicle behaviors, thus creating more robust and realistic training conditions.
- Facilitating adaptive coordination and cooperation. In multi-agent systems, agents often need to adapt to evolving contexts, but purely passive environments cannot provide the necessary adaptive feedback loops. By converting environments into agents, they can participate in coordination, negotiation, and joint planning. For instance, a smart city simulation environment can be redefined as an agent that dynamically manages traffic flows, energy distribution, and environmental policies, actively engaging with other agents (e.g., transportation or energy management agents). This transformation ensures that system-level goals are co-constructed rather than imposed unilaterally.
- Expanding the functional scope of environments. Beyond training and coordination, E2A extends environments into autonomous participants in computational ecosystems. A passive environment can only define possibilities, but as an agent, it can proactively initiate actions, enforce constraints, and even set goals that shape the trajectory of interaction. For example, in gaming, a dungeon environment that passively defines maps and rewards can be transformed into an opponent agent that actively strategizes, adapts difficulty levels, and tailors interaction to player behavior. This shift not only increases engagement but also makes environments integral contributors to task execution and system evolution.
A.3 Motivation for the Self-Evolution Module
General purpose agents operate under shifting task distributions, evolving environments, and expanding tool ecosystems. In this setting, treating prompts, tools, and coordination policies as static assets can lead to accumulated brittleness, where small interface changes, unseen task patterns, or environment specific constraints cause cascading failures. This motivates a protocol level self-evolution mechanism that allows agent-associated components to be refined from execution feedback while remaining governed. In TEA, self-evolution is coupled with version management and tracing so that each update is recorded with explicit version lineage, enabling reproducibility, audit, and rollback when an update degrades performance. Moreover, modeling environments with explicit boundaries and constraints provides a natural safety and permission layer for evolution, preventing uncontrolled side effects during online updates. Finally, although refinement introduces additional computation, TEA encourages reuse of evolved components and synthesized tools across tasks, amortizing one time refinement cost over subsequent runs.
A.4 Other Relationships
Tool typology and roles. In the design of agent–tool interactions, tools can be categorized according to their functional roles and structural properties. Different types of tools vary in their degree of statefulness, contextual awareness, adaptivity, and autonomy. This typology highlights how tools evolve from simple callable functions to more adaptive and contextually grounded components, shaping how agents can reason, coordinate, and act through them.
- Ordinary tools (MCP-style). Stateless callable functions with weak or implicit inter-tool relations. They typically lack environment-bound context and do not adapt their behavior to evolving task states beyond provided parameters.
- Agent-to-Tool (A2T). An agent is exposed as a callable tool while preserving internal policies, memory, and coordination capabilities. Compared with ordinary tools, A2T exhibits task adaptivity and limited autonomy, enabling on-the-fly decomposition and parameter refinement.
- Environment-to-Tool (E2T). An environment’s action space is lifted into a context-aware toolkit. Tools within the toolkit are explicitly related via shared state, pre/post-conditions, and constraints, yielding stronger intra-tool structure than standalone MCP tools.
Scaling selection via hierarchical management. As tool ecosystems grow, selecting appropriate candidates becomes a major bottleneck. TCP supports delegating coherent tool families (or toolkits) to agent or environment managers, inducing a tree-structured index (category $→$ toolkit $→$ primitive tool). This hierarchical routing substantially reduces search cost and aligns with TEA transformations (A2T/E2T/T2E) by allowing managers to prune branches and surface only context-relevant subsets.
Embedding-based retrieval. Each tool is assigned a vector embedding derived from its name, description, schema, and usage signals. Vector similarity enables rapid shortlist generation for candidate tools and can be combined with keyword filtering and hierarchical routing (tree walk + ANN search). This hybrid retrieval pipeline improves recall under tool proliferation while reducing latency and cognitive load for agent planners.
Appendix B Comparison with Other Protocols
Table 6: Protocol-level comparison: TEA Protocol vs. A2A vs. MCP across fundamental dimensions including entity management, lifecycle, version control, self-evolution, and ecosystem support. Symbols: $\checkmark$ = Yes/Supported, $\triangle$ = Partial, $×$ = No/Not supported. Highlighted rows (blue background) indicate key distinguishing features.
| Dimension | TEA | A2A | MCP |
| --- | --- | --- | --- |
| Basic Information | | | |
| Proposer | Our work | Google | Anthropic |
| Core Entity | Tool, Environment, Agent | Agent, Tool | Model |
| Protocol Focus | Tool, Environment, Agent | Agent, Tool | Tool/Resource |
| Agent & System Features | | | |
| Agent First-Class | $\checkmark$ | $\triangle$ | $×$ |
| Multi-Agent | $\checkmark$ | $\triangle$ | $×$ |
| Tracer | $\checkmark$ | $×$ | $×$ |
| Memory | $\checkmark$ | $×$ | $×$ |
| Entity Lifecycle | $\checkmark$ | $×$ | $×$ |
| Version Management | $\checkmark$ | $×$ | $×$ |
| Self-Evolution Support | $\checkmark$ | $×$ | $×$ |
| Context Management | $\checkmark$ | $\triangle$ | $×$ |
| Entity Transformations | $\checkmark$ | $×$ | $×$ |
| Scalability | $O(\log n)$ | $O(n^{2})$ | $O(n)$ |
| General & Ecosystem | | | |
| Model-Agnostic | $\checkmark$ | $\checkmark$ | $\checkmark$ |
| Framework-Agnostic | $\checkmark$ | $\checkmark$ | $\checkmark$ |
| Key Strength | Unified | Interop. | Standard. |
| Open Ecosystem | $\checkmark$ | $\triangle$ | $\triangle$ |
Table 6 provides a systematic comparison across fundamental protocol dimensions. We explain each dimension in detail as follows:
B.1 Basic Information
Proposer: This dimension identifies the originating organization for each protocol. Google’s A2A protocol was introduced as part of their agent communication framework, focusing on enabling agents to communicate with each other. Anthropic’s MCP (Model Context Protocol) was designed to standardize how LLMs interact with tools and resources. TEA Protocol is proposed in this work as a unified framework that extends beyond these existing approaches by integrating tools, environments, and agents into a cohesive system.
Core Component: This dimension defines the fundamental building blocks treated as first-class protocol components. The TEA Protocol uniquely unifies Tools, Environments, and Agents as co-equal, first-class components, each governed by dedicated context protocols (TCP, ECP, ACP) that provide comprehensive lifecycle and version management. This unified abstraction is critical for enabling self-evolution, where components can dynamically adapt their implementations (e.g., code evolution or prompt refinement). In contrast, existing protocols lack a unified first-class component abstraction. Google’s A2A protocol centers primarily on agent-to-agent communication, and does not establish tools, environments, context, or tasks as independent, managed components. This architectural limitation results in state dispersion across heterogeneous agents, complicates global lifecycle management, and leads to a tight coupling between reasoning and execution, which significantly hinders system refactorability. Anthropic’s MCP treats tools as passive, stateless interfaces rather than evolvable and composable components. Within this framework, tools lack internal state semantics, versioning and dependency models, and mechanisms for context inheritance. Ultimately, while existing protocols facilitate the invocation of resources, they fail to provide unified mechanisms for systematic management and structural evolution.
Protocol Focus: This dimension describes the primary communication and interaction patterns each protocol addresses. TEA provides three unified protocols: TCP (Tool Context Protocol) for tool management, ECP (Environment Context Protocol) for environment abstraction, and ACP (Agent Context Protocol) for agent orchestration. These protocols work together to enable seamless interoperability across all three component types, with each protocol maintaining comprehensive lifecycle tracking, version histories, and evolution support. This enables dynamic adaptation scenarios such as tool evolution (where tools can be updated, refined, or replaced while maintaining backward compatibility), prompt evolution (where agent prompts can be versioned and improved over time), and agent capability evolution (where agents can learn and adapt their behaviors). A2A focuses specifically on agent-to-agent messaging and coordination, providing communication primitives but not addressing tools or environments directly, and lacks any version or evolution management. MCP handles tool and resource integration for LLMs, standardizing how models invoke tools and access resources, but treats tools as static components without lifecycle or version management, making it impossible to support tool evolution or prompt refinement workflows.
B.2 Agent & System Features
Agent First-Class: First-class support signifies that agents are modeled as independent, managed protocol components with their own semantic schemas, state metadata, and lifecycle mechanisms. TEA’s ACP provides full first-class status to agents, capturing their roles, competencies, and objectives within a unified schema that enables seamless registration, discovery, and orchestration. A2A provides only partial support; although it enables communication, it treats agents more as opaque RPC endpoints with service-level identifiers rather than semantically rich components with managed internal states. MCP does not define agents as protocol components at all, focusing instead on model-to-tool interactions, thereby overlooking the agent as a primary unit of orchestration and management.
Multi-Agent: Multi-agent support refers to mechanisms for coordinating multiple agents in collaborative, competitive, or hierarchical configurations. TEA’s ACP formalizes multi-agent dynamics through structured relationship representations, supporting hierarchical organization (where high-level agents coordinate low-level agents), cooperative configurations (where agents collaborate toward shared goals), and competitive scenarios (where agents may have conflicting objectives). A2A enables call-level agent interactions, allowing agents to invoke each other as services, but lacks structured collaboration patterns, or negotiation mechanisms. MCP does not address multi-agent scenarios at all, as it focuses on model-tool interactions rather than agent coordination.
Tracer: Tracer refers to mechanisms for recording and tracking the complete execution process of agents, capturing detailed execution traces, decision points, tool invocations, state transitions, and intermediate results throughout task execution. TEA provides comprehensive tracing capabilities through its tracer system, which meticulously records the agent execution process for each task, enabling persistent task tracking, progress monitoring, error handling, and post-execution analysis. This allows developers to understand how agents reason, act, and evolve throughout task completion, facilitating debugging, optimization, and continuous improvement of agent behaviors. A2A and MCP lack tracing mechanisms, meaning execution tracking must be implemented ad-hoc in each application, leading to inconsistent logging and difficulty in understanding agent decision-making processes and debugging complex workflows.
Memory: Memory interfaces provide mechanisms for storing, retrieving, and managing information across agent interactions and sessions. TEA provides a dedicated memory manager that coordinates different manager components (tool managers, environment managers, agent managers) through session-based management. The memory manager operates as a workflow agent that records complete execution histories, automatically determines when to summarize information, and extracts task insights to assist future task completion. Critically, the session-based management ensures that concurrent calls do not result in resource conflicts, maintaining data consistency and preventing race conditions across multiple agent interactions. This enables agents to build upon past experiences and maintain long-term knowledge while ensuring reliable concurrent access. A2A and MCP do not define memory management protocols, leaving memory concerns to be handled entirely at the application layer, which can lead to inconsistent memory management, difficulty in sharing knowledge across agents, and potential resource conflicts in concurrent scenarios.
Component Lifecycle: Component lifecycle management refers to comprehensive lifecycle tracking and management for all component types (tools, environments, and agents) throughout their operational lifetime. TEA provides unified component lifecycle management through its context protocols (TCP, ECP, ACP), handling creation, registration, state tracking, execution monitoring, and controlled decommissioning for all three component types. This enables dynamic maintenance of instance code, proper resource allocation, state coherence, and graceful termination. Critically, TEA’s lifecycle management supports self-evolution scenarios where components can be updated, refined, or replaced while maintaining operational continuity. A2A and MCP lack comprehensive lifecycle management at this level: A2A only provides basic agent communication without lifecycle tracking for tools or environments, while MCP treats tools as static resources with no lifecycle management, making it impossible to support dynamic updates or evolution.
Version Management: Version management refers to mechanisms for tracking, maintaining, and managing multiple versions of components (tools, environments, and agents) including their code, prompts, and capabilities. TEA provides comprehensive version management through lifecycle and version systems embedded in TCP, ECP, and ACP. This enables critical self-evolution scenarios: code evolution where tool and environment implementations can be versioned, updated, and maintained with backward compatibility; prompt evolution where agent prompts can be versioned, A/B tested, and incrementally improved based on performance feedback; and capability evolution where agents can maintain multiple capability versions and gradually deploy improvements. Each component maintains version metadata, change histories, and evolution trajectories, enabling rollback, comparison, and gradual deployment of improvements. This is essential for building adaptive systems that improve over time. A2A and MCP completely lack version management: A2A treats agents as static service endpoints without versioning support, while MCP treats tools as immutable resources with no version control, making it impossible to support tool evolution, prompt refinement, or adaptive capability development.
Self-Evolution Support: Self-evolution support refers to comprehensive mechanisms that enable components (tools, environments, and agents) to evolve, adapt, and improve over time. TEA provides full self-evolution support by combining component lifecycle management and version management systems, enabling components to dynamically update, refine, and evolve while maintaining operational continuity and backward compatibility. This enables critical self-evolution scenarios: tool evolution where tools can be dynamically updated, refined, or replaced while maintaining version histories; prompt evolution where agent prompts can be versioned, A/B tested, and incrementally improved based on performance feedback; and agent capability evolution where agents can learn from experiences, adapt their behaviors, and maintain multiple capability versions. The combination of lifecycle and version management enables rollback, comparison, gradual deployment, and continuous improvement workflows that are essential for building adaptive systems that improve over time. A2A and MCP completely lack self-evolution support: A2A treats agents as static service endpoints without lifecycle or versioning mechanisms, while MCP treats tools as immutable resources with no lifecycle or version management, making it impossible to support any form of evolution, refinement, or adaptive capability development.
B.3 Context & System Capabilities
Context Management: Context management refers to mechanisms for capturing, organizing, and retrieving contextual information about tools, environments, agents, and their relationships. TEA offers comprehensive context management through its three context protocols: TCP maintains tool context with embedding-based retrieval and semantic relationship modeling, ECP manages environment state and execution context, and ACP tracks agent states and coordination context. This enables intelligent tool selection, environment-aware execution, and context-aware agent orchestration. A2A provides limited context sharing between agents through message passing, but lacks structured context management or relationship modeling. MCP uses flat tool descriptions without modeling inter-tool relationships, toolkits, or contextual execution environments, making it difficult to select appropriate tools in large-scale systems.
Component Transformations: Component transformations enable components (tools, environments, and agents) to dynamically change their roles (e.g., an agent becoming a tool, or an environment becoming an agent). TEA uniquely supports six transformation types: Agent-to-Tool (A2T) encapsulates agent capabilities as reusable tools, Tool-to-Agent (T2A) designates tools as agent actuators, Environment-to-Tool (E2T) converts environment actions into toolkits, Tool-to-Environment (T2E) elevates tool sets into environment abstractions, Agent-to-Environment (A2E) encapsulates agents as interactive environments for hierarchical modeling, and Environment-to-Agent (E2A) infuses reasoning into environments. These transformations enable dynamic role reconfiguration and flexible system architectures. A2A and MCP do not support component transformations, meaning components have fixed roles that cannot be dynamically adapted to changing task requirements.
Scalability: In an open ecosystem with $n$ coordinatable resources, the fundamental difference in coordination overhead stems from the presence or absence of hierarchical component abstraction and routing mechanisms. A2A adopts a flat multi-agent peer-to-peer collaboration model, where coordination can grow quickly with system scale due to dense pairwise interactions and state alignment. MCP reduces tool integration costs through unified interfaces, but still relies on traversing a large candidate pool or explicit application-level orchestration during resource discovery and capability matching, which can make coordination grow with the number of resources. In contrast, TEA unifies agents, tools, and environments as managed components through transformations (e.g., A2E), and utilizes tree-structured indexing and hierarchical routing for resource localization and task distribution. Under hierarchical capability organization, this can reduce coordination to logarithmic-depth routing, since each decision only considers a small, context-relevant subset at each level.
B.4 General & Ecosystem
Model-Agnostic and Framework-Agnostic: Model-agnostic means protocols work with diverse LLM backends (GPT, Claude, Gemini, etc.), while framework-agnostic means they can be integrated into different application frameworks. All three protocols are designed with these properties: TEA provides a unified LLM interface at the infrastructure layer that abstracts model heterogeneity, A2A’s agent communication is independent of the underlying models, and MCP’s tool interface works with any LLM that supports function calling. This ensures broad compatibility and allows developers to choose models and frameworks based on their specific needs rather than protocol constraints.
Key Strength: This dimension highlights each protocol’s primary advantage. TEA’s strength lies in its unified integration of Tools, Environments, and Agents into a single cohesive framework, enabling seamless interoperability and dynamic transformations between component types. A2A excels at agent interoperability, providing efficient mechanisms for agents to communicate and coordinate. MCP provides robust tool standardization, making it easy to integrate diverse tools with LLMs through a consistent interface.
Open Ecosystem Support: Open ecosystem support refers to whether a protocol can independently enable a thriving ecosystem of interoperable agents, tools, and environments without requiring additional frameworks. TEA provides a complete protocol stack with all necessary components (tool management, environment abstraction, agent orchestration, transformations, context management, etc.) to support an open ecosystem where different developers can create compatible agents, tools, and environments that seamlessly interoperate. A2A and MCP provide partial ecosystem support: A2A enables agent-to-agent interoperability but lacks tool and environment management, requiring additional frameworks for complete ecosystem support; MCP enables tool integration and standardization but lacks agent coordination and environment management, also requiring additional frameworks to achieve full ecosystem capabilities.
Appendix C Details of TEA Protocol
We provide a detailed presentation of the TEA Protocol in this section, as illustrated in Figure 1. The protocol architecture is fundamentally designed around coroutine-based asynchronous execution, enabling concurrent and parallel execution across all system components. This design supports multiple execution patterns: a single agent can concurrently execute multiple independent tasks without state interference, multiple agents can coordinate on shared tasks through collaborative mechanisms, and multiple agents can operate on distinct tasks in parallel.
The TEA Protocol comprises three architectural layers: i) Basic Managers provide foundational services through six specialized managers: the model manager abstracts heterogeneous LLM backends through a unified interface, ensuring model-agnostic interoperability; the prompt manager handles prompt lifecycle management, versioning, and retrieval for agent systems; the memory manager coordinates memory operations across different component managers via session-based concurrency control, preventing resource conflicts in concurrent scenarios; the dynamic manager implements serialization and deserialization mechanisms, converting components (prompts, memory, agents, tools, environments) and their associated code into JSON representations for persistence and restoration; the version manager maintains version histories for all components, where modifications generate new versions while preserving backward compatibility, and component access by identifier retrieves the most recent version by default; and the tracer captures comprehensive execution traces, recording decision points, tool invocations, state transitions, and intermediate results for post-execution analysis and debugging. ii) Core Protocols define three context protocols: the Tool Context Protocol (TCP), Environment Context Protocol (ECP), and Agent Context Protocol (ACP), each managing their respective component types with dedicated schemas, metadata registries, and lifecycle management. iii) Protocol Transformations establish bidirectional conversion relationships among TCP, ECP, and ACP, enabling dynamic role reconfiguration and seamless resource orchestration across component boundaries.
Additionally, the protocol incorporates a Self-Evolution Module that addresses the critical requirement for adaptive agent capabilities by encapsulating evolvable components, including prompts, tool implementations, agent architectures, memory strategies, environment actions, and successful execution solutions, as differentiable variables. The module integrates textgrad optimization and self-reflection mechanisms, allowing agents to iteratively refine these components during task execution. Optimized components are automatically registered as new versions through the version manager, ensuring that subsequent tasks leverage improved capabilities while maintaining access to historical versions for comparative analysis and rollback.
C.1 Basic Managers
The Basic Managers constitute the foundation of the TEA Protocol, providing essential services that enable higher-level functionalities. These managers include:
- Model Manager provides a unified interface for diverse large language models across multiple providers (OpenAI, Anthropic, Google, OpenRouter, etc.), supporting various model types including chat/completions, responses API, embeddings, and transcriptions. The manager maintains a centralized registry of model configurations, each encapsulating provider-specific parameters, capabilities (streaming, function calling, vision), and fallback mechanisms. It abstracts provider heterogeneity through a standardized invocation interface, enabling seamless model switching and ensuring consistent interaction patterns regardless of the underlying API. The manager supports asynchronous execution, tool/function calling, structured output formats, and automatic fallback to alternative models upon failures, ensuring robust and reliable model access across the system.
- Prompt Manager manages the complete lifecycle of prompts for agents, providing comprehensive version control, template rendering, and dynamic updates. The manager maintains a centralized registry of prompt configurations, each encapsulating system prompts, agent message templates, metadata, and version histories. It supports modular template rendering with dynamic variable substitution, enabling flexible prompt composition through configurable modules. The manager implements automatic versioning where prompt updates create new versions while preserving historical versions, enabling rollback and comparative analysis. It provides asynchronous registration, retrieval, and update operations with concurrent initialization support, ensuring efficient prompt management across multiple agents. The manager integrates with the self-evolution module by exposing trainable variables within prompts, allowing optimization algorithms to refine prompt content while maintaining version consistency. Prompts are persisted as JSON configurations and can be exported as contract documents, ensuring reproducibility and documentation of prompt evolution.
- Memory Manager provides comprehensive memory support to agents, managing the complete lifecycle of memory systems through registration, initialization, and session coordination. The manager implements session-based concurrency control, where each agent task operates within isolated memory sessions identified by session IDs, agent names, and task IDs. This session isolation ensures that concurrent calls from multiple agents or tasks do not result in resource conflicts or data corruption. The manager supports event-based memory operations, allowing agents to record execution events, step information, and contextual data throughout task execution. Memory systems are registered with configurations and can be dynamically retrieved, updated, and versioned, enabling agents to maintain persistent state and learn from historical interactions while ensuring thread-safe concurrent access.
- Dynamic Manager provides runtime code execution and serialization capabilities for all components (prompts, memory, agents, tools, environments) and their associated code. The manager enables dynamic loading of Python classes and functions from source code strings, creating virtual modules in memory without requiring disk-based files. It implements intelligent code analysis to automatically detect and inject necessary imports based on symbol usage, supporting context-aware import injection for different component types. The manager provides serialization and deserialization mechanisms for parameter schemas, converting Pydantic models to JSON representations and reconstructing them when needed. This enables components and their code to be stored as JSON configurations, loaded dynamically at runtime, and shared across different execution contexts, facilitating code evolution, version management, and dynamic component instantiation.
- Version Manager provides unified version management for all component types (tools, environments, agents, prompts, memory, etc.), maintaining comprehensive version histories with metadata, descriptions, and timestamps. The manager implements semantic versioning with automatic version generation, supporting major, minor, and patch version increments based on the nature of changes. It maintains version histories for each component, tracking the evolution trajectory and enabling access to any historical version for rollback, comparison, or analysis. The manager supports version lifecycle operations including deprecation and archiving, allowing controlled phase-out of older versions while preserving historical records. Version information is persisted as JSON, and component access by name automatically retrieves the latest version by default, while explicit version specification enables precise version control. This unified versioning system ensures consistent evolution tracking across all component types and enables seamless rollback capabilities when needed.
- Tracer provides comprehensive execution tracing capabilities for recording and analyzing agent execution processes throughout task completion. The tracer maintains session-based record management, where each execution step is captured as a structured record containing observation data, tool invocations, session identifiers, task identifiers, timestamps, and unique record IDs. Records are organized by session ID, enabling isolation of execution traces across different agent sessions and tasks while supporting cross-session analysis. The tracer implements flexible query mechanisms, allowing retrieval of records by session ID, task ID, record index, or record ID, facilitating both real-time monitoring and post-execution analysis. It supports persistent storage through JSON serialization with file locking mechanisms to ensure thread-safe concurrent access, enabling execution traces to be saved, loaded, and shared across different execution contexts. The tracer captures the complete execution trajectory including decision points, state transitions, tool call sequences, and intermediate results, providing a comprehensive audit trail for debugging, performance analysis, behavior understanding, and continuous improvement of agent capabilities.
These components work together to support the coroutine-based asynchronous framework, enabling parallel execution and concurrent task handling.
C.2 Core Protocols
The TEA Protocol defines three core context protocols: the Tool Context Protocol (TCP), the Environment Context Protocol (ECP), and the Agent Context Protocol (ACP). These protocols share a unified architectural design, each implemented through two core components: a context manager and a server. The context manager serves as the central orchestrator, responsible for context engineering (maintaining contextual information and relationships between components), lifecycle management (handling component registration, versioning, state tracking, and resource allocation), and semantic retrieval (enabling efficient component discovery through vector embeddings). The server component encapsulates the context manager and exposes a unified interface, providing operations for component registration, retrieval, execution, version management, and lifecycle control to other system modules. Each protocol generates a unified contract document (similar to Anthropic’s Agent Skills (Anthropic, 2025)) that aggregates all registered components’ descriptions, providing a comprehensive overview of available tools, environments, and agents with their capabilities, parameters, and usage guidelines. This architectural pattern ensures consistent access patterns across tools, environments, and agents while maintaining separation of concerns between internal management logic and external service interfaces.
C.2.1 Tool Context Protocol
MCP (Anthropic, 2024b) is the most widely adopted tool protocol and is defined by three components: tools, prompts, and resources, corresponding respectively to model-controlled functions, user-initiated interactive templates, and client-managed data. However, despite its widespread adoption, MCP suffers from several fundamental limitations that hinder its effectiveness in complex multi-agent systems (see Table 6). First, MCP lacks context management capabilities, meaning that tool execution environments cannot be adaptively provided to agents, constraining the system’s ability to maintain coherent context across tool invocations. Second, MCP provides no version management system, preventing tools from evolving over time while maintaining backward compatibility and version history. Third, MCP lacks component lifecycle management, meaning that tools cannot be dynamically registered, updated, or retired with proper lifecycle control, limiting the system’s ability to manage tool resources effectively.
To address these limitations, we propose the Tool Context Protocol (TCP), a comprehensive framework that fundamentally extends MCP’s capabilities through several key innovations. TCP is implemented through two core components: the ToolContextManager and the TCPServer. The ToolContextManager serves as the central orchestrator for tool lifecycle management, supporting tool loading from both local registries (via the TOOL registry system) and persistent JSON configurations, enabling seamless integration of tools across different deployment scenarios. During tool registration, TCP automatically generates multiple representation formats for each tool: function-calling schemas for LLM function calling interfaces, natural language text descriptions for human-readable documentation, and structured argument schemas (Pydantic BaseModel types) for type-safe parameter validation, providing LLMs with rich semantic information for accurate parameter inference. TCP incorporates comprehensive version management, maintaining complete version history for each tool and supporting version restoration, enabling tools to evolve while preserving backward compatibility. The protocol employs a semantic retrieval mechanism that stores each tool’s description and metadata as vector embeddings using FAISS, enabling efficient similarity-based tool discovery through query–embedding comparisons. Additionally, TCP generates tool contracts that aggregate all registered tools into unified documentation, facilitating tool discovery and usage. The TCPServer provides a unified API interface that encapsulates the ToolContextManager, exposing operations for tool registration, retrieval, execution, version management, and lifecycle control, ensuring consistent tool access patterns across the system.
C.2.2 Environment Context Protocol
In reinforcement learning, frameworks such as Gym (Brockman et al., 2016) provide standardized interfaces for training and testing environments, where each environment specifies its own observation and action spaces. The core abstraction of an environment consists of two fundamental components: observation (the current state of the environment, accessible through state queries) and action (operations that agents can perform to interact with and modify the environment state). However, most existing research on general-purpose agent systems either focuses on single environments or relies on ad-hoc adaptations to independent environments, seldom addressing the need for unified environment interfaces. Recent attempts to encapsulate environments as MCP tools allow agents to interact with them, but this approach lacks mechanisms to capture inter-tool dependencies and to manage the contextual execution environments required by tools.
To overcome these limitations, we introduce the Environment Context Protocol (ECP), a comprehensive framework that establishes unified interfaces and contextual management across diverse computational environments. ECP follows a similar architecture to TCP, implemented through two core components: the EnvironmentContextManager and the ECPServer. At its core, ECP recognizes that each environment provides a set of actions that agents can invoke, where each action represents an operation that agents can perform to interact with the environment. Each environment maintains its own state (observation) accessible through state queries, while actions provide the means for agents to interact with and modify this state. Similar to TCP, ECP supports environment loading from both local registries and persistent configurations, automatically discovers and registers all actions defined within each environment, and incorporates comprehensive version management, semantic retrieval mechanisms, and contract generation. The key distinction is that ECP manages environments (which encapsulate observation and action spaces) rather than standalone tools, enabling agents to interact with computational environments through standardized action interfaces while maintaining environment state coherence.
C.2.3 Agent Context Protocol
Existing agent frameworks or protocols, such as A2A (Google, 2025), typically rely on ad-hoc strategies for defining and managing agents, where each agent is associated with specific roles, capabilities, and policies. However, despite their utility, such systems suffer from several fundamental limitations that hinder their effectiveness in complex multi-agent systems (see Table 6). First, existing frameworks lack standardized representations of agent attributes, making it difficult to systematically capture and reason about agents’ roles, competencies, and objectives, leading to poor interoperability across different agent implementations. Second, existing approaches provide insufficient means to capture and formalize inter-agent interactions, such as delegation, collaboration, or hierarchical organization, limiting the system’s ability to support structured multi-agent coordination patterns. Third, existing frameworks fail to explicitly encode the contextual relationships between agents and the environments or tools they operate with, thereby complicating consistent state maintenance and coordination in multi-agent scenarios.
To overcome these shortcomings, we introduce the Agent Context Protocol (ACP), which establishes a unified schema for registering, representing, and coordinating agents within the TEA Protocol. ACP follows a similar architecture to TCP and ECP, implemented through two core components: the AgentContextManager and the ACPServer. Similar to TCP and ECP, ACP supports agent loading from both local registries and persistent configurations, and incorporates comprehensive version management, semantic retrieval mechanisms, and contract generation. The key distinction is that ACP manages agents (autonomous components with reasoning capabilities) rather than tools or environments, enabling agents to be registered, orchestrated, and coordinated through standardized interfaces. ACP establishes a unified schema for representing agents through semantically enriched metadata that captures agents’ roles, competencies, and objectives. The protocol formalizes the modeling of inter-agent dynamics, allowing for cooperative, competitive, and hierarchical configurations through structured relationship representations. ACP enables persistent state tracking across tasks and sessions, ensuring continuity and context preservation in multi-agent interactions. By embedding contextualized descriptions of agents and their interactions, ACP facilitates flexible orchestration, adaptive collaboration, and systematic integration with TCP and ECP, laying the groundwork for scalable and extensible multi-agent architectures.
C.3 Protocol Transformations
While TCP, ECP, and ACP provide independent specifications for tools, environments, and agents, practical deployment requires interoperability across these protocols. Thus, communication mechanisms and well-defined transformation pathways are indispensable for enabling components to assume alternative roles and exchange contextual information in a principled manner. For instance, when an agent must operate as a tool within a larger workflow, an explicit agent-to-tool transformation becomes necessary. More generally, we identify six fundamental categories of protocol transformations: Agent-to-Tool (A2T), Environment-to-Tool (E2T), Agent-to-Environment (A2E), Tool-to-Environment (T2E), Tool-to-Agent (T2A), and Environment-to-Agent (E2A). Together, these transformations constitute the foundation for dynamic role reconfiguration, enabling computational components to flexibly adapt their functional scope in response to task requirements and system constraints. This design not only ensures seamless interoperability across heterogeneous contexts but also enhances the adaptability and scalability of multi-entity systems.
- Agent-to-Tool (A2T). The A2T transformation encapsulates an agent’s capabilities and reasoning into a standardized tool interface, preserving contextual awareness while enabling seamless integration with existing tool ecosystems. For example, it can instantiate a deep researcher workflow that first generates queries, then extracts insights, and finally produces summaries, thereby providing a general-purpose tool for internet-scale retrieval tasks.
- Tool-to-Agent (T2A). The T2A transformation designates tools as the operational actuators of an agent, mapping the agent’s goals or policies into parameterized tool invocations. In this view, the agent reasons at a higher level while delegating concrete execution steps to tools, ensuring alignment between the agent’s decision space and the tool’s functional constraints. For example, a data analysis agent may employ SQL tools to query structured databases, or a design agent may invoke image editing tools to implement creative modifications. This separation allows agents to focus on strategic reasoning while relying on tools as reliable execution mechanisms.
- Environment-to-Tool (E2T). The E2T transformation converts environment-specific actions and capabilities into standardized tool interfaces, enabling agents to interact with environments through consistent tool calls. It maintains environment state coherence and exposes contextual information about available actions, allowing agents to operate across heterogeneous environments without bespoke adaptations. For example, in a browser environment, actions such as Navigate, GoBack, and Click can be consolidated into a context-aware toolkit that is directly accessible to agents.
- Tool-to-Environment (T2E). The T2E transformation elevates a collection of tools into an environment abstraction, where individual tool functions are treated as actions within a coherent action space governed by shared state and contextual rules. This conversion allows agents to interact with toolkits not merely as isolated functions but as structured environments, thereby supporting sequential decision-making, context preservation, and adaptive control. For example, a software development toolkit comprising tools for code editing, compilation, and debugging can be encapsulated as a programming environment, enabling agents to plan and execute development tasks while maintaining consistent state across tool invocations.
- Agent-to-Environment (A2E). The A2E transformation encapsulates an agent as an interactive environment, exposing its decision rules, behaviors, and state dynamics as an operational context for other agents. This conversion enables agents to function not only as autonomous components but also as adaptable environments in which other agents can act, thereby supporting multi-agent training, hierarchical control, and interactive simulations. For example, in a multi-agent simulation, a market agent can be represented as an environment that provides trading rules and dynamic market responses, allowing other agents to engage in transactions and learn adaptive strategies. Similarly, in human-in-the-loop interaction, a human agent can be modeled as an environment, enabling artificial agents to interpret user feedback and constraints as contextual signals for decision-making.
- Environment-to-Agent (E2A). The E2A transformation embeds reasoning and adaptive decision-making into the state dynamics and contextual rules of an environment, thereby elevating it into an autonomous agent. In this way, the environment is no longer a passive setting for action execution but becomes an active participant capable of initiating behaviors, coordinating with other agents, and enforcing constraints. For example, in adversarial gaming scenarios, an environment that originally only defines the state and action spaces can be transformed into an opponent agent that not only formulates strategies and responds proactively to player actions but also dynamically adjusts difficulty and interaction patterns, providing a more challenging training and evaluation platform. This transformation expands the functional role of environments within agent systems and offers a more dynamic and realistic testbed for multi-agent cooperation and competition research.
These six transformation categories establish a comprehensive framework for dynamic resource orchestration within the TEA Protocol. By enabling seamless transitions between tools, environments, and agents, the protocol transformations support adaptive architectures that reconfigure functional components in response to task requirements and contextual constraints.
C.4 Self-Evolution Module
The Self-Evolution Module addresses the growing need for agent evolution capabilities in modern AI systems. This module enables agents to continuously improve their performance by optimizing various components during task execution. The module wraps evolvable components as evolvable variables, including: prompts that guide agent behavior and reasoning; tool code that implements agent capabilities; agent code that defines agent architectures and decision-making logic; memory code that manages information storage and retrieval; environment code that defines interaction spaces; and agent execution solutions that represent successful task completion strategies. The module employs two key algorithms for optimization: textgrad (Yuksekgonul et al., 2025) provides gradient-based optimization for text-based components, enabling fine-grained improvements through iterative refinement; and self-reflection enables agents to analyze their own performance, identify weaknesses, and propose improvements. When components are optimized during task execution, the optimized versions are automatically registered as new versions through the version manager, ensuring that subsequent tasks can leverage the improved components while maintaining access to previous versions for rollback and comparison. This self-evolution capability enables agents to adapt and improve over time, learning from experience and continuously refining their capabilities without manual intervention.
TextGrad. TextGrad (Yuksekgonul et al., 2025) treats a target component (e.g., a prompt template or a code snippet) as an optimizable variable and uses feedback from execution to drive iterative updates. In our setting, the feedback signal can be defined from task outcomes and trace data, such as success or failure, constraint violations, tool error messages, intermediate correctness checks, and any available scalar scores. Given a current variable state, the system first runs the component in a controlled setting and collects a run trace via the tracer. It then constructs a differentiable style supervision signal by prompting an LLM to attribute errors to specific spans of the variable and to produce gradient-like edit directions. The optimizer applies the suggested edits to obtain an updated variable, reruns a lightweight validation on held-out traces or the current task, and keeps the update only if it improves the chosen criteria. This loop repeats for a small number of iterations, after which the final variant is registered as a new component version with its lineage and associated trace.
Self-reflection. Self-reflection treats agent-associated components as optimizable variables and improves them through structured critique and revision rather than gradient-style updates. Concretely, after a run, the system summarizes the trace into a compact diagnosis that highlights failure points, missing information, incorrect assumptions, or unsafe actions, and then selects which variables to optimize based on their causal contribution to the observed failures. A reflection prompt then guides the model to propose targeted changes to the selected variables, such as rewriting a prompt instruction, refining a tool description or schema, adjusting a planning heuristic, or generating a patch to a tool implementation. Candidate changes are evaluated through re-execution under the same environment boundaries and constraints, using the tracer to verify that the revised component improves task outcomes and does not introduce new violations. Accepted changes are committed as new versions with rollback support, enabling future runs to select improved variants while preserving historical baselines.
C.5 Formalization
In this subsection, we present a formal definition of the TEA protocol and its basic properties.
**Definition 1 (TEA Protocol)**
*Let $\mathcal{T},\mathcal{E},\mathcal{A}$ denote the sets of tools, environments, and agents; let TCP/ECP/ACP be the context protocols defined in this appendix; and let $\mathcal{M}$ denote the set of basic managers, including the model manager, prompt manager, memory manager, dynamic manager, version manager, and tracer, which provide foundational services for the protocol. The TEA Protocol is defined as the tuple
$$
\mathrm{TEA}\;=\;\langle\mathrm{TCP},\,\mathrm{ECP},\,\mathrm{ACP},\,\mathcal{M},\,\mathcal{P}_{\mathrm{TEA}}\rangle,
$$
where $\mathcal{P}_{\mathrm{TEA}}$ is a family of typed transformations over $\mathcal{T}\cup\mathcal{E}\cup\mathcal{A}$
$$
\{\mathrm{A2T},\,\mathrm{E2T},\,\mathrm{T2E},\,\mathrm{T2A},\,\mathrm{A2E},\,\mathrm{E2A}\}\subseteq\mathcal{P}_{\mathrm{TEA}}
$$
that satisfy: (i) interface consistency (exposed I/O signatures remain well-typed under the target protocol), and (ii) closure/compositionality (the composition of valid transformations is again an element of $\mathcal{P}_{\mathrm{TEA}}$ whenever domains and codomains match).*
**Definition 2 (Tool)**
*A tool is defined as a tuple
$$
T=\langle n_{T},d_{T},m_{T},g_{T},\phi_{T}\rangle,
$$
where $n_{T}$ is the tool name, $d_{T}$ is the description, $m_{T}$ is the metadata dictionary, $g_{T}∈\{\mathrm{True},\mathrm{False}\}$ indicates whether the tool supports self-evolution (i.e., whether its code can be optimized during task execution), and $\phi_{T}:\mathcal{I}_{T}→\mathcal{O}_{T}$ is the functional mapping from input space $\mathcal{I}_{T}$ to output space $\mathcal{O}_{T}$ that implements the tool’s behavior.*
**Definition 3 (Tool Configuration)**
*A tool configuration is defined as
$$
\mathrm{ToolConfig}=\langle T,v_{T},C_{T},\mathcal{F}_{T}\rangle,
$$
where $T=\langle n_{T},d_{T},m_{T},g_{T},\phi_{T}\rangle$ is the tool definition, $v_{T}$ is the version string, $C_{T}$ is the source code string, and $\mathcal{F}_{T}=\{F_{\mathrm{fc},T},F_{\mathrm{text},T},F_{\mathrm{schema},T}\}$ is the set of tool representations (function-calling schema, natural language text, and structured argument schema).*
**Definition 4 (Tool Context Protocol (TCP))**
*We formalize TCP as the tuple
$$
\mathrm{TCP}=\langle\mathcal{T},\mathcal{C},\mathcal{S},\mathcal{I}\rangle,
$$
where:
- $\mathcal{T}$ is the set of registered tools, each $T∈\mathcal{T}$ defined as $\langle n_{T},d_{T},m_{T},g_{T},\phi_{T}\rangle$ and associated with a $\mathrm{ToolConfig}$ that maintains version history $\mathcal{H}_{T}:\mathbb{V}\rightharpoonup\mathrm{ToolConfig}$ (a partial function mapping version strings to configurations).
- $\mathcal{C}$ is the tool context manager that maintains state and implements all core functionalities: (i) state mappings $\rho:\mathbb{S}\rightharpoonup\mathrm{ToolConfig}$ (active registry) and $\eta:\mathbb{S}×\mathbb{V}\rightharpoonup\mathrm{ToolConfig}$ (version history), (ii) embedding service $\xi:(d_{T},m_{T})→\mathbb{R}^{d}$ with semantic retrieval via vector database, and (iii) lifecycle operations including loading from registries and code, building instances, version management, and contract generation.
- $\mathcal{S}$ is the TCP server that encapsulates $\mathcal{C}$ and exposes a unified interface, delegating all operations to the context manager while providing consistent access patterns.
- $\mathcal{I}$ is the set of interfaces exposed by $\mathcal{S}$ :
- $\mathtt{init}$ - initialize tools from registry and code, build instances, initialize vector database
- $\mathtt{register}$ - create instance, build ToolConfig, store in registry
- $\mathtt{get}$ - get tool instance by name from active registry
- $\mathtt{info}$ - get tool configuration by name from active registry
- $\mathtt{retrieve}$ - retrieve similar tools via semantic search using vector database
- $\mathtt{list}$ - list all registered tool names
- $\mathtt{update}$ - update existing tool with new implementation, generate new version
- $\mathtt{copy}$ - duplicate existing tool with optional new name and version
- $\mathtt{unregister}$ - remove tool from active registry and version history
- $\mathtt{restore}$ - restore specific historical version of tool by name and version
- $\mathtt{vars}$ - extract tool source code as Variable objects for self-evolution
- $\mathtt{setvars}$ - update tool code variables for self-evolution, generate new version
- $\mathtt{invoke}$ - execute tool by name with structured input, return ToolResponse
- $\mathtt{contract}$ - generate unified documentation by aggregating all tools’ descriptions
- $\mathtt{save}$ - serialize tool configurations and version history to JSON file
- $\mathtt{load}$ - deserialize tool configurations and version history from JSON file
Given a request $r=(\mathtt{tool\_name},\mathtt{tool\_args})$ , $\mathcal{S}$ delegates to $\mathcal{C}$ , which uses $\mathtt{get}$ to obtain the tool instance from $\rho$ using $\mathtt{tool\_name}$ , and then invokes it with $\mathtt{tool\_args}$ via the $\mathtt{invoke}$ operation, returning a ToolResponse with execution results.*
Note. TCP explicitly supports the TEA transformations A2T via an exposure operator $\iota_{A}:A\mapsto T$ and E2T via a lifting operator $\Lambda:E\mapsto(\mathcal{S}_{E},K_{E})$ .
**Definition 5 (Environment)**
*An environment is defined as a tuple
$$
E=\langle n_{E},d_{E},m_{E},g_{E},\mathcal{A}_{E},\sigma_{E},\tau_{E}\rangle,
$$
where $n_{E}$ is the environment name, $d_{E}$ is the description, $m_{E}$ is the metadata dictionary, $g_{E}∈\{\mathrm{True},\mathrm{False}\}$ indicates whether the environment supports self-evolution, $\mathcal{A}_{E}$ is the action space (a dictionary mapping action names to action configurations), $\sigma_{E}:\bot→\mathcal{S}_{E}$ is the state retrieval function that returns the current state $\mathcal{S}_{E}$ of the environment, and $\tau_{E}:\mathbb{S}×\mathcal{D}→\mathcal{O}_{a}$ is the action execution function that takes an action name and input dictionary and returns the action result.*
**Definition 6 (Environment Configuration)**
*An environment configuration is defined as
$$
\mathrm{EnvironmentConfig}=\langle E,v_{E},C_{E},\mathcal{A}_{E},R_{E}\rangle,
$$
where $E=\langle n_{E},d_{E},m_{E},g_{E},\mathcal{A}_{E},\sigma_{E},\tau_{E}\rangle$ is the environment definition, $v_{E}$ is the version string, $C_{E}$ is the source code string, $\mathcal{A}_{E}$ is the action space (dictionary of action configurations with multi-format representations), and $R_{E}$ is the rules string (generated environment rules for interaction).*
**Definition 7 (Environment Context Protocol (ECP))**
*We formalize ECP as the tuple
$$
\mathrm{ECP}=\langle\mathcal{E},\mathcal{C},\mathcal{S},\mathcal{I}\rangle,
$$
where:
- $\mathcal{E}$ is the set of registered environments, each $E∈\mathcal{E}$ defined as $\langle n_{E},d_{E},m_{E},g_{E},\mathcal{A}_{E},\sigma_{E},\tau_{E}\rangle$ and associated with an $\mathrm{EnvironmentConfig}$ that maintains version history $\mathcal{H}_{E}:\mathbb{V}\rightharpoonup\mathrm{EnvironmentConfig}$ (a partial function mapping version strings to configurations).
- $\mathcal{C}$ is the environment context manager that maintains state and implements all core functionalities: (i) state mappings $\rho:\mathbb{S}\rightharpoonup\mathrm{EnvironmentConfig}$ (active registry) and $\eta:\mathbb{S}×\mathbb{V}\rightharpoonup\mathrm{EnvironmentConfig}$ (version history), (ii) embedding service $\xi:(d_{E},m_{E},\mathcal{A}_{E})→\mathbb{R}^{d}$ with semantic retrieval via vector database, and (iii) lifecycle operations including loading from registries and code, building instances, action discovery, version management, and contract generation.
- $\mathcal{S}$ is the ECP server that encapsulates $\mathcal{C}$ and exposes a unified interface, delegating all operations to the context manager while providing consistent access patterns.
- $\mathcal{I}$ is the set of interfaces exposed by $\mathcal{S}$ :
- $\mathtt{init}$ - initialize environments from registry and code, build instances, initialize vector database
- $\mathtt{register}$ - create instance, discover actions, build EnvironmentConfig, store in registry
- $\mathtt{get}$ - get environment instance by name from active registry
- $\mathtt{info}$ - get environment configuration by name from active registry
- $\mathtt{state}$ - get current state of environment by name via get_state method
- $\mathtt{retrieve}$ - retrieve similar environments via semantic search using vector database
- $\mathtt{list}$ - list all registered environment names
- $\mathtt{update}$ - update existing environment with new implementation, generate new version
- $\mathtt{copy}$ - duplicate existing environment with optional new name and version
- $\mathtt{unregister}$ - remove environment from active registry and version history
- $\mathtt{restore}$ - restore specific historical version of environment by name and version
- $\mathtt{vars}$ - extract environment source code as Variable objects for self-evolution
- $\mathtt{setvars}$ - update environment code variables for self-evolution, generate new version
- $\mathtt{invoke}$ - execute environment action by name and action name with structured input
- $\mathtt{contract}$ - generate unified documentation by aggregating all environments’ rules
- $\mathtt{save}$ - serialize environment configurations and version history to JSON file
- $\mathtt{load}$ - deserialize environment configurations and version history from JSON file
Given a request $r=(\mathtt{env\_name},\mathtt{action\_name},\mathtt{action\_args})$ , $\mathcal{S}$ delegates to $\mathcal{C}$ , which uses $\mathtt{get}$ to obtain the environment instance from $\rho$ using $\mathtt{env\_name}$ , and then invokes the action with $\mathtt{action\_name}$ and $\mathtt{action\_args}$ via the $\mathtt{invoke}$ operation, returning an action result.*
Note. ECP explicitly supports the TEA transformations A2E via an encapsulation operator $\Omega_{A}:A\mapsto\widehat{E}$ that presents an agent as an interactive environment, and T2E via an abstraction operator $\Gamma:(\mathcal{S},K)\mapsto\widehat{E}$ that consolidates a toolkit into an environment abstraction.
**Definition 8 (Agent)**
*An agent is defined as a tuple
$$
A=\langle n_{A},d_{A},m_{A},g_{A}\rangle,
$$
where $n_{A}$ is the agent name, $d_{A}$ is the description, $m_{A}$ is the metadata dictionary, and $g_{A}∈\{\mathrm{True},\mathrm{False}\}$ indicates whether the agent supports self-evolution.*
**Definition 9 (Agent Configuration)**
*An agent configuration is defined as
$$
\mathrm{AgentConfig}=\langle A,v_{A},C_{A},\mathcal{F}_{A}\rangle,
$$
where $A=\langle n_{A},d_{A},m_{A},g_{A}\rangle$ is the agent definition, $v_{A}$ is the version string, $C_{A}$ is the source code string, and $\mathcal{F}_{A}$ is the set of agent representations (function-calling schemas, natural language descriptions, and Pydantic BaseModel argument schemas).*
**Definition 10 (Agent Context Protocol (ACP))**
*We formalize ACP as the tuple
$$
\mathrm{ACP}=\langle\mathcal{A},\mathcal{C},\mathcal{S},\mathcal{I}\rangle,
$$
where:
- $\mathcal{A}$ is the set of registered agents, each $A∈\mathcal{A}$ defined as $\langle n_{A},d_{A},m_{A},g_{A}\rangle$ and associated with an $\mathrm{AgentConfig}$ that maintains version history $\mathcal{H}_{A}:\mathbb{V}\rightharpoonup\mathrm{AgentConfig}$ (a partial function mapping version strings to configurations).
- $\mathcal{C}$ is the agent context manager that maintains state and implements all core functionalities: (i) state mappings $\rho:\mathbb{S}\rightharpoonup\mathrm{AgentConfig}$ (active registry) and $\eta:\mathbb{S}×\mathbb{V}\rightharpoonup\mathrm{AgentConfig}$ (version history), (ii) embedding service $\xi:(d_{A},m_{A})→\mathbb{R}^{d}$ with semantic retrieval via vector database, and (iii) lifecycle operations including loading from registries and code, building instances, version management, and contract generation.
- $\mathcal{S}$ is the ACP server that encapsulates $\mathcal{C}$ and exposes a unified interface, delegating all operations to the context manager while providing consistent access patterns.
- $\mathcal{I}$ is the set of interfaces exposed by $\mathcal{S}$ :
- $\mathtt{init}$ - initialize agents from registry and code, build instances, initialize vector database
- $\mathtt{register}$ - create instance, build AgentConfig, store in registry
- $\mathtt{get}$ - get agent instance by name from active registry
- $\mathtt{info}$ - get agent configuration by name from active registry
- $\mathtt{retrieve}$ - retrieve similar agents via semantic search using vector database
- $\mathtt{list}$ - list all registered agent names
- $\mathtt{update}$ - update existing agent with new implementation, generate new version
- $\mathtt{copy}$ - duplicate existing agent with optional new name and version
- $\mathtt{unregister}$ - remove agent from active registry and version history
- $\mathtt{restore}$ - restore specific historical version of agent by name and version
- $\mathtt{vars}$ - extract agent source code as Variable objects for self-evolution
- $\mathtt{setvars}$ - update agent code variables for self-evolution, generate new version
- $\mathtt{invoke}$ - execute agent method by name with structured input, return agent response
- $\mathtt{contract}$ - generate unified documentation by aggregating all agents’ descriptions
- $\mathtt{save}$ - serialize agent configurations and version history to JSON file
- $\mathtt{load}$ - deserialize agent configurations and version history from JSON file
Given a request $r=(\mathtt{agent\_name},\mathtt{input})$ , $\mathcal{S}$ delegates to $\mathcal{C}$ , which uses $\mathtt{get}$ to obtain the agent instance from $\rho$ using $\mathtt{agent\_name}$ , and then invokes it with $\mathtt{input}$ via the $\mathtt{invoke}$ operation, returning an agent response.*
Note. ACP explicitly supports the TEA transformations T2A via a designation operator $\kappa_{T}:T\mapsto\widehat{A}$ and E2A via an elevation operator $\Psi_{E}:\widehat{E}\mapsto\widehat{A}$ that embeds reasoning/decision capabilities into an environment to obtain an agent abstraction.
Appendix D The AgentOrchestra Implemented by TEA Protocol
AgentOrchestra is a concrete instantiation and implementation of the TEA Protocol, demonstrating how the protocol’s core principles and transformations can be applied to build a practical hierarchical multi-agent system. This section first introduces the fundamental design principles that govern agent behavior and interaction within the framework, including the definitions of key components such as agents, environments, models, memory, observations, and actions. We then present the specific agents and tools that constitute AgentOrchestra, including the planning agent for task decomposition and coordination, the deep researcher agent for comprehensive information gathering, the deep analyzer agent for complex reasoning tasks, the browser use agent for automated web interaction, the tool generator agent for intelligent tool evolution and management, and the reporter agent for automated report generation and citation management.
D.1 Agent Design Principles
Agent. An agent is an autonomous computational entity that perceives and interprets the environment, maintains a history of actions and observations, and flexibly generates actions to accomplish a wide variety of user-specified tasks across diverse domains. Within the TEA Protocol framework, agents are managed through the ACP, which provides standardized registration, representation, and coordination mechanisms.
Environment. The environment represents the external context and resources within which the agent operates, providing the interface for action execution and information access. Within the TEA Protocol framework, environments are managed through the ECP, which provides unified inputs, outputs, and environment rules across multiple environments.
Model. LLMs are the core drivers of this framework, providing the reasoning and decision-making capabilities for agents. Within the TEA Protocol framework, models are managed through the Infrastructure Layer, which provides a unified interface for diverse LLMs. This design enables agents to dynamically select and switch between different LLMs during task execution, aligning each model’s unique strengths with specific requirements.
Memory. Memory serves as a fundamental component of the agent, persistently recording the complete history of agent execution. Within the TEA Protocol framework, memory is managed through the Infrastructure Layer as a workflow agent that operates based on sessions, automatically recording agent execution paths across multiple tasks. This memory system automatically determines when to summarize and extract task insights to assist in task completion.
Observation. An observation primarily consists of the task description, attached files, the agent’s execution history, the environment state, and the set of available tools and sub-agents, providing the agent with a comprehensive view of the ongoing process.
Action. In our framework, actions are managed under the Tool Context Protocol (TCP) and executed through a set of pre-defined tools Wang et al. (2024b); Liang et al. (2025); Roucher et al. (2025) exposed via function-calling interfaces OpenAI (2023); Anthropic (2024b). Actions are not equivalent to tools. A single tool can support multiple actions by accepting different parameters. For example, a planning tool may support create, update and delete through a unified interface.
An agent operates in a perception–interpretation–action cycle. It observes the environment and stores information in memory, interprets context with the unified LLMs interface, and determines an action. The action is executed in a sandbox, with results recorded back to memory to refine reasoning and adaptation. This loop continues until objectives are achieved or a termination condition is met.
D.2 Planning Agent
The planning agent serves as the central orchestrator in our hierarchical framework, dedicated to high-level reasoning, task decomposition, and adaptive planning. The planning agent utilizes structured thinking and unified invocation mechanisms to coordinate specialized sub-agents and tools for complex task completion. As illustrated in Figure 4, the planning agent implements a systematic iterative workflow that integrates structured reasoning, context management, and execution coordination with continuous monitoring and adaptive adjustments.
<details>
<summary>x6.png Details</summary>

### Visual Description
## Diagram: Planning Agent Workflow
### Overview
The image is a diagram illustrating the workflow of a Planning Agent. It outlines the process from receiving a task to executing it, including steps for interpreting, decomposing, and assigning sub-tasks. The diagram also incorporates feedback loops and considerations for objective shifts and unexpected errors.
### Components/Axes
* **Title:** Planning Agent
* **Elements:**
* Task (Left side, vertical rectangle)
* Todo (Top section)
* Actions (Table within Todo)
* add: Add a new plan (Pink)
* clear: Delete the plan (Red)
* update: Update the plan (Purple)
* complete: Mark step as completed (Orange)
* Pipeline (Middle section)
* Interpret user tasks (Rounded rectangle with target icon)
* Decompose into manageable sub-tasks (Rounded rectangle with stacked box icon)
* Assign to specialized sub-agents (Rounded rectangle with assign icon)
* Sub-agent A (Light blue)
* Sub-agent B (Light green)
* Tool C (Light orange)
* Feedback (Double arrow between "Assign" and sub-agents)
* Execute (Top-right, blue rounded rectangle)
* Create, update, and manage plans for complex tasks simultaneously
* Track execution states
* Objective Shifts (Update Plans) (Bottom-left rounded rectangle)
* Unexpected Errors (Bottom-right rounded rectangle)
* **Flow:** The primary flow is from left to right through the "Pipeline." Arrows indicate the direction of the process.
* **Connections:**
* Task -> Interpret user tasks (Yellow arrow)
* Interpret user tasks -> Decompose into manageable sub-tasks (Black arrow)
* Decompose into manageable sub-tasks -> Assign to specialized sub-agents (Black arrow)
* Assign to specialized sub-agents <-> sub-agents (Feedback loop)
* Interpret user tasks -> Objective Shifts (Yellow arrow)
* sub-agents -> Unexpected Errors (Black arrow)
* Objective Shifts & Unexpected Errors -> Assign to specialized sub-agents (Black arrow)
### Detailed Analysis or Content Details
The diagram presents a structured approach to task management. The "Todo" section outlines actions related to planning, while the "Pipeline" details the steps involved in processing a task. The "Execute" section describes the actions performed on the plans.
* **Task:** Represents the initial input or goal.
* **Todo:** Manages the planning actions. The actions are:
* Add a new plan (Pink)
* Delete the plan (Red)
* Update the plan (Purple)
* Mark step as completed (Orange)
* **Pipeline:**
1. **Interpret user tasks:** Understands the task requirements.
2. **Decompose into manageable sub-tasks:** Breaks down the task into smaller, actionable units.
3. **Assign to specialized sub-agents:** Distributes sub-tasks to appropriate agents or tools.
4. **Feedback:** Allows for iterative adjustments based on the performance of sub-agents.
* **Execute:** Involves creating, updating, and managing plans, as well as tracking execution states.
* **Objective Shifts & Unexpected Errors:** Represents potential changes or issues that may arise during the process, requiring adjustments to the plan.
### Key Observations
* The diagram emphasizes a modular and iterative approach to task management.
* Feedback loops and error handling are integrated into the workflow.
* The use of specialized sub-agents suggests a distributed or parallel processing architecture.
### Interpretation
The diagram illustrates a sophisticated planning and execution system. The "Planning Agent" is designed to handle complex tasks by breaking them down into smaller, manageable units and assigning them to specialized agents. The inclusion of feedback loops and error handling mechanisms ensures adaptability and robustness. The "Todo" section provides a clear set of actions for managing plans, while the "Pipeline" outlines the steps involved in processing a task. The diagram suggests a system capable of handling dynamic and unpredictable environments.
</details>
Figure 4: Planning Agent Workflow.
Structured Reasoning. The planning agent employs a structured thinking framework that guides each execution step, capturing reasoning processes, evaluation of previous goals, memory insights, next objectives, and tool/agent selections. This structured approach ensures systematic reasoning, explicit progress tracking, and transparent decision-making. The agent dynamically builds a unified interface that combines sub-agents from ACP and tools from TCP (including those transformed from environments via E2T and from agents via A2T), enabling seamless coordination of both specialized agents and domain-specific tools within a single execution framework.
Pipeline Workflow. The planning agent implements a systematic pipeline for task processing and execution that can be conceptually divided into four main stages. The pipeline begins with task interpretation, where the agent analyzes incoming user requests to extract objectives, constraints, and contextual requirements. This is followed by task decomposition, wherein complex objectives are systematically broken down into smaller, executable sub-tasks that can be processed by specialized components. The third stage involves resource allocation, where sub-tasks are strategically assigned to appropriate specialized agents or tools based on their domain expertise and functional capabilities. Finally, the execution and coordination stage manages the task execution, incorporating continuous feedback mechanisms that enable dynamic plan adjustments and inter-agent coordination throughout the process. The implementation incorporates session management for maintaining context across multiple interactions, memory storage and retrieval systems for learning from past experiences, and execution tracking for observability and debugging.
Adaptive Planning and Error Handling. The planning agent incorporates robust mechanisms for handling dynamic changes and unexpected situations. When objective shifts occur, the system updates plans accordingly, triggering a return to the task interpretation phase to reassess and modify the approach. Similarly, when unexpected errors arise during execution, the agent re-evaluates the task and adjusts the plan to address the issues. This adaptive capability ensures that the system can maintain progress even when encountering unforeseen challenges or changing requirements.
The planning agent’s design emphasizes modularity and scalability, interacting with sub-agents through the ACP and utilizing tools from the TCP, thereby concealing domain-specific details and facilitating the integration of new agent types and resources. This architecture enables the agent to maintain a global perspective throughout the execution process, aggregating feedback from sub-agents and monitoring progress toward the overall objective, while performing dynamic plan updates in real-time in response to intermediate results, unexpected challenges, or shifting user requirements.
D.3 Deep Researcher Agent
The deep researcher agent is a specialized component designed for comprehensive information gathering through multi-round research workflows with multimodal capabilities. As illustrated in Figure 5, the agent implements a systematic pipeline workflow for research execution that begins with task analysis and query generation, followed by multi-engine web search across various platforms, result evaluation and completeness assessment, and iterative refinement through follow-up queries until comprehensive information is gathered. The agent leverages the Reporter Agent (detailed in Section D.7) to generate structured research reports with proper citations and references.
<details>
<summary>x7.png Details</summary>

### Visual Description
## Diagram: Deep Researcher Agent Workflow
### Overview
The image is a diagram illustrating the workflow of a "Deep Researcher Agent." It outlines the process from task initiation to summarizing insights, incorporating search engines and iterative refinement.
### Components/Axes
* **Title:** Deep Researcher Agent
* **Task:** A box on the left side, representing the initial task input.
* **Search:** A section containing search engines.
* **Engines:**
* baidu (purple box)
* bing (light purple box)
* firecrawl (pink box)
* google (orange box)
* **Execute:** A rounded rectangle in the top-right corner.
* Iteratively gather insights across multiple rounds (with a checkmark)
* Obtain and summarize the final search results (with a checkmark)
* **Pipeline:** A sequence of steps forming the core research process.
* Analyze: generate fitting query (magnifying glass icon)
* Search: query-based web search (Google "G" icon)
* Insight: extract insights about task (lightbulb icon)
* Summarize: summarize the insights (stacked papers icon)
* **Next Round (Update Query):** A box at the bottom, indicating a loop back to the "Analyze" step.
* **Check Results:** A box at the bottom, connected to both "Next Round" and "Summarize".
* **Connectors:** Arrows indicating the flow of information and control.
### Detailed Analysis
The workflow begins with a "Task." The "Search" section uses various search engines (baidu, bing, firecrawl, google) to gather information. The "Pipeline" then processes this information through the following steps:
1. **Analyze:** The initial task is analyzed to generate a fitting query.
2. **Search:** A query-based web search is performed using the generated query.
3. **Insight:** Insights are extracted from the search results.
4. **Summarize:** The extracted insights are summarized.
The process can loop back to the "Analyze" step via "Next Round (Update Query)," allowing for iterative refinement of the search query. "Check Results" is connected to both "Next Round" and "Summarize," suggesting that results are checked before either updating the query or finalizing the summary.
### Key Observations
* The diagram emphasizes an iterative approach to research, with the ability to refine the search query based on intermediate results.
* Multiple search engines are used, suggesting a comprehensive search strategy.
* The "Execute" section highlights the iterative gathering of insights and the final summarization of search results.
### Interpretation
The diagram illustrates a sophisticated research process that combines automated search with iterative refinement. The "Deep Researcher Agent" leverages multiple search engines and a structured pipeline to extract and summarize insights from web-based information. The iterative nature of the process allows the agent to adapt its search strategy based on the results obtained, potentially leading to more relevant and comprehensive findings. The "Check Results" step implies a quality control mechanism, ensuring the accuracy and reliability of the extracted insights.
</details>
Figure 5: Deep Researcher Agent Workflow.
Search Engines. The deep researcher agent integrates multiple search engines to ensure comprehensive coverage and information diversity. The system supports six primary search engines: Baidu for Chinese-language content, Bing, Brave and DuckDuckGoSearch for general web search, Firecrawl for comprehensive web crawling and content extraction with full webpage content retrieval, and Google for comprehensive global search. Additionally, the agent can utilize specialized LLM-based search models for enhanced information retrieval. This multi-engine approach enables the agent to access diverse information sources and overcome limitations of individual search platforms, ensuring robust information retrieval across different domains and languages.
Pipeline Workflow. The core pipeline implements a systematic multi-stage process for research execution. The workflow begins with task analysis and query generation, where the agent generates optimized search queries based on the research objectives, contextual requirements, and previous search history. This initial analysis transforms vague research requests into specific, actionable search queries that can effectively target relevant information sources. This is followed by parallel web search, wherein the agent performs targeted searches across multiple engines and LLM-based search models simultaneously using the generated queries. The multi-engine approach is essential because different search platforms have varying coverage, indexing strategies, and content biases, ensuring comprehensive information retrieval while mitigating the limitations of individual search engines. The third stage involves result merging and evaluation, where the agent consolidates search results from multiple sources and evaluates whether the gathered information provides a complete answer to the research task. This evaluation step is necessary because it determines whether additional research rounds are needed or if sufficient information has been collected. Finally, the report generation stage uses the Reporter Agent to consolidate all research rounds into a structured markdown report with proper citations and references, and generates a comprehensive summary from the final report content.
Iterative Research Process. The deep researcher agent incorporates a sophisticated iterative mechanism for comprehensive research. After each round’s evaluation, the system checks whether the gathered information provides a complete answer. When additional research is required, the agent enters the next round, where it updates and refines search queries based on previous findings and identified knowledge gaps. Each round’s content, including queries, search results, and evaluations, is systematically added to the Reporter Agent, which maintains proper citation tracking throughout the research process. This iterative process continues until a complete answer is found or predefined research limits (maximum rounds) are reached. Upon completion, the Reporter Agent generates a final structured report with all citations properly numbered and referenced, ensuring not only comprehensive coverage of complex research topics but also proper source attribution and balanced control over exploration depth, efficiency, and resource consumption.
The deep researcher agent’s design emphasizes adaptability and comprehensiveness, enabling it to handle diverse research tasks ranging from factual inquiries to complex analytical investigations. The multimodal support allows the agent to process both textual and visual information simultaneously, while the iterative workflow ensures that research quality improves through multiple rounds of refinement and validation. The integration with the Reporter Agent ensures that all research findings are properly documented with citations, making the research process transparent and verifiable.
D.4 Deep Analyzer Agent
The deep analyzer agent is a specialized component designed for complex reasoning tasks involving diverse data sources through a workflow-oriented approach with multimodal data support. As illustrated in Figure 6, the agent implements a systematic pipeline workflow for complex reasoning and analysis that begins with file classification and validation, followed by overall file summary assessment, type-specific analysis strategies, and iterative multi-round refinement until answers are found or analysis limits are reached. The agent leverages the Reporter Agent (detailed in Section D.7) to generate structured analysis reports with proper citations and references.
<details>
<summary>x8.png Details</summary>

### Visual Description
## Diagram: Deep Analyzer Agent Workflow
### Overview
The image is a diagram illustrating the workflow of a "Deep Analyzer Agent." It outlines the process from input files to summarized insights, including steps for file conversion, task enhancement, insight extraction, and summarization. The diagram also shows how the process can loop back for further analysis.
### Components/Axes
* **Title:** Deep Analyzer Agent
* **Files:** This section describes the types of files the agent can process.
* `image`: Caption Image (color: purple-red)
* `audio`: Transcribe Audio (color: purple)
* `text`: Read the text file (color: red-orange)
* `zip`: Read the contents of zip (color: orange)
* **Pipeline:** This section outlines the main processing steps.
* `Mdify`: convert files to markdown (includes a markdown icon)
* `Enhance`: get the enhanced task (includes a magic wand icon)
* `Insight`: extract insights about task (includes a lightbulb icon)
* `Summarize`: summarize the insights (includes a stack of papers icon)
* **Execute:** This section describes the actions performed.
* Generate an enhanced task from the input text and files (includes a checkmark icon)
* Obtain and summarize the final analysis results (includes a checkmark icon)
* **Task&Files:** Input to the process.
* **Next Round (Task & Summaries):** Output and feedback loop.
* **Check Results:** Feedback loop component.
### Detailed Analysis or ### Content Details
1. **Input Files:**
* The agent accepts four types of files: images, audio, text, and zip archives.
* Images are used for captioning.
* Audio files are transcribed.
* Text files are read directly.
* Zip archives are unpacked and their contents are read.
2. **Pipeline Stages:**
* **Mdify:** Converts input files to markdown format. An arrow points downward from the "M" in "Mdify," suggesting a conversion or transformation process.
* **Enhance:** Enhances the task based on the converted files.
* **Insight:** Extracts insights from the enhanced task.
* **Summarize:** Summarizes the extracted insights.
3. **Execution:**
* The agent generates an enhanced task from the input text and files.
* The agent obtains and summarizes the final analysis results.
4. **Feedback Loop:**
* The "Next Round (Task & Summaries)" component represents the output of the pipeline.
* An arrow connects this component back to the "Task&Files" input, indicating a feedback loop for further analysis.
* The "Check Results" component is connected to both the "Summarize" stage and the "Next Round" component, suggesting a quality control or validation step.
### Key Observations
* The diagram illustrates a sequential workflow with a feedback loop.
* The "Mdify" stage appears to be a crucial step for converting various file types into a unified format (markdown).
* The "Execute" section highlights the core functionalities of the agent: task generation and result summarization.
### Interpretation
The diagram describes a system designed to analyze various types of input files, extract meaningful insights, and provide a summary of the findings. The feedback loop suggests an iterative process where the agent can refine its analysis based on previous results. The "Mdify" stage indicates that the system uses markdown as an intermediate format, which likely simplifies subsequent processing steps. The overall architecture suggests a sophisticated tool for automated data analysis and summarization.
</details>
Figure 6: Deep Analyzer Agent Workflow.
File Classification and Support. The deep analyzer agent supports comprehensive file formats including text files, PDFs, images, audio, video, and compressed archives, with support for both local files and URLs. The system employs LLM-based file type classification to determine the appropriate analysis strategy for each file. For URLs, the system automatically detects file types based on URL patterns and extensions, while for local files, it uses both LLM classification and extension-based fallback mechanisms. This classification stage is essential because different file types require different analysis approaches: text and PDF files benefit from chunk-based markdown analysis, images and audio require direct multimodal LLM analysis, and videos may need both direct analysis and transcript-based processing.
Pipeline Workflow. The core pipeline implements a systematic multi-stage process for complex reasoning and analysis. The workflow begins with file validation and classification, where the system validates file accessibility and size constraints, then classifies each file by type (text, PDF, image, audio, video) to determine appropriate analysis strategies. This is followed by overall file summary, where the agent generates a preliminary summary based on file metadata (sizes, types, timestamps) to determine if the task can be answered from file information alone, enabling early termination when sufficient information is available. The third stage involves type-specific analysis, where the agent processes each file according to its type: text files are converted to markdown and analyzed in chunks; PDF files first attempt direct LLM analysis, then fall back to markdown conversion and chunk-based analysis if needed; images first attempt direct multimodal LLM analysis, then proceed to multi-step analysis if the answer is not found; audio files are analyzed directly through multimodal LLM; and video files first attempt direct LLM analysis, then convert to markdown transcripts for chunk-based analysis if needed. Each analysis step checks whether the answer has been found, enabling early stopping when sufficient information is obtained. Finally, the report generation stage uses the Reporter Agent to consolidate all analysis rounds into a structured markdown report with proper citations and references, and generates comprehensive summaries from the final report content.
Iterative Multi-Round Analysis. The deep analyzer agent incorporates a sophisticated iterative mechanism for comprehensive analysis refinement. The system executes multiple analysis rounds, with each round processing all files according to their classified types. After each round, the system synthesizes summaries from all file analyses and evaluates whether a complete answer has been found. When additional analysis is required, the agent enters the next round, where it continues processing files with refined strategies based on previous findings. Each round’s content, including file classifications, analysis results, and answer evaluations, is systematically added to the Reporter Agent, which maintains proper citation tracking throughout the analysis process. This iterative process continues until a complete answer is found or predefined analysis limits (maximum rounds) are reached. Upon completion, the Reporter Agent generates a final structured report with all citations properly numbered and referenced, ensuring not only comprehensive coverage of complex reasoning tasks but also proper source attribution and balanced control over analysis depth, efficiency, and resource consumption.
Task-Only Analysis. When no files are provided, the deep analyzer agent can directly analyze tasks such as text games, math problems, logic puzzles, or reasoning challenges. The system performs multi-round analysis where each round applies step-by-step reasoning, breaks down task components, identifies key information and constraints, and generates insights and partial solutions. This capability enables the agent to handle diverse reasoning tasks that do not require file-based information, making it a versatile tool for both file-based and file-free analysis scenarios.
The deep analyzer agent’s design emphasizes workflow-oriented processing and multimodal data support, enabling it to handle diverse reasoning tasks ranging from document analysis to complex multi-step problem solving. The adaptive file type handling ensures optimal analysis strategies for each data source, while the iterative workflow guarantees that analysis quality improves through multiple rounds of refinement and validation. The integration with the Reporter Agent ensures that all analysis findings are properly documented with citations, making the analysis process transparent and verifiable.
D.5 Browser Use Agent
The browser use agent is a specialized component designed for automated web interaction and task completion through the browser_use library. As illustrated in Figure 7, the agent implements a systematic workflow for web interaction and task execution that begins with task initialization and report setup, followed by browser agent execution with intelligent web navigation and interaction, result extraction and evaluation, and comprehensive report generation with execution records. The agent leverages the Reporter Agent (detailed in Section D.7) to generate structured browser task reports with proper documentation.
<details>
<summary>x9.png Details</summary>

### Visual Description
## Diagram: Browser Use Agent
### Overview
The image is a diagram illustrating the architecture and workflow of a Browser Use Agent. It outlines the interaction between the agent, the browser/computer, and the task it's designed to perform. The diagram includes components for defining actions, a pipeline for executing tasks, and mechanisms for updating goals and checking results.
### Components/Axes
* **Title:** Browser Use Agent
* **Task:** A rounded rectangle on the left side labeled "Task".
* **Browser&Computer:** A rectangle encompassing the "Actions" and "Pipeline" sections.
* **Actions:** A section within "Browser&Computer" that defines possible actions:
* **goto:** Go to the URL (colored purple)
* **input:** Input a text (colored light purple)
* **scroll:** Scroll down or up (colored light red)
* **click:** Click a button or position (colored light orange)
* **Pipeline:** A sequence of steps for task execution:
* **Prepare:** prepare browser environment (hourglass icon)
* **Generate:** generate next actions list (arrow icon)
* **Execute:** execute the actions list (cursor icon)
* **Evaluate:** check the answer (checkmark icon)
* **Record:** record execution state (document icon)
* **Execute (Blue Box):** A blue rounded rectangle containing:
* Iteratively generate, execute, and summarize actions
* Generate next goal until task completion
* **Next Step (Update Next Goal):** A rounded rectangle at the bottom.
* **Check Results:** A rounded rectangle at the bottom.
### Detailed Analysis or ### Content Details
* **Task to Prepare:** A yellow arrow connects the "Task" box to the "Prepare" step in the "Pipeline".
* **Prepare to Generate:** A black arrow connects the "Prepare" step to the "Generate" step.
* **Generate to Execute:** A black arrow connects the "Generate" step to the "Execute" step.
* **Execute to Evaluate:** A black arrow connects the "Execute" step to the "Evaluate" step.
* **Evaluate to Record:** A black arrow connects the "Evaluate" step to the "Record" step.
* **Record to Generate:** A black arrow connects the "Record" step back to the "Generate" step, forming a loop.
* **Record to Next Step:** A black arrow connects the "Record" step to the "Next Step (Update Next Goal)" box.
* **Prepare to Next Step:** A yellow arrow connects the "Prepare" step to the "Next Step (Update Next Goal)" box.
* **Next Step & Check Results:** The "Next Step (Update Next Goal)" box is connected to the "Check Results" box with an "&" symbol.
* **Check Results to Record:** A black arrow connects the "Check Results" box to the "Record" step.
### Key Observations
* The diagram illustrates a cyclical process where the agent prepares the environment, generates actions, executes them, evaluates the results, records the state, and then generates new actions based on the recorded state.
* The "Next Step (Update Next Goal)" and "Check Results" components provide feedback and control mechanisms for the agent.
* The "Actions" section defines the basic operations the agent can perform within the browser.
### Interpretation
The diagram presents a high-level overview of a Browser Use Agent's architecture and workflow. The agent operates in a loop, continuously generating, executing, and evaluating actions to achieve a given task. The "Next Step" and "Check Results" components suggest that the agent can adapt its strategy based on the outcome of previous actions. The separation of "Actions" from the "Pipeline" suggests a modular design where the agent's capabilities can be extended by adding new actions. The diagram highlights the key components and their interactions, providing a clear understanding of how the agent operates within a browser environment.
</details>
Figure 7: Browser Use Agent Workflow.
Browser Agent Integration. The browser use agent leverages the browser_use library, which provides an intelligent browser automation framework with LLM-driven decision-making capabilities. The system integrates ChatOpenAI as the underlying language model for both task planning and page content extraction, enabling the agent to understand web page structures, generate appropriate actions, and extract relevant information. The browser agent supports comprehensive web interactions including URL navigation, form filling, element clicking, scrolling, and content extraction. The integration addresses the complexity of modern web applications by providing semantic understanding of page content and intelligent action selection, enabling the agent to handle dynamic web pages, JavaScript-rendered content, and complex user interfaces that require contextual understanding.
Pipeline Workflow. The core pipeline implements a systematic multi-stage process for web interaction and task execution. The workflow begins with task initialization and report setup, where the agent initializes a Report instance to track the browser task execution, records the task description, and prepares for result documentation. This initialization stage is essential because it establishes a structured framework for capturing execution details, enabling comprehensive documentation and post-execution analysis. This is followed by browser agent execution, wherein the browser_use Agent is instantiated with the specified task and LLM configuration, then executes the task through intelligent web navigation and interaction. The browser agent operates with a maximum step limit (typically 50 steps) to ensure task completion within reasonable bounds, and employs sophisticated page understanding mechanisms to extract content and generate appropriate actions. During execution, the agent generates visual execution records (GIF animations) and conversation logs, providing detailed traces of the interaction process. The third stage involves result extraction, where the agent extracts the final results from the browser agent’s execution history. The system attempts multiple extraction strategies: first checking for extracted content summaries, then falling back to final results, and finally extracting from the last step’s action results if available. This multi-strategy approach ensures robust result extraction even when the browser agent’s output format varies. Finally, the report generation stage uses the Reporter Agent to consolidate the task description and execution results into a structured markdown report with proper formatting. The report includes the original task, execution results, and references to generated execution records (GIF files and logs), ensuring comprehensive documentation of the browser interaction process.
Concurrent Execution Support. The browser use agent incorporates robust mechanisms for handling concurrent task execution. Each browser task execution is assigned a unique call identifier (call_id), which is used to create isolated subdirectories for execution artifacts, preventing file conflicts when multiple browser tasks run simultaneously. The system generates unique paths for GIF animations, conversation logs, and report files based on the call_id, ensuring that concurrent executions do not interfere with each other. This concurrent execution support is essential for multi-agent scenarios where multiple browser tasks may be initiated simultaneously, enabling scalable and reliable browser automation in distributed agent systems.
Execution Record Generation. The browser use agent automatically generates comprehensive execution records during task execution. The system creates visual execution traces in GIF format, capturing the sequence of browser interactions and page states throughout the task execution. Additionally, the agent saves detailed conversation logs that record all LLM interactions, action decisions, and page content extractions. These execution records provide valuable debugging information, enable post-execution analysis, and support transparency in browser automation tasks. The records are organized in per-call subdirectories, making it easy to trace specific task executions and analyze browser interaction patterns.
The browser use agent’s design emphasizes intelligent web automation and comprehensive documentation, enabling it to handle diverse web-based tasks ranging from simple information retrieval to complex multi-step interactions. The integration with browser_use library provides sophisticated web understanding capabilities, while the Reporter Agent ensures that all browser interactions are properly documented with execution traces, making the automation process transparent and verifiable.
D.6 Tool Generator Agent
The tool generator agent is a specialized component designed for intelligent tool evolution through automated creation, dynamic retrieval, and systematic reuse mechanisms under the TCP. As illustrated in Figure 8, the agent implements a systematic pipeline workflow for intelligent tool lifecycle management that begins with task analysis and tool retrieval, followed by tool creation and evaluation, and tool registration in TCP.
<details>
<summary>x10.png Details</summary>

### Visual Description
## Diagram: Tool Generator Agent Workflow
### Overview
The image is a diagram illustrating the workflow of a Tool Generator Agent. It depicts the process from analyzing user tasks to executing or discarding a generated tool. The workflow is divided into two main sections: TCP (Tools) and Pipeline.
### Components/Axes
* **Title:** Tool Generator Agent
* **Sections:**
* **TCP (Tools):** Contains available tools.
* **Pipeline:** Describes the workflow stages.
* **Task:** Input to the pipeline.
* **Tools (TCP):**
* `read`: Read an URL (Pink)
* `query`: Query a wiki page (Orange-Red)
* `code`: Run a code (Light Purple)
* `locate`: Locate a position (Orange)
* **Pipeline Stages:**
* Analyze user tasks (Target Icon)
* Retrieve TCP tools retrieval (Magnifying Glass Icon)
* Create generate tool code (Wrench Icon)
* Evaluate tool evaluation (Chat Bubble Icon)
* Discard discard the tool (Exclamation Point Icon)
* **Execute:**
* Get tool context information (Blue)
* Get parameters and execute tool (Blue)
* **Arrows:** Indicate the flow of the process.
* Yellow arrows indicate the main flow and reuse.
* Black arrows indicate registration and feedback loops.
* **Decision Points:**
* "Yes" and "No" labels indicate branching based on evaluation.
### Detailed Analysis
1. **Task Input:** The process begins with a "Task" input on the left.
2. **Pipeline Flow:**
* The task is first analyzed ("Analyze user tasks").
* Then, relevant tools are retrieved ("Retrieve TCP tools retrieval").
* Tool code is generated ("Create generate tool code").
* The generated tool is evaluated ("Evaluate tool evaluation").
3. **Decision Point:**
* If the tool evaluation is negative ("No"), the process loops back to "Retrieve TCP tools retrieval".
* If the tool evaluation is positive ("Yes"), the tool is registered.
4. **Tool Reuse:**
* Tools from the TCP section can be reused. The `read`, `query`, and `locate` tools can be reused.
5. **Execution:**
* The selected tool is executed ("Execute"), which involves getting tool context information and executing the tool with parameters.
6. **Discard:**
* If the tool is not suitable, it is discarded ("Discard discard the tool").
### Key Observations
* The diagram illustrates a cyclical process of tool generation, evaluation, and refinement.
* The TCP section provides a set of tools that can be used in the pipeline.
* The "Yes/No" decision point is crucial for determining whether a tool is registered or discarded.
* The "Reuse" arrow indicates that existing tools can be incorporated into the process.
### Interpretation
The diagram represents a system for automatically generating and evaluating tools. The pipeline aims to create tools that effectively address user tasks. The feedback loop ensures that only suitable tools are registered and used. The TCP section acts as a repository of reusable tools, promoting efficiency and consistency. The process emphasizes iterative refinement and quality control in tool generation.
</details>
Figure 8: Tool Generator Agent Workflow.
Pipeline Workflow. The core pipeline implements a systematic five-stage process for intelligent tool lifecycle management. The workflow begins with task analysis, where the agent analyzes task requirements and extracts tool specifications including tool name, description, parameter schema, and implementation plan. This is followed by tool retrieval, wherein the agent uses TCP’s semantic search to retrieve similar tools from the registry. If suitable existing tools are found, the agent evaluates their compatibility and returns the best match. The third stage involves tool creation, where the agent generates new tool implementations using LLM-based code generation when no suitable existing tools are found. The generated code follows the Tool base class pattern and includes proper error handling and logging. The fourth stage is tool evaluation, where the agent validates newly created tools by loading the tool class, checking for required attributes (name, description, __call__ method), and verifying structural correctness. Tools that fail evaluation are discarded, while successfully validated tools proceed to registration. Finally, the tool registration stage registers validated tools in TCP, which automatically handles version management, contract generation, and persistence to JSON manifests, making the tools immediately available to all agents through the unified TCP interface.
TCP Integration. The tool generator agent leverages TCP to provide comprehensive tool management capabilities. Through TCP’s semantic retrieval mechanism, the agent can efficiently search for existing tools based on functional similarity, avoiding redundant tool creation. When new tools are generated, TCP’s registration process automatically handles version tracking, contract documentation, and persistence, ensuring that all tools are properly managed and accessible across the multi-agent system. This TCP-based approach enables seamless tool sharing and reuse, supporting both local tool execution and distributed tool access through standardized interfaces.
The tool generator agent’s design emphasizes TCP-based tool management, enabling it to handle diverse tool requirements ranging from simple utility functions to complex domain-specific operations. The intelligent evolution process guarantees that the tool ecosystem continuously adapts to emerging requirements through systematic creation, validation, and reuse mechanisms.
D.7 Reporter Agent
The Reporter Agent is a specialized component designed for managing and generating structured markdown reports with proper citation and reference management. As illustrated in Figure 9, the agent implements a systematic workflow for report management that encompasses two primary phases: (i) the Pipeline for incremental content addition and processing, and (ii) the Report generation process with automated citation management. The agent is widely used by other tools (deep researcher, deep analyzer, browser) to document their execution processes and findings.
<details>
<summary>x11.png Details</summary>

### Visual Description
## Diagram: Reporter Agent Workflow
### Overview
The image is a diagram illustrating the workflow of a "Reporter Agent." It outlines the process from task initiation to report generation, detailing the actions, pipeline stages, and execution steps involved.
### Components/Axes
* **Title:** Reporter Agent
* **Sections:**
* Task (left side)
* Report (top section)
* Pipeline (middle section)
* Next Round (Task & Summaries) & Check Results (bottom section)
* **Report Actions Legend:** Located in the top-left of the "Report" section.
* `add` (pink): Add content
* `cite` (red): Add citations
* `ref` (purple): Extract ref
* `export` (orange): Generate report
* **Execute Box:** Located in the top-right of the "Report" section.
* Add content to list (with a checkmark icon)
* If check completed, generate report (with a checkmark icon)
* **Pipeline Stages:**
* Add: add research contexts (magnifying glass icon)
* Extract: extract references (bar graph icon)
* cite: cite the references (document icon)
* Export: generate report (upload icon)
### Detailed Analysis
* **Task:** A vertical box on the left side labeled "Task."
* **Report:** Contains "Actions" and an "Execute" box. The "Actions" section lists the different actions that can be performed, each associated with a specific color. The "Execute" box contains two tasks with checkmarks.
* **Pipeline:** A horizontal sequence of steps:
1. **Add:** "add research contexts" with a magnifying glass icon.
2. **Extract:** "extract references" with a bar graph icon.
3. **cite:** "cite the references" with a document icon.
4. **Export:** "generate report" with an upload icon.
* **Flow:**
* The "Task" box has a yellow arrow pointing to the "Add" stage in the "Pipeline."
* The "Add" stage in the "Pipeline" has a yellow arrow pointing to the "Next Round (Task & Summaries)" box.
* The "Pipeline" stages are connected sequentially with black arrows.
* The "Export" stage in the "Pipeline" has a black arrow pointing to the "Check Results" box.
* The "Next Round (Task & Summaries)" box is connected to the "Check Results" box with an "&" symbol.
* The "Check Results" box has a black arrow pointing back to the "Add" stage in the "Pipeline."
### Key Observations
* The diagram illustrates a cyclical workflow, where the output of the "Pipeline" ("Check Results") feeds back into the "Add" stage for the next round.
* The "Report" section seems to represent a higher-level set of actions, while the "Pipeline" breaks down the report generation process into more granular steps.
* The "Execute" box in the "Report" section suggests a conditional execution based on whether a check is completed.
### Interpretation
The diagram describes the workflow of a "Reporter Agent," likely a software or automated system designed to generate reports. The process begins with a "Task," which initiates the "Pipeline." The "Pipeline" consists of stages for adding research contexts, extracting references, citing references, and exporting the report. The "Next Round" and "Check Results" components suggest an iterative process where the results of one round inform the next. The "Report" section provides a higher-level view of the actions involved, while the "Execute" box indicates a conditional execution based on completion checks. The overall workflow is cyclical, allowing for continuous refinement and improvement of the generated reports.
</details>
Figure 9: Reporter Agent Workflow.
Pipeline Workflow. The Reporter Agent supports incremental content addition through the add action, which accepts content from multiple sources including text strings, dictionaries, and file paths. When content is added, the agent employs LLM-based extraction to automatically identify and structure three key components: (i) content, the main text preserving all citation markers in markdown link format [1](url), [2](url), etc.; (ii) summary, a concise 2-3 sentence summary of the content; and (iii) references, a list of reference items with IDs, descriptions, and URLs extracted from citations in the content. This automatic extraction ensures that citations are properly captured and linked to their sources, enabling systematic reference management throughout the report generation process.
Reference Management and Deduplication. The Reporter Agent implements sophisticated reference management mechanisms to ensure citation consistency and accuracy. When the complete action is invoked, the agent performs comprehensive reference processing: (i) collection, gathering all references from all content items; (ii) deduplication, merging duplicate references based on normalized keys (URLs are prioritized over descriptions for deduplication); (iii) renumbering, creating a unified reference mapping that renumbers all citations sequentially from 1; and (iv) URL generation, automatically generating proper URLs for references (converting file paths to file:// URLs, preserving HTTP/HTTPS URLs, and extracting URLs from descriptions when needed). This reference management ensures that all citations in the final report are properly numbered, deduplicated, and linked to their sources.
Generation and Completion. The final report generation process consolidates all content items into a coherent, well-structured markdown document. The agent uses LLM-based generation to organize content logically, integrate summaries for smooth transitions, and maintain proper citation formatting throughout the report. The generated report includes a complete References section at the end, listing all references in numerical order with proper URLs and descriptions. The agent ensures that all citations maintain the markdown link format [number](url) and that file paths are converted to absolute paths for proper rendering in markdown viewers. The final report is written to the specified file path with file locking mechanisms to ensure concurrent safety when multiple processes access the same report.
Integration with Other Tools. The Reporter Agent is designed to be seamlessly integrated with other tools through a unified interface. Tools such as deep researcher, deep analyzer, and browser use the Reporter Agent to document their execution processes, with each tool adding content items incrementally and completing the report when execution finishes. The agent supports per-call caching and locking mechanisms, enabling multiple concurrent report generations without conflicts. This integration ensures that all tools can generate comprehensive, properly cited reports that document their findings and execution traces, making the entire system’s operations transparent and verifiable.
The Reporter Agent’s design emphasizes automatic citation management and structured report generation, enabling other tools to produce professional, well-documented reports without manual citation formatting. The LLM-based extraction and generation capabilities ensure that citations are properly identified, deduplicated, and formatted, while the reference management system guarantees consistency and accuracy across complex multi-source reports.
Appendix E Detailed Analysis of Benchmark Results
E.1 GAIA Benchmark
As shown in Figure ˜ 3 and Table ˜ 4, AgentOrchestra achieves state-of-the-art performance on GAIA (89.04% average) by mitigating the dimensionality curse and semantic drift that arise in large-scale agentic planning. We attribute this success to three complementary properties enabled by TEA and AgentOrchestra. First, hierarchical decoupling of the action space reduces planning complexity: hierarchical routing decomposes the global task into locally tractable sub-problems and assigns them to domain-appropriate sub-agents, preserving abstract reasoning under long horizons even amid low-level browser events. Second, ECP formalizes epistemic environment boundaries: GAIA’s multi-domain tasks require temporal and cross-modal state coherence, and baselines often degrade during domain transitions such as browser retrieval followed by local Python analysis. By treating environments as first-class managed components, TEA preserves and propagates session-critical state (e.g., authentication tokens, downloaded files, and transient file-system mutations) across agent boundaries, reducing contextual forgetting. Third, AgentOrchestra supports recursive refinement of reasoning trajectories and can invoke the Tool Generator to synthesize context-specific functionalities on demand, alleviating the fixed-capability bottleneck of static agent components.
Qualitatively, the Tool Generator is effective on tasks involving structured sources (e.g., Wikipedia or tabular data), where it can synthesize query wrappers and extraction utilities with clear I/O contracts. However, we observe limitations on fine-grained visual tasks (e.g., localizing specific colored digits or subtle UI elements), suggesting that tool synthesis alone cannot replace robust visual grounding and that tighter integration with vision-centric models remains necessary.
Across train and test, the Tool Generator produced over 50 TCP-registered tools spanning multiple domains, and approximately 30% were reused in subsequent tasks. This reuse rate indicates a practical balance between specialization and generalization, where the system expands capability coverage while retaining transferable utilities for recurring sub-problems.
On GAIA validation, AgentOrchestra achieves 92.45% on Level 1, 83.72% on Level 2, and 57.69% on Level 3 (82.42% average), consistently outperforming strong baselines such as AWorld (77.58%) and Langfun Agent (76.97%). Notably, the gap between Level 1 and Level 3 highlights the challenge of long-horizon, multi-domain tasks, where TEA-style environment state management and on-demand tool synthesis become increasingly important.
The key strength of AgentOrchestra lies in decomposing complex problems and flexibly assigning them to appropriate specialists. For example, in a Level 3 GAIA scenario that required extracting numerical data from an embedded table within a PDF and then performing multi-step calculations, the Planning Agent invoked the Browser Use Agent to locate and download the file, delegated parsing and verification to the Deep Analyzer, and then synthesized the final answer. When existing tools were inadequate, the Tool Generator created task-specific utilities (e.g., custom extractors for particular document layouts or scripts for bespoke computations), improving coverage and reliability. We note that frequent inter-agent exchanges can introduce latency and overhead; thus, the system minimizes unnecessary switching and motivates future work on adaptive routing and resource selection for improved efficiency and scalability.
E.2 SimpleQA Benchmark
As shown in Table ˜ 3, AgentOrchestra achieves state-of-the-art performance on SimpleQA with 95.3% accuracy, substantially outperforming monolithic LLM baselines (e.g., o3 at 49.4% and gemini-2.5-pro-preview-05-06 at 50.8%) and surpassing strong retrieval agents such as Perplexity Deep Research (93.9%). We attribute this improvement to systematic reduction of epistemic uncertainty through our hierarchical verification pipeline. Hallucinations in short-form factuality often arise from conflicting web-based evidence or reliance on internal parametric memory. AgentOrchestra mitigates these issues by enforcing cross-agent consensus: the Planning Agent orchestrates a retrieve-verify-synthesize cycle where the Deep Researcher performs multi-engine breadth-first searches while the Deep Analyzer evaluates evidence consistency across heterogeneous sources. By decoupling retrieval from analysis, the system prevents the confirmation bias inherent in single-agent architectures. Furthermore, the integration with the Reporter Agent ensures traceable attribution, grounding every factual claim in a re-verified source.
E.3 HLE Benchmark
On HLE, AgentOrchestra achieves 37.46%, outperforming strong baselines including o3 (20.3%) and Perplexity Deep Research (21.1%). This gain highlights the framework’s capacity for long-horizon analytical reasoning and adaptive capability expansion in expert-level domains. HLE demands synthesizing disparate, highly specialized knowledge beyond simple retrieval. We attribute the gain to hierarchical coordination and adaptive capability expansion. The Planning Agent maintains global objective coherence, prunes the hypothesis space via delegation to specialists, and assigns technical validation to agents such as the Deep Analyzer. As a result, the final solution is both analytically rigorous and cross-verified against multimodal evidence, yielding robust performance on challenging expert-level tasks.
E.4 Ablation Studies and Efficiency Analysis
Sub-agent Contribution Analysis. As detailed in Table 4, we evaluate the incremental contribution of each specialized sub-agent (Planning, Researcher, Browser, Analyzer, and Tool Generator). The synergy between these components is most evident in network-dependent tasks. While the baseline Planning agent (P) achieves 36.54% accuracy, the integration of coarse-grained retrieval via the Deep Researcher (R) and fine-grained interaction via the Browser (B) elevates performance to 72.76%. This doubling of efficacy underscores the complementarity of high-level information gathering and low-level DOM manipulation. Furthermore, the Deep Analyzer (A) provides an 8.67% improvement by resolving complex multi-step reasoning within documents and media, while the Tool Generator (T) adds a final 12.61% boost. This last leap validates that static toolkits are insufficient for the diverse requirements of GAIA, necessitating on-demand tool synthesis to bridge functional gaps.
System Efficiency and Resource Consumption. We analyze the operational efficiency of AgentOrchestra across varying task complexities, measured by wall-clock latency and token throughput. Simple factual queries typically complete within 30 seconds (approx. 5k tokens), while medium-complexity research tasks average 3 minutes (approx. 25k tokens). High-complexity multimodal or long-horizon scenarios require approximately 10 minutes and 100k tokens. Compared to monolithic baselines, our hierarchical architecture optimizes resource allocation by invoking resource-intensive sub-agents only when necessary. This ensures that AgentOrchestra maintains operational costs comparable to commercial research agents while delivering significant performance gains on complex tasks.
Self-Evolution Effectiveness. To evaluate the TEA Protocol’s support for iterative optimization, we assess self-evolution mechanisms on mathematical reasoning benchmarks (GPQA-Diamond, AIME24/25). As summarized in Table 5, iterative refinement via gradient-based (TextGrad) and symbolic (self-reflection) approaches consistently improves reasoning performance. For instance, using gpt-4.1 with self-reflection boosts AIME25 accuracy from 20.00% to 33.34%, highlighting the benefits of recursive trajectory refinement. By leveraging execution feedback via TEA’s versioning and tracer mechanisms, the system identifies and corrects logical inconsistencies in its planning. Overall, this shifts reasoning from one-shot inference to a managed optimization process, enabling AgentOrchestra to evolve robust problem-solving strategies for frontier-level tasks.
Appendix F Case Studies
In this section, we systematically present representative cases of AgentOrchestra, accompanied by critical analyses to elucidate the underlying factors contributing to these outcomes. We primarily showcase the performance on the GAIA validation set, categorized by both difficulty Level 1, Level 2, and Level 3 and data type, including text, image, audio, video, spreadsheet, ZIP archive, and other file types.
Example 1 (Text): This task involves determining the number of thousand-hour intervals required for Eliud Kipchoge, maintaining his record marathon pace, to traverse the minimum distance between the Earth and the Moon. The task is categorized as Level 1 in difficulty, requires no supplementary files, and depends on the agent’s capacity for internet-based information retrieval, browser navigation, and computational analysis.
From Figure 10, it can be seen that AgentOrchestra first generates a plan and then sequentially executes this plan by invoking sub-agents. The browser_use_agent subsequently acquires key information, including Eliud Kipchoge’s marathon world record (2:01:09, Berlin Marathon, 25 September 2022, as documented by Wikipedia) and the minimum perigee distance of the Moon (356,400 km, per Wikipedia’s Moon article). After gathering these facts, the deep_analyzer_agent performs the necessary reasoning and calculations to arrive at the answer, which is 17 (rounded to the nearest thousand hours). Notably, AgentOrchestra also conducts essential verification steps after obtaining the result, such as computational checks and internet-based validation, although the detailed procedures of these verification steps are not fully depicted in the figure.
<details>
<summary>x12.png Details</summary>

### Visual Description
## Flow Diagram: Task Execution
### Overview
The image is a flow diagram illustrating the steps taken by a system to answer a question. The system uses a combination of agents (Planning Agent, Browser Use Agent, Deep Analyzer Agent) to gather information and perform calculations, ultimately arriving at a final answer.
### Components/Axes
* **Task ID:** e1fc63a2-da7a-432f-be78-7c4a95598703. Level 1
* **Question:** If Eliud Kipchoge could maintain his record-making marathon pace indefinitely, how many thousand hours would it take him to run the distance between the Earth and the Moon at its closest approach? Please use the minimum perigee value on the Wikipedia page for the Moon when carrying out your calculation. Round your result to the nearest 1000 hours and do not use any comma separators if necessary.
* **Attached File:** Null
* **Planning Agent:**
* **Plan:** Calculate time for Kipchoge to run Earth-Moon distance at marathon pace (ID: kipchoge\_moon\_calculation)
* **Progress:** 0/5 steps completed (0.0%)
* **Status:** 0 completed, 0 in progress, 0 blocked, 5 not started
* **Steps:**
1. \[ \] Use browser\_use\_agent to find Eliud Kipchoge's marathon world record time and pace from relevant web pages
2. \[ \] Use browser\_use\_agent to find the minimum perigee distance between Earth and Moon from Wikipedia page for the Moon
3. \[ \] Use deep\_analyzer\_agent to calculate how long it would take Kipchoge to run the Earth-Moon distance at his record pace
4. \[ \] Verify calculations and convert result to thousands of hours, rounded to nearest 1000
5. \[ \] Provide final answer without comma separators
* **Browser Use Agent (Left):**
* **Task:** Find Eliud Kipchoge's marathon world record time and pace according to Wikipedia page.
* **Multi-turn Interactions With Browser:** (Screenshot of a webpage)
* **Result:** Eliud Kipchoge's marathon world record (as per Wikipedia): Time: 2:01:09 (Berlin Marathon, 25 September 2022)
* **Browser Use Agent (Right):**
* **Task:** Find the minimum perigee distance of the Moon as stated on this Wikipedia page.
* **Multi-turn Interactions With Browser:** (Screenshot of a webpage)
* **Result:** Minimum perigee distance of the Moon (per Wikipedia's Moon article): Minimum perigee distance: 356,400 km
* **Deep Analyzer Agent:**
* **Task:** If Eliud Kipchoge could... if necessary. Given data found by browser agent:
* Kipchoge's marathon world record: 2:01:09 for 42.195 km marathon distance
* His pace: 2:52 min/km
* Minimum perigee distance Earth to Moon: 356,400 km
* Please verify and perform the detailed calculations step by step.
* **Result:** To determine how many thousand hours it would take Eliud Kipchoge to run the distance between the Earth and the Moon at its closest approach, we need to follow these steps: ... Final Calculation Check: Speed = 42.195 / (1 + 1/60 + 9/3600) = 42.195 / (1.01866666666) = 41.326267 km/h Distance = 356400 km Time = 356400 / 41.32626740947075 = 15844.4858859 hours Rounding 15844.4858859 hours to the nearest 1000 hours gives 16000 hours This is 16 thousand hours. The number of thousand hours is 16. Final Answer: "16"
* **Final Answer:** 16
### Detailed Analysis or ### Content Details
The diagram illustrates a step-by-step process:
1. The **Planning Agent** outlines the steps required to answer the question.
2. Two instances of the **Browser Use Agent** are used to extract information from web pages: one to find Eliud Kipchoge's marathon record and another to find the minimum perigee distance between the Earth and the Moon.
3. The **Deep Analyzer Agent** takes the extracted information and performs calculations to determine the time it would take Kipchoge to run the Earth-Moon distance.
4. The final answer is rounded to the nearest 1000 hours, resulting in "16".
### Key Observations
* The system relies on multiple agents to perform different tasks.
* The Browser Use Agent is used to extract information from web pages.
* The Deep Analyzer Agent performs calculations based on the extracted information.
* The final answer is rounded to the nearest 1000 hours.
### Interpretation
The diagram demonstrates a system's ability to answer a complex question by breaking it down into smaller tasks and using specialized agents to perform each task. The system effectively combines information retrieval and calculation to arrive at a final answer. The rounding of the final answer suggests a tolerance for approximation in the context of the question. The use of multiple agents highlights a modular approach to problem-solving.
</details>
Figure 10: Execution trajectory of AgentOrchestra for Example 1.
Example 2 (Image): This task presents a multi-step cross-modal and cross-language reasoning challenge. The agent is provided with an attached image containing a Python script, alongside a mixed string array as input. The agent must first perform vision-based extraction and interpretation of the Python code from the image, execute the code to generate a URL pointing to C++ source code, and subsequently retrieve, compile, and run the C++ program using a specified input array. The final answer is derived by reasoning over the program’s output. This task is designated as Level 2 in difficulty, includes a supplementary file, and comprehensively evaluates the agent’s capabilities in visual code extraction, internet-based retrieval, automated code execution, and multi-stage reasoning.
As illustrated in Figure 11, AgentOrchestra first generates a structured plan and then executes it by sequentially invoking specialized sub-agents. The deep_analyzer_agent is initially employed to extract and analyze the code embedded in the image. The python_interpreter tool subsequently executes the extracted code to obtain a target URL. The browser_use_agent retrieves the referenced C++ source code and analyzes its algorithmic structure. Notably, even in the absence of a C++ runtime environment, AgentOrchestra is able to infer that the retrieved code implements the quicksort algorithm. Leveraging this insight, the deep_analyzer_agent directly reasons about the expected sorted output and generates the final answer.
<details>
<summary>x13.png Details</summary>

### Visual Description
## Flow Diagram: Task Execution with Multiple Agents
### Overview
The image depicts a flow diagram illustrating a task being executed by multiple agents. The task involves using a Python script to generate a URL, retrieving C++ code from that URL, compiling and running the C++ code against an array of numbers, and returning the sum of the third and fifth integers in the sorted list. The diagram shows the sequential steps and the agents responsible for each step.
### Components/Axes
The diagram consists of the following components:
1. **Task ID Box (Top-Left)**: Contains the task ID and level.
2. **Planning Agent Box (Bottom-Left)**: Describes the plan and steps involved.
3. **Deep Analyzer Agent Boxes (Two, Left-Center and Right-Center)**: Executes Python scripts and C++ code.
4. **Browser Use Agent Box (Center)**: Retrieves C++ source code from a URL.
5. **Final Answer Box (Right)**: Contains the final answer.
6. **Arrows**: Indicate the flow of information between the agents.
### Detailed Analysis or ### Content Details
**1. Task ID Box (Top-Left)**:
* **Task ID**: b7f857e4-d8aa-4387-af2a-0e844df5b9d8
* **Level**: 2
* **Question**: The attached image contains a Python script. Run the Python code against an array of strings, listed below. The output of the Python script will be a URL containing C++ source code. Compile and run this C++ code against the array [35, 12, 8, 99, 21, 5] and return the sum of the third and fifth integers in the sorted list.
* **Array**: \['ok', 'glg', 'C++', 'J++', 'tps', 'XQ', 'pqr', 'sta', 'V', 'Yose', 'Vwr', 'yz1', '234', 'ta', '567', '867', 'a', 'or', 'q', 'W&W', 'Y', 'ing', 'kot', 'abc', 'sor', 'ir', 'hms', 'VQ', 'ue', 'kset', 'W', 'h'\]
* **Attached File**: b7f857e4-d8aa-4387-af2a-0e844df5b9d8.jpg
**2. Planning Agent Box (Bottom-Left)**:
* **Plan**: Execute Python Scripts, Generate C++ Code and Solve Array Problem (ID: python\_cpp\_task\_plan)
* **Progress**: 0/5 steps completed (0.0%)
* **Status**: 0 completed, 0 in progress, 0 blocked, 5 not started
* **Steps**:
* 0. \[ \] Use deep\_analyzer\_agent to analyze the attached image file 'b7f857e4-d8aa-4387-af2a-0e844df5b9d8.png' to extract the Python script code.
* 1. \[ \] Execute the extracted Python script against the provided array of strings to get a URL containing C++ source code.
* 2. \[ \] Use browser\_use\_agent to access the URL generated by the Python script and retrieve the C++ source code.
* 3. \[ \] Use deep\_analyzer\_agent to compile and run the C++ code against the array [35, 12, 8, 99, 21, 5] to get the sorted list.
* 4. \[ \] Calculate the sum of the third and fifth integers in the sorted list and provide the final answer.
**3. Deep Analyzer Agent Box 1 (Left-Center)**:
* **Task**: Execute the extracted Python script against the provided array of strings to get a URL.
* **Result**:
* Joining these strings: "ok" + "glg" + "C++" + "J++" + "tps" + "XQ" + "pqr" + "sta" + "V" + "Yose" + "Vwr" + "yz1" + "234" + "ta" + "567" + "867" + "a" + "or" + "q" + "W&W" + "Y" + "ing" + "kot" + "abc" + "sor" + "ir" + "hms" + "VQ" + "ue" + "kset" + "W" + "h" + "C++"
* Result: https://rosettacode.org/wiki/Sorting\_algorithms/QuickSortC++
* Now, prepend the 'archive\_prefix': url = "https://web.archive.org/web/20230609112831/"
* https://rosettacode.org/wiki/Sorting\_algorithms/QuickSortC++
* So the final URL is: https://web.archive.org/web/20230609112831/https://rosettacode.org/wiki/Sorting\_algorithms/QuickSortC++
**4. Browser Use Agent Box (Center)**:
* **Task**: https://web.archive.org/web/20230609112831/https://rosettacode.org/wiki/Sorting\_algorithms/QuickSortC++ and retrieve the C++ source code for the Quicksort algorithm. I need to extract the complete C++ code that can be compiled and executed to sort an array.
* **Multi-turn Interactions With Browser**: [Screenshot of a webpage]
* **Result**: The task is to compile and run "this" C++ code (or rather, the sorting logic it represents) against the array [35, 12, 8, 99, 21, 5]. We don't need to actually compile it if we can determine the sorting behavior. Quicksort sorts an array. The C++ code on Rosetta Code for Quicksort will sort the input array in ascending order.
**5. Deep Analyzer Agent Box 2 (Right-Center)**:
* **Task**: Compile and run the C++ code against the array [35, 12, 8, 99, 21, 5] to get the sorted list.
* **Result**:
* The input array for the C++ code is [35, 12, 8, 99, 21, 5].
* Sorting this array in ascending order:
* Original: [35, 12, 8, 99, 21, 5]
* Sorted: [5, 8, 12, 21, 35, 99]
* The sorted list is [5, 8, 12, 21, 35, 99].
* Array indexing is 0-based.
* 1st integer (index 0): 5
* 2nd integer (index 1): 8
* 3rd integer (index 2): 12
* 4th integer (index 3): 21
* 5th integer (index 4): 35
* 6th integer (index 5): 99
* The third integer is 12.
* The fifth integer is 35.
* The sum of the third and fifth integers is 12 + 35.
* Sum = 47.
**6. Final Answer Box (Right)**:
* 47
### Key Observations
* The task involves a sequence of steps executed by different agents.
* The Python script generates a URL that leads to C++ code for the Quicksort algorithm.
* The C++ code is used to sort an array of numbers.
* The final answer is the sum of the third and fifth integers in the sorted array.
### Interpretation
The diagram illustrates a complex task that requires the coordination of multiple agents. The task involves using a Python script to generate a URL, retrieving C++ code from that URL, compiling and running the C++ code against an array of numbers, and returning the sum of the third and fifth integers in the sorted list. The diagram shows the sequential steps and the agents responsible for each step. The use of multiple agents allows for the task to be broken down into smaller, more manageable steps. The final answer is 47, which is the sum of the third (12) and fifth (35) integers in the sorted array [5, 8, 12, 21, 35, 99].
</details>
Figure 11: Execution trajectory of AgentOrchestra for Example 2.
Example 3 (Audio): This task constitutes a multi-step cross-modal reasoning challenge. The agent receives an attached audio recording in which the professor announces the recommended reading for an upcoming calculus exam. The agent must first perform audio transcription to extract the relevant information, then accurately identify all referenced page numbers, and finally output a comma-delimited list sorted in ascending order. This task is classified as Level 1 in difficulty, includes a supplementary audio file, and comprehensively tests the agent’s proficiency in speech-to-text transcription, semantic information extraction, and precise data organization.
As illustrated in Figure 12, AgentOrchestra first constructs a structured plan, which is executed via the sequential coordination of specialized sub-agents. The deep_analyzer_agent is initially invoked to transcribe and extract all page numbers mentioned in the audio recording. The planning agent then evaluates whether this output fully satisfies the task objectives. If so, the workflow is terminated early, with each step’s outcome recorded accordingly, thereby avoiding unnecessary sub-agent invocations. Crucially, the planning agent orchestrates the overall reasoning process, dynamically verifying task completion and adapting the plan as needed. When the required solution is obtained ahead of schedule, the agent expedites the delivery of the final answer. Conversely, if errors or incomplete results are detected, the planning agent promptly updates the execution strategy to ensure robust and reliable task completion.
<details>
<summary>x14.png Details</summary>

### Visual Description
## Workflow Diagram: Audio Analysis and Page Number Extraction
### Overview
The image depicts a workflow diagram outlining the process of extracting page numbers from an audio recording related to a calculus professor's recommendations. The workflow starts with a task description, proceeds through a "Deep Analyzer Agent," and then involves multiple iterations of a "Planning Agent" before arriving at the "Final Answer."
### Components/Axes
* **Task ID:** 1f975693-876d-457b-a649-393859e79bf3. Level 1
* This is the starting point, containing the initial question and attached file.
* **Deep Analyzer Agent:**
* Analyzes the attached audio file to extract page numbers.
* **Planning Agent:**
* Iteratively refines the page number extraction process. There are multiple instances of this agent shown.
* **Final Answer:**
* The final extracted page numbers.
### Detailed Analysis or ### Content Details
1. **Task Description (Top-Left):**
* **Task ID:** 1f975693-876d-457b-a649-393859e79bf3. Level 1
* **Question:** "Hi, I was out sick from my classes on Friday, so I'm trying to figure out what I need to study for my Calculus mid-term next week. My friend from class sent me an audio recording of Professor Willowbrook giving out the recommended reading for the test, but my headphones are broken so I couldn't you please listen to the recording for me and tell me the page numbers I'm supposed to go over? I've attached a file called Homework.mp3 that has the recording. Please provide just the page numbers as a comma-delimited list. And please provide the list in ascending order."
* **Attached File:** 1f975693-876d-457b-a649-393859e79bf3.mp3
* A microphone icon is present.
2. **Deep Analyzer Agent (Top-Center):**
* **Task:** "Analyze the attached audio file '1f975693-876d-457b-a649-393859e79bf3.mp3' to extract page numbers mentioned by Professor Willowbrook for the Calculus mid-term."
* **Result:** "Okay, I've 'listened' to the recording for you. Based on Professor Willowbrook's recommendations, here are the page numbers you should go over: 132, 133, 134, 197, 245"
3. **Planning Agent (Bottom-Left):**
* **Plan:** "Extract Page Numbers from Calculus Professor Audio Recording (ID: calculus\_audio\_analysis)"
* **Progress:** "0/5 steps completed (0.0%)"
* **Status:** "0 completed, 1 in progress, 0 blocked, 4 not started"
* **Steps:**
* "0. \[+\] Use deep\_analyzer\_agent to analyze the attached audio file '1f975693-876d-457b-a649-393859e79bf3.mp3' to extract page numbers mentioned by Professor Willowbrook for the Calculus mid-term."
* "1. \[ \] Parse and identify all page numbers mentioned in the audio recording"
* "2. \[ \] Sort the page numbers in ascending order"
* "3. \[ \] Format the results as a comma-delimited list"
* "4. \[ \] Provide the final answer with the page numbers in the required format"
4. **Planning Agent (Top-Right):**
* **Plan:** "Extract Page Numbers from Calculus Professor Audio Recording (ID: calculus\_audio\_analysis)"
* **Progress:** "1/5 steps completed"
* **Status:** "1 completed"
* **Steps:**
* "0. \[✓] Use deep\_analyzer\_agent to analyze the attached audio file '1f975693-876d-457b-a649-393859e79bf3.mp3' to extract page numbers mentioned by Professor Willowbrook for the Calculus mid-term."
* "1. \[ \] Parse and identify all page numbers mentioned in the audio recording"
* "2. \[ \] Sort the page numbers in ascending order"
* "3. \[ \] Format the results as a comma-delimited list"
* "4. \[ \] Provide the final answer with the page numbers in the required format"
* **Notes:** "Successfully analyzed audio file and received the page numbers: 132, 133, 134, 197, 245"
5. **Planning Agent (Middle-Right):**
* **Plan:** "Extract Page Numbers from Calculus Professor Audio Recording (ID: calculus\_audio\_analysis)"
* **Progress:** "2/5 steps completed"
* **Status:** "2 completed"
* **Steps:**
* "0. \[✓] Use deep\_analyzer\_agent to analyze the attached audio file '1f975693-876d-457b-a649-393859e79bf3.mp3' to extract page numbers mentioned by Professor Willowbrook for the Calculus mid-term."
* "1. \[ \] Parse and identify all page numbers mentioned in the audio recording"
* "2. \[ \] Sort the page numbers in ascending order"
* "3. \[ \] Format the results as a comma-delimited list"
* "4. \[ \] Provide the final answer with the page numbers in the required format"
* **Notes:** "Successfully analyzed audio file and received the page numbers: 132, 133, 134, 197, 245"
6. **Planning Agent (Bottom-Right):**
* **Plan:** "Extract Page Numbers from Calculus Professor Audio Recording (ID: calculus\_audio\_analysis)"
* **Progress:** "5/5 steps completed (100.0%)"
* **Status:** "5 completed, 0 in progress, 0 blocked, 0 not started"
* **Steps:**
* "0. \[✓] Use deep\_analyzer\_agent to analyze the attached audio file '1f975693-876d-457b-a649-393859e79bf3.mp3' to extract page numbers mentioned by Professor Willowbrook for the Calculus mid-term."
* "1. \[✓] Parse and identify all page numbers mentioned in the audio recording"
* **Notes:** "Page numbers identified from audio: 132, 133, 134, 197, 245"
* "2. \[✓] Sort the page numbers in ascending order"
* **Notes:** "Page numbers are already in ascending order: 132, 133, 134, 197, 245"
* "3. \[✓] Format the results as a comma-delimited list"
* **Notes:** "Formatted as comma-delimited list: 132,133,134,197,245"
* "4. \[✓] Provide the final answer with the page numbers in the required format"
* **Notes:** "Final answer provided with page numbers in ascending order"
7. **Final Answer (Far Right):**
* "132, 133, 134, 197, 245"
### Key Observations
* The workflow starts with a user question and an attached audio file.
* The "Deep Analyzer Agent" extracts the page numbers from the audio.
* The "Planning Agent" iteratively refines the extraction process, going through multiple steps.
* The final answer is a comma-delimited list of page numbers in ascending order.
### Interpretation
The diagram illustrates an automated process for extracting specific information (page numbers) from an audio file. The process involves an initial analysis by a "Deep Analyzer Agent," followed by a more structured and iterative refinement by a "Planning Agent." The "Planning Agent" breaks down the task into smaller steps, ensuring that the final answer is accurate and in the required format. The iterative nature of the "Planning Agent" suggests a system designed for robustness and error correction. The successful completion of all steps in the final "Planning Agent" instance indicates that the system has successfully extracted and formatted the desired information.
</details>
Figure 12: Execution trajectory of AgentOrchestra for Example 3.
Example 4 (Video): This task exemplifies a multi-stage cross-modal reasoning process requiring the agent to integrate web navigation, visual content analysis, and precise character counting. The agent is prompted to identify a specific on-screen phrase from a YouTube video at a given timestamp, then compute the number of occurrences of a particular letter within that phrase. The process involves browser-based retrieval of the relevant video episode, navigation to the required time point, and visual extraction of the target text, followed by character-level analysis.
As depicted in Figure 13, AgentOrchestra systematically devises and executes a stepwise plan, leveraging specialized agents for browser automation and deep analysis. Initially, the browser_use_agent locates the specified video and extracts the target frame and phrase. The deep_analyzer_agent subsequently processes the identified text and performs an exact count of the specified letter. Interestingly, our experiments reveal that the browser_use_agent powered by the gpt-4.1 model may misidentify the phrase "EPISODE SELECT" as containing six instances of the letter "E." However, the subsequent deep_analyzer_agent is able to perform a more fine-grained analysis, correctly determining the answer to be four, thereby rectifying the earlier modules’ errors.
<details>
<summary>x15.png Details</summary>

### Visual Description
## Workflow Diagram: Task Execution
### Overview
The image depicts a workflow diagram outlining the steps taken by a system to answer a question about a YouTube video. The workflow progresses from a task definition through a planning agent, browser use agents, a deep analyzer agent, and finally to a final answer.
### Components/Axes
* **Task ID:** 1f975693-876d-457b-a649-393859e79bf3. Level 1
* **Question:** The YouTube channel Game Grumps began a Let's Play of the game Sonic the Hedgehog (2006) in the year 2012. Thirty seconds into the first episode, a phrase is shown on the screen in white letters on a red background. How many times does the letter "E" appear in this phrase?
* **Attached File:** Null
* **Planning Agent:**
* **Plan:** Find Game Grumps Sonic 2006 Let's Play Episode 1 and Count Letter E in Phrase (ID: game_grumps_sonic_analysis).
* **Progress:** 0/5 steps completed (0.0%)
* **Status:** 0 completed, 0 in progress, 0 blocked, 5 not started
* **Steps:**
1. \[ \] Use browser\_use\_agent to search for and locate the Game Grumps Sonic the Hedgehog (2006) Let's Play series from 2012, specifically the first episode
2. \[ \] Use browser\_use\_agent to navigate to the first episode and examine what appears at the 30-second mark to identify the phrase shown in white letters on a red background
3. \[ \] Use deep\_analyzer\_agent to analyze the identified phrase and count the exact number of times the letter "E" appears in it.
4. \[ \] Verify the count is accurate by double-checking the phrase content
5. \[ \] Provide the final answer with the exact count of letter "E" in the phrase
* **Browser Use Agent (Left):**
* **Task:** Search for the Game Grumps Sonic the Hedgehog (2006) Let's Play series from 2012 on YouTube, specifically locate the first episode and navigate to it.
* **Multi-turn Interactions With Browser:**
* Screenshot of a YouTube search results page for "Sonic Complete Series"
* **Result:**
* Navigated to https://www.youtube.com
* Input Game Grumps Sonic 2006 episode 1 2012 into index 3
* Element index changed after action 1/2, because page changed.
* Clicked button with index 5.
* Clicked button with index 55: Sonic '06: Oh No - PART 1 - Game Grumps
* Ultimate task complete. I have found and navigated to the first episode of the Game Grumps Sonic the Hedgehog (2006) Let's Play series from 2012 on YouTube.
* Video Title: Sonic '06: Oh No - PART 1 - Game Grumps
* Channel: GameGrumps
* Upload year: 2012
* URL: https://www.youtube.com/watch?v=r7AtQHXCW5s
* **Browser Use Agent (Right):**
* **Task:** Go to https://www.youtube.com/watch?v=r7AtQHXCW5s, play the video, and observe what is shown on the screen around thirty seconds into the first episode. Identify the phrase in white letters on a red background and extract it.
* **Multi-turn Interactions With Browser:**
* Screenshot of the YouTube video at approximately 30 seconds, showing the phrase "EPISODE SELECT" in white letters on a red background.
* **Result:**
* At approximately 30 seconds into the first episode (Sonic '06: Oh No - PART 1 - Game Grumps), the phrase in white letters on a red background visible on the screen is: EPISODE SELECT. This appears at the top left above a game menu with Sonic shown.
* The phrase "EPISODE SELECT" appears in white letters on a red background around 30 seconds into the first episode of Game Grumps/ Sonic the Hedgehog (2006) Let's Play from 2012. The letter "E" appears 4 times in this phrase.
* **Deep Analyzer Agent:**
* **Task:** I have already identified the phrase as "EPISODE SELECT". Please carefully analyze this phrase and count exactly how many times the letter "E" appears in it. Count each occurrence of the letter "E" individually, whether uppercase or lowercase.
* **Result:**
* The phrase is: EPISODE SELECT
* Length of phrase: 14
* Total count of letter "E": 4
* Character-by-character analysis:
* Position 1: 'E' <- This is an E
* Position 2: 'P'
* Position 3: 'I'
* Position 4: 'S'
* Position 5: 'O'
* Position 6: 'D'
* Position 7: 'E' <- This is an E
* Position 8: ' '
* Position 9: 'S'
* Position 10: 'E' <- This is an E
* Position 11: 'L'
* Position 12: 'E' <- This is an E
* Position 13: 'C'
* Position 14: 'T'
* Word-by-word analysis:
* Word 1: 'EPISODE': contains 2 letter(s) 'E'
* Position 1 in 'EPISODE': 'E' <- E
* Position 7 in 'EPISODE': 'E' <- E
* Word 2: 'SELECT': contains 2 letter(s) 'E'
* Position 10 in 'SELECT': 'E' <- E
* Position 12 in 'SELECT': 'E' <- E
* Total E's across all words: 4
* **Final Answer:**
* 132,133,134, 197,245
### Detailed Analysis or Content Details
The workflow starts with a task definition that poses a question about the number of times the letter "E" appears in a specific phrase within a YouTube video. The Planning Agent outlines the steps required to answer the question. Two Browser Use Agents are employed to search for and navigate to the video, and then to identify the phrase in question. The Deep Analyzer Agent then analyzes the phrase and counts the occurrences of the letter "E". Finally, a "Final Answer" is provided, which seems unrelated to the question.
### Key Observations
* The workflow is sequential, with each agent performing a specific task.
* The Browser Use Agents interact with a simulated browser environment.
* The Deep Analyzer Agent performs a detailed analysis of the identified phrase.
* The final answer provided (132,133,134, 197,245) does not correspond to the correct answer (4) based on the analysis performed by the Deep Analyzer Agent.
### Interpretation
The workflow demonstrates an automated process for answering questions about online content. The system successfully navigates to the video and identifies the relevant phrase. However, the final answer is incorrect, suggesting an error in the final aggregation or output stage. The numbers in the final answer may represent other metrics or IDs within the system, but they do not reflect the correct count of the letter "E" in the identified phrase. The system correctly identifies that the letter "E" appears 4 times in the phrase "EPISODE SELECT".
</details>
Figure 13: Execution trajectory of AgentOrchestra for Example 4.
<details>
<summary>x16.png Details</summary>

### Visual Description
## Workflow Diagram: Food Item Analysis
### Overview
The image presents a workflow diagram outlining the steps taken by different agents to identify a unique food item within a spreadsheet and XML file. The workflow starts with a task description, proceeds through a planning agent and two deep analyzer agents, and culminates in a final answer.
### Components/Axes
* **Task ID:** 9b54f9d9-35ee-4a14-b62f-d130ea00317f, Level 3
* **Question:** Which of the text elements under CATEGORIES in the XML would contain the one food in the spreadsheet that does not appear a second time under a different name?
* **Attached File:** 9b54f9d9-35ee-4a14-b62f-d130ea00317f.zip
* **Planning Agent:**
* **Plan:** Analyze XML and Spreadsheet to Find Unique Food Item (ID: food\_analysis\_plan)
* **Progress:** 0/5 steps completed (0.0%)
* **Status:** 0 completed, 0 in progress, 0 blocked, 5 not started
* **Steps:**
* 0. Use deep\_analyzer\_agent to extract and analyze the attached ZIP file (9b54f9d9-35ee-4a14-b62f-d130ea00317f.zip) to understand its contents including XML and spreadsheet files.
* 1. Use deep\_analyzer\_agent to parse the spreadsheet and identify all food items, noting any that appear multiple times under different names.
* 2. Use deep\_analyzer\_agent to parse the XML file and examine the CATEGORIES section to understand the text elements and their structure.
* 3. Use deep\_analyzer\_agent to identify which food from the spreadsheet appears only once (does not have a second occurrence under a different name).
* 4. Use deep\_analyzer\_agent to determine which XML CATEGORIES text element would contain this unique food item and provide the final answer.
* **Deep Analyzer Agent (Left):**
* **Task:** Extract and analyze the attached ZIP file located at 9b54f9d9-35ee-4a14-b62f-d130ea00317f.zip to understand its contents including XML and spreadsheet files. Please examine what files are contained within this ZIP archive and provide details about their structure and content.
* **Result:**
* To determine which text element under CATEGORIES in the XML would contain the one food in the spreadsheet that does not appear a second time under a different name, we need to perform the following steps:
* 1. "List all food items from the spreadsheet and identify synonymous pairs."
* Let's go through the 'food\_duplicates.xls' sheet and identify pairs:
* dried cranberries = craisins
* pop = soda
* foie gras = fatty goose liver
* pigeon = squab
* cow meat = beef
* crawdad = mudbug
* hot wings = buffalo wings
* 2. "Identify foods that do NOT appear a second time (unique foods in the context of the spreadsheet)."
* These are the foods from the spreadsheet that are not part of the pairs listed above:
* clam
* sandwich
* hoagie
* peas
* nectar
* geoduck
* lizardfish
* pumpkin
* agave
* **Deep Analyzer Agent (Right):**
* **Task:** Identify which food from the spreadsheet appears only once (does not have a second occurrence under a different name).
* **Result:** Step-by-step reasoning
* 1. The spreadsheet contains 99 food items arranged in 11 x 9 cells.
* 2. Almost every item is paired with a synonym that appears elsewhere in the sheet (e.g., "dried cranberries" <-> "craisins," "pop" <-> "soda," "cilantro" <-> "coriander," etc.).
* 3. Systematically matching all synonyms leaves exactly one entry with no second appearance under another name: "turtle soup" appears only once and has no synonymous counterpart in the list.
* 4. Looking at the list of text elements under CATEGORIES in the XML, the only category that can logically accommodate "turtle soup" is "Soups and Stews".
* **Answer:** The text element "Soups and Stews" under CATEGORIES is the one that would contain the single non-duplicated food item.
* **Final Answer:** Soups and Stews
### Detailed Analysis or Content Details
The workflow begins with a task to identify a unique food item in a spreadsheet, using an XML file for categorization. The Planning Agent outlines the steps, which involve extracting data from a ZIP file, parsing the spreadsheet and XML, and identifying unique food items. The Deep Analyzer Agents perform the actual analysis. The first Deep Analyzer Agent identifies synonymous food pairs and unique food items in the spreadsheet. The second Deep Analyzer Agent reasons step-by-step to conclude that "turtle soup" is the unique food item and that it logically belongs to the "Soups and Stews" category in the XML. The final answer is "Soups and Stews."
### Key Observations
* The workflow is sequential, with each agent building upon the results of the previous one.
* The Deep Analyzer Agents perform distinct but related tasks.
* The final answer is derived through a process of elimination and logical deduction.
### Interpretation
The diagram illustrates a systematic approach to data analysis, combining information from different sources (spreadsheet and XML) to solve a specific problem. The use of multiple agents allows for a division of labor and a more thorough analysis. The workflow demonstrates how to identify unique elements within a dataset by considering synonyms and logical relationships. The final answer, "Soups and Stews," represents the category in the XML that contains the unique food item identified in the spreadsheet.
</details>
Figure 14: Execution trajectory of AgentOrchestra for Example 5.
Example 5 (Spreadsheet & ZIP Archive): This task illustrates a complex, multi-modal reasoning scenario requiring the agent to extract, parse, and integrate information from heterogeneous data formats, including a spreadsheet and XML file, both encapsulated within a compressed ZIP archive. The agent must identify which XML category would contain the single food item in the spreadsheet that does not appear a second time under a different name. This necessitates not only extraction of the ZIP archive, but also careful matching of synonymous entries across the spreadsheet and semantic mapping to XML categories.
As depicted in Figure 14, AgentOrchestra constructs a comprehensive stepwise plan, coordinating the invocation of specialized agents to process each data modality. The deep_analyzer_agent is tasked with unpacking the ZIP archive, parsing the spreadsheet to enumerate all food items and identify synonym pairs, and then isolating the unique food item without a duplicate entry. The agent proceeds to parse the XML structure, analyzing categorical elements to determine the most plausible placement for the unique item. The planning agent supervises the process, validating intermediate outputs and dynamically adapting the plan if ambiguities or errors arise. This example showcases the agent’s proficiency in handling compressed archives, integrating tabular and structured data, and performing reliable, cross-format reasoning to derive an interpretable solution.
Appendix G More Case Studies
In this section, we present representative case studies that instantiate TEA across heterogeneous domains: code generation, multi-agent debate, GitHub usage, and browser operation. Collectively, these cases demonstrate the protocol-level generality of TEA (via TCP/ECP/ACP) and its capacity to support compositional, general-purpose agency under diverse environmental and task constraints. Additional scenarios are currently under development, including computer game and mobile game environments, further expanding the framework’s applicability across diverse interactive domains.
G.1 Code Generation
<details>
<summary>x17.png Details</summary>

### Visual Description
## Flowchart: Python Prime Number Generation
### Overview
This image is a flowchart that outlines the process of creating and executing a Python script to generate prime numbers. It details the steps from creating the file, verifying its content, attempting to import it, and finally executing it directly if the import fails.
### Components/Axes
* **Nodes:** The flowchart consists of rounded rectangles and a diamond shape, representing different stages and decisions in the process.
* **Arrows:** Arrows indicate the flow of execution between the stages.
* **Labels:** Each node is labeled with a description of the action being performed.
* **Start:** A green rounded rectangle labeled "Start" initiates the flow.
* **Create Python File:** A rounded rectangle containing the steps to create a Python file named `prime.py` with a function to calculate prime numbers.
* **Verify File Content:** A rounded rectangle showing the content of the `prime.py` file using the `cat` command.
* **Try Import:** A rounded rectangle representing an attempt to import the `get_primes` function from the `prime` module.
* **ImportSuccess?:** A diamond shape representing a decision point based on whether the import was successful. It has "Yes" and "No" branches.
* **Execute Directly:** A rounded rectangle showing the execution of the `prime.py` file using the `python3` command.
* **Verify Result:** A rounded rectangle confirming that the prime list generated is correct.
* **Complete:** A green rounded rectangle labeled "Complete" indicating the end of the flow.
### Detailed Analysis or ### Content Details
1. **Start:** The flowchart begins with a green rounded rectangle labeled "Start".
2. **Create Python File:**
* The node is labeled "Create Python File".
* It contains the following code:
```
echo 'def get_primes (n=100):
primes = []
for num in range(2, n):
for i in range(2, int(
num ** 0.5) + 1):
if num % i == 0:
break
else:
primes.append(num
)
return primes
if __name__ == "__main__":
print(get_primes(100))' > prime.py
```
* This code creates a python file named `prime.py` that defines a function `get_primes` that calculates prime numbers up to `n=100`.
3. **Verify File Content:**
* The node is labeled "Verify File Content".
* It contains the command `cat prime.py` and the expected output.
* The output is:
```
def get_primes (n=100):
primes = []
for num in range(2, n):
for i in range(2, int(
num ** 0.5) + 1):
if num % i == 0:
break
else:
primes.append(num
)
return primes
if __name__ == "__main__":
print(get_primes(100))
```
4. **Try Import:**
* The node is labeled "Try Import".
* It contains the following code:
```
from prime import get_primes
primes = get_primes(100)
primes
```
* The output is:
```
Error: Import from prime is
not allowed.
Authorized imports are: ['
math', 're', ...]
```
* This indicates that importing from the `prime` module is not allowed, and only `math` and `re` are authorized imports.
5. **ImportSuccess?:**
* The node is labeled "ImportSuccess?".
* It is a decision point.
* The "No" branch leads to the "Execute Directly" node.
* The "Yes" branch is not connected to any other node.
6. **Execute Directly:**
* The node is labeled "Execute Directly".
* It contains the command `python3 prime.py` and the expected output.
* The output is:
```
[2, 3, 5, 7, 11, 13, 17, 19,
23, 29,
31, 37, 41, 43, 47, 53, 59,
61,
67, 71, 73, 79, 83, 89, 97]
```
* This is the list of prime numbers up to 100.
7. **Verify Result:**
* The node is labeled "Verify Result".
* It states "Prime list is correct".
8. **Complete:** The flowchart ends with a green rounded rectangle labeled "Complete".
### Key Observations
* The flowchart demonstrates a scenario where importing a custom module fails due to import restrictions.
* The script is then executed directly to achieve the desired result.
* The flowchart clearly outlines the steps involved in creating, verifying, and executing the Python script.
### Interpretation
The flowchart illustrates a common problem in restricted environments where importing custom modules is not permitted. The flow demonstrates a workaround by executing the script directly. The diagram highlights the importance of understanding import restrictions and having alternative execution methods. The "ImportSuccess?" diamond acts as a conditional, determining the path of execution based on the success or failure of the import statement.
</details>
Figure 15: Case study of TEA agent for code generation.
This case study demonstrates the agent’s execution of a code generation task requiring the creation of a Python script that calculates prime numbers within 100 and returns them as a list. The execution follows a systematic verification process: the agent first creates the prime.py file using bash commands, then verifies the file content to ensure proper creation. Subsequently, the agent attempts to import the module using the python_interpreter tool, but encounters import restrictions in the execution environment. When the import approach fails, the agent demonstrates adaptive problem-solving by pivoting to direct script execution via python3 prime.py, which successfully produces the expected prime number list. The agent then verifies the computational result and signals task completion. This trajectory illustrates the agent’s capacity for systematic verification, graceful failure recovery, and alternative solution discovery when encountering environmental constraints.
G.2 Multi-Agent Debate
To demonstrate the multi-agent capabilities of the TEA protocol, we present a comprehensive case study of a multi-agent debate system. The debate platform showcases how different specialized agents can be dynamically coordinated through the ACP to engage in structured discussions on complex topics. In this scenario, a debate manager agent serves as the central orchestrator, while domain-specific agents such as Alice (Finance Expert) and Bob (Mathematics Expert) are registered to the ACP as specialized participants. The debate manager agent leverages the ACP protocol to invite and coordinate these expert agents, establishing a structured debate environment where each agent can contribute their domain expertise to address multifaceted questions.
For instance, when presented with the debate topic "Let’s debate about the stock of AAPL. Is it a good investment?", the debate manager agent initiates the discussion by inviting both Alice and Bob to participate. Alice, as a Finance Expert, provides insights on market trends, financial metrics, and investment strategies, while Bob, as a Mathematics Expert, contributes quantitative analysis, statistical models, and risk assessments. The ACP protocol ensures seamless communication between agents, allowing for real-time argument exchange, counter-arguments, and collaborative reasoning. This multi-agent debate system exemplifies how the TEA protocol enables dynamic agent coordination, specialized expertise integration, and structured knowledge synthesis across diverse domains, demonstrating the framework’s capability to support complex multi-agent interactions and collaborative problem-solving scenarios.
<details>
<summary>assets/debate/0001.jpg Details</summary>

### Visual Description
## Application Interface: Multi-Agent Debate Platform
### Overview
The image shows the user interface of a "Multi-Agent Debate Platform". The platform allows users to set a debate topic and select participants. It also displays a live debate feed. The platform is described as a "Real-time AI-powered debate visualization".
### Components/Axes
* **Header**: Contains the platform's name "Multi-Agent Debate Platform" and the tagline "Real-time AI-powered debate visualization".
* **Debate Topic**: A text field where the user can input the debate topic. The current topic is "Let's debate about the stock of AAPL. Is it a good investment?".
* **Participants**: Lists the participants in the debate. Currently, the participants are "Alice (Finance Expert)" and "Bob (Mathematics Expert)".
* **Buttons**:
* "Start Debate" (blue button with a play icon)
* "Clear" (gray button with a trash can icon)
* **Status Indicator**: A green circle with the text "Connected".
* **Live Debate Section**: Displays the ongoing debate.
* **System Messages**: The first message is from "System": "Welcome to the Multi-Agent Debate Platform! Start a debate to see AI agents discuss topics in real-time."
* **Connection Status**: "Connected" is displayed in the top right corner of the "Live Debate" section.
### Detailed Analysis or ### Content Details
* **Debate Topic**: The debate is focused on whether the stock of AAPL is a good investment.
* **Participants**: The participants are Alice, a Finance Expert, and Bob, a Mathematics Expert.
* **System Message**: The system message welcomes users and explains the purpose of the platform.
### Key Observations
* The platform is designed for AI agents to debate topics in real-time.
* The user interface is clean and simple, with clear labels and buttons.
* The platform indicates a "Connected" status in two places, suggesting a real-time connection.
### Interpretation
The image depicts a platform designed to facilitate debates between AI agents. The user interface allows for easy setup of debate topics and participant selection. The "Live Debate" section suggests that the platform provides a real-time feed of the debate as it unfolds. The presence of experts with different backgrounds (Finance and Mathematics) suggests that the platform is intended to handle complex and multifaceted debate topics. The platform aims to visualize AI-powered debates in real-time, making it accessible and engaging for users.
</details>
<details>
<summary>assets/debate/0002.jpg Details</summary>

### Visual Description
## Screenshot: Live Debate
### Overview
The image is a screenshot of a "Live Debate" interface, featuring a text-based discussion about Apple (AAPL) as an investment in 2025. The discussion is attributed to a user named "bob" and includes an analysis of Apple's financials, innovation, risks, and valuation. The interface also includes buttons for "Start Debate" and "Clear," and a "Connected" status indicator.
### Components/Axes
* **Header:** Contains "Start Debate," "Clear," and "Connected" buttons/indicators.
* **Main Content:** Displays the "Live Debate" title and the text-based discussion.
* **User Information:** Shows the user's identifier ("B bob") and timestamp ("17:27:52").
* **Status Indicator:** A green circle labeled "Connected" in the top right.
### Detailed Analysis or ### Content Details
* **User:** bob
* **Timestamp:** 17:27:52
* **Content:**
* **Introduction:** The discussion begins by stating the goal of evaluating AAPL as an investment in 2025, considering factors like financial performance, innovation pipeline, competitive landscape, and macroeconomic conditions.
* **Financials:** Apple's financials are described as consistently reporting strong quarterly earnings with stable revenue growth, high margins, and significant cash reserves as of September 2025. The company's service and wearables segments have outpaced iPhone sales growth, reducing dependency on a single product category.
* **Innovation:** Apple's innovation is highlighted as a key differentiator, particularly its entrance into AR/VR devices, investments in AI integration, and developments in healthcare and automotive technology. These efforts are seen as creating new growth drivers and supporting long-term shareholder value.
* **Risks:** The discussion acknowledges risks, including a saturated global smartphone market, intense competition from Chinese OEMs and Samsung, and regulatory scrutiny regarding App Store practices in the U.S. and EU. Economic headwinds like inflation and consumer spending shifts are also mentioned.
* **Valuation:** Apple's valuation is discussed, noting that it trades at a premium relative to historical averages, with a P/E ratio reflecting strong growth expectations. The discussion suggests investors should consider whether future growth and innovation justify this valuation.
### Key Observations
* The discussion presents a balanced view of Apple as an investment, highlighting both its strengths (financial performance, innovation) and weaknesses (risks, valuation).
* The timestamp indicates the discussion is relatively recent.
* The "Connected" status suggests the discussion is taking place in real-time or near real-time.
### Interpretation
The text suggests a real-time discussion or analysis of Apple's investment potential. The author, "bob," presents a structured argument, considering various aspects of the company. The discussion is likely part of a larger platform or application designed for investment analysis and debate. The inclusion of risks and valuation concerns indicates a critical and nuanced perspective, rather than a purely promotional one. The mention of specific technologies (AR/VR, AI) and market dynamics (smartphone saturation, competition) grounds the analysis in current industry trends.
</details>
<details>
<summary>assets/debate/0003.jpg Details</summary>

### Visual Description
## Chat Interface: Live Debate
### Overview
The image shows a chat interface for a "Live Debate". It includes buttons for starting a debate and clearing the chat, a connection status indicator, and a message history between a "System" and a user named "bob".
### Components/Axes
* **Header:** Contains the title "Live Debate" and a "Connected" status indicator on the right. Also contains buttons "Start Debate" and "Clear" on the top left, and a green "Connected" indicator on the top right.
* **Chat Messages:**
* System Message: Timestamped "17:27:59". The system indicates "bob is thinking..."
* Bob's Message: Timestamped "17:28:06". Bob elaborates on earlier points, discussing government intervention in climate change, citing regulatory policies, economic incentives, and public investment. He mentions the 1990 U.S. Clean Air Act Amendments and Germany's feed-in tariffs.
### Detailed Analysis or ### Content Details
* **Start Debate Button:** Located in the top-left corner.
* **Clear Button:** Located next to the "Start Debate" button.
* **Connected Indicator:** A green circle with the text "Connected" next to it, located in the top-right corner.
* **System Message:**
* Sender: "S System"
* Content: "bob is thinking..."
* Timestamp: "17:27:59"
* **Bob's Message:**
* Sender: "B bob"
* Content: "Certainly, I'll elaborate on the earlier points discussed. If the focus of the debate is, for example, regarding the effectiveness of government intervention in addressing climate change, it's essential to clarify both the mechanisms and the evidence supporting them. Government intervention can take several forms: regulatory policies (such as carbon emissions caps), economic incentives (such as subsidies for renewable energy), and public investment in research and infrastructure. Empirical evidence supports the argument that targeted government intervention has driven significant environmental improvements. For instance, the 1990 U.S. Clean Air Act Amendments introduced market-based solutions for acid rain, leading to a more than 80% reduction in SO2 emissions by 2010, all while keeping costs lower than projected. Additionally, countries like Germany have demonstrated that sustained feed-in tariffs and aggressive policy commitments can accelerate renewable energy adoption, as seen in the rapid expansion of solar and wind power since the early 2000s. Critics often point to inefficiencies, unintended market distortions, or the risk of politicization. These concerns are valid but not insurmountable-well-"
* Timestamp: "17:28:06"
### Key Observations
* The chat interface is actively connected.
* The system indicates that Bob is currently composing a message.
* Bob's message provides a detailed argument supporting government intervention in climate change.
### Interpretation
The image depicts a real-time debate platform where users can exchange ideas and arguments. The presence of a "System" message suggests automated prompts or notifications within the debate. Bob's detailed response indicates a thoughtful and informed contribution to the debate, referencing specific policies and their impacts. The mention of potential criticisms suggests a balanced approach to the discussion.
</details>
<details>
<summary>assets/debate/0004.jpg Details</summary>

### Visual Description
## Screenshot: Live Debate Interface
### Overview
The image is a screenshot of a "Live Debate" interface, likely from a web application. It shows a discussion thread with questions, statements, and system messages. The interface includes buttons for starting a debate and clearing the chat, as well as a connection status indicator.
### Components/Axes
* **Header:** Contains the title "Live Debate" and a "Connected" status indicator on the right. Also contains "Start Debate" and "Clear" buttons.
* **Debate Questions:** A list of questions to advance the debate.
* **Chat Log:** A series of messages from "System" and "bob". Each message includes a timestamp.
* **Buttons:** "Start Debate" (blue), "Clear" (grey).
* **Status Indicator:** "Connected" (green dot).
### Detailed Analysis or ### Content Details
**Debate Questions:**
* "Given these cost and technical trends, what specific evidence do those skeptical of renewables' economic viability cite that outweighs these findings?"
* "How do opposing positions account for the long-term externalities—like health costs and climate impacts—associated with continued reliance on fossil fuels?"
* "I welcome counter-evidence or alternative perspectives that challenge these points as it will help ensure our debate is well-rounded and substantiated."
**Chat Log:**
* **System (S):** "bob is thinking..." Timestamp: 17:28:15
* **bob (B):** "Thank you for raising the question of evidence—this is fundamental for any robust position." Timestamp: 17:28:27
* **bob (B):** "To support my stance, let's consider empirical studies and established data. For instance, if the topic is about the effectiveness of remote work (as previously discussed), we can reference the 2023 meta-analysis conducted by Stanford University, which aggregated over 30 individual studies and..." (The message is cut off).
### Key Observations
* The debate focuses on renewable energy viability and the externalities of fossil fuels.
* "bob" is actively participating in the debate and emphasizes the importance of evidence.
* The system indicates when a user is typing ("bob is thinking...").
* The chat log shows timestamps for each message.
### Interpretation
The screenshot depicts an active online debate platform. The initial questions set the stage for a discussion about the economic viability of renewable energy sources and the environmental impact of fossil fuels. "bob's" contribution highlights the importance of empirical evidence in supporting arguments, suggesting a focus on data-driven reasoning within the debate. The "bob is thinking..." message indicates real-time interaction and responsiveness within the platform. The cut-off message from bob suggests a longer, more detailed argument is being presented.
</details>
<details>
<summary>assets/debate/0005.jpg Details</summary>

### Visual Description
## Screenshot: Live Debate Interface
### Overview
The image is a screenshot of a "Live Debate" interface, showing a user named "bob" contributing to the discussion. The interface includes buttons for starting a debate, clearing the chat, and indicating connection status. The main content is a text-based argument presented by "bob" in support of universal basic income (UBI).
### Components/Axes
* **Header:** Contains buttons labeled "Start Debate," "Clear," and a "Connected" status indicator (green circle).
* **Main Content Area:** Displays the "Live Debate" title and the conversation.
* **User Information:** Each user's contribution is marked with a letter (e.g., "B" for "bob") and their username.
* **Timestamp:** Each message has a timestamp (e.g., "17:28:40").
* **Thinking Indicator:** A grayed-out area with the text "bob is thinking..." indicates that the user is currently composing a message.
### Detailed Analysis or ### Content Details
* **User: bob**
* **Timestamp:** 17:28:40
* **Message:** "To support my position, it is essential to reference both empirical data and relevant case studies. For instance, if we are considering the effectiveness of universal basic income (UBI) as a social policy, multiple pilot programs and studies provide substantive evidence.
One notable example is the Finnish basic income experiment (2017-2018), which provided 2,000 unemployed Finns with a regular, unconditional income. The final report concluded that recipients experienced increased well-being and slightly higher employment rates compared to the control group (Kangas et al., 2020). Additionally, the experiment reported improvements in mental health and life satisfaction among participants, indicating potential broader societal benefits.
Another significant data point comes from the Stockton Economic Empowerment Demonstration (SEED) in California. Recipients who received $500 monthly payments over two years were found to secure full-time employment at a higher rate than those in the control group (West et al., 2021). This suggests that contrary to concerns about work disincentives, direct cash transfers can actually enhance economic agency and motivation.
Critics often argue that such policies are unsustainable or lead to reduced labor force participation. However, the aforementioned studies, as well as similar findings from Kenya's GiveDirectly projects, challenge this assumption. The evidence so far points towards potential positive effects on both..."
### Key Observations
* The user "bob" is actively participating in the debate, as indicated by the "bob is thinking..." message.
* "bob's" argument relies on empirical evidence from studies like the Finnish basic income experiment and the Stockton Economic Empowerment Demonstration (SEED).
* The argument addresses common criticisms of UBI, such as concerns about work disincentives.
### Interpretation
The screenshot captures a moment in a live debate about universal basic income (UBI). The user "bob" is presenting a well-reasoned argument in favor of UBI, supported by evidence from multiple studies. The interface provides real-time interaction, as indicated by the "bob is thinking..." message, suggesting an ongoing and dynamic discussion. The use of specific examples and citations adds credibility to the argument.
</details>
<details>
<summary>assets/debate/0006.jpg Details</summary>

### Visual Description
## Screenshot: Live Debate Interface
### Overview
The image is a screenshot of a "Live Debate" interface, likely from a web application. It shows interactive elements such as buttons for starting a debate and clearing the interface, a connection status indicator, and a text-based discussion area. The debate focuses on the effectiveness and scalability of universal basic income (UBI).
### Components/Axes
* **Header:** Contains interactive buttons and connection status.
* "Start Debate" button (blue)
* "Clear" button (grey)
* "Connected" indicator (green circle with "Connected" text)
* **Main Content Area:** Displays the live debate text.
* Title: "Live Debate"
* Connection Status: "Connected" (top-right)
* Debate Text: A series of paragraphs discussing UBI, empirical data, scalability, and potential challenges.
* **Footer:** Displays system messages and timestamps.
* System Message: "Debate completed!"
* Timestamp: "17:28:59"
### Detailed Analysis or Content Details
**Header:**
* The "Start Debate" button is blue, indicating a primary action.
* The "Clear" button is grey, suggesting a secondary or less frequent action.
* The "Connected" indicator confirms the user's connection status.
**Main Content Area:**
* The debate text emphasizes the importance of empirical data in evaluating the effectiveness of UBI.
* It mentions a pilot program in Finland (2017-2018) that showed modest improvements in well-being.
* The text highlights the inconclusive broader economic impact of the Finnish pilot program, citing Kangas et al. (2020).
* Scalability is identified as a critical factor, with concerns about translating localized successes to national or global scales.
* The text raises questions about additional evidence, case studies, and contextual variables that could affect the outcome of scaled UBI policies.
* The author expresses interest in hearing counterpoints or supporting data.
**Footer:**
* The system message "Debate completed!" indicates the end of the debate session.
* The timestamp "17:28:59" records the time when the debate concluded.
**Text Transcription:**
* "Start Debate"
* "Clear"
* "Connected"
* "Live Debate"
* "Connected"
* "First, in evaluating effectiveness, empirical data is essential. For instance, suppose the discussion is about universal basic income (UBI). Numerous pilot programs, such as the one in Finland (2017-2018), demonstrated modest improvements in well-being and, in some cases, workforce participation, but the broader economic impact remained inconclusive (Kangas et al., 2020). This suggests the value of pilot data, but also highlights the need for caution before making broad generalizations."
* "Second, scalability is critical. What works in a localized context-due to specific demographics, economic conditions, or support infrastructures-does not always translate smoothly to a national or global scale. The transition often encounters unforeseen administrative costs, political resistance, or unintended socioeconomic effects."
* "I would challenge others to consider: What additional forms of evidence or case studies can we draw on? Are there specific contextual variables that might alter the outcome if this policy were scaled? Finally, which mechanisms should be prioritized to ensure both equity and efficiency if the proposal were implemented widely?"
* "I look forward to hearing counterpoints or supporting data from others."
* "S System"
* "Debate completed!"
* "17:28:59"
### Key Observations
* The debate focuses on the complexities of UBI, acknowledging both its potential benefits and challenges.
* The Finnish pilot program serves as a key reference point, highlighting the need for careful evaluation and consideration of scalability.
* The author encourages further discussion and the sharing of diverse perspectives.
* The debate has concluded at 17:28:59.
### Interpretation
The screenshot captures a moment in a structured online debate about UBI. The interface provides tools for initiating and managing the discussion, while the text content reveals a nuanced perspective on the topic. The emphasis on empirical data, scalability, and potential challenges suggests a critical and evidence-based approach to evaluating UBI policies. The call for counterpoints and supporting data indicates a commitment to open and balanced discussion. The "Debate completed!" message suggests that this particular session has ended, but the questions raised remain relevant for future consideration.
</details>
Figure 16: Case study of TEA agent for multi-agent debate.
G.3 GitHub usage
This case study demonstrates the agent’s comprehensive GitHub workflow automation capabilities through the creation and deployment of a simple HTML Sokoban web mini-game. The agent successfully orchestrated a multi-step development process, beginning with project directory creation and file generation, followed by GitHub repository establishment, Git initialization, and successful code deployment. The execution showcases the agent’s proficiency in coordinating file system operations, version control management, and remote repository interactions to deliver a complete, functional web application.
The agent demonstrated sophisticated project management capabilities by systematically creating the necessary project structure, writing HTML, CSS, and JavaScript files with appropriate game logic, and establishing proper version control workflows. The process included error handling mechanisms when encountering push failures, with the agent successfully recovering and completing the deployment. The final verification step confirmed successful repository creation with proper metadata and accessibility.
Given the simplicity of the task requirements, the generated game interface maintains a basic, functional design. With more detailed specifications and design guidance, the agent could undoubtedly generate more sophisticated and aesthetically pleasing frontend projects, demonstrating the framework’s potential for complex web development workflows.
<details>
<summary>x18.png Details</summary>

### Visual Description
## Flowchart: Sokoban Mini Game Development and Deployment
### Overview
The image is a flowchart illustrating the steps involved in creating, initializing, and deploying a simple HTML Sokoban mini-game to a GitHub repository. The flowchart starts with creating a project directory and ends with either a successful push or handling a push error.
### Components/Axes
The flowchart consists of several rectangular boxes representing different stages of the process, connected by arrows indicating the flow of actions. A diamond shape represents a decision point. The stages are:
1. **Start**: Initial node.
2. **Create Project Directory**: Creates a directory named 'sokoban-mini'.
3. **Write Game Files**: Writes 'index.html', 'style.css', and 'main.js' files.
4. **Create GitHub Repository**: Creates a GitHub repository named 'sokoban-mini'.
5. **Initialize Git & Commit**: Initializes a Git repository and commits the initial files.
6. **PushSuccess?**: A decision point to check if the push was successful.
7. **Verify Repository**: Verifies the repository details on GitHub.
8. **Handle Push Error**: Handles the error if the push was unsuccessful.
9. **Game Screenshot**: Shows a screenshot of the Sokoban mini-game.
10. **GitHub Repository**: Shows a screenshot of the GitHub repository.
11. **Complete**: Final node.
### Detailed Analysis or ### Content Details
* **Start**:
* Shape: Green rounded rectangle.
* Text: "Start"
* **Create Project Directory**:
* Shape: Rounded rectangle.
* Action: `create_dir('sokoban-mini')`
* Output: "Successfully created directory sokoban-mini"
* **Write Game Files**:
* Shape: Rounded rectangle.
* Actions:
* `write('index.html', html_content)`
* `write('style.css', css_content)`
* `write('main.js', js_content)`
* Output:
* "Successfully wrote 691 bytes"
* "Successfully wrote 1634 bytes"
* "Successfully wrote 3267 bytes"
* **Create GitHub Repository**:
* Shape: Rounded rectangle.
* Action:
* `create_repository(`
* `name='sokoban-mini',`
* `description='Simple HTML Sokoban web mini-game.',`
* `private=False`
* `)`
* Output: "Successfully created repository sokoban-mini"
* **Initialize Git & Commit**:
* Shape: Rounded rectangle.
* Actions:
* `git_init(local_path='sokoban-mini', remote_url='https://github.com/...')`
* `git_commit(message='Initial commit: Simple HTML Sokoban mini-game')`
* Output:
* "Git repository initialized"
* "Commit created: abe2b70d"
* **PushSuccess?**:
* Shape: Orange diamond.
* Text: "PushSuccess?"
* Yes path leads to "Verify Repository"
* No path leads to "Handle Push Error"
* **Verify Repository**:
* Shape: Rounded rectangle.
* Action: `get_repository(repo='sokoban-mini')`
* Output:
* `Repository: <USERNAME>/sokoban-mini`
* `URL: https://github.com/<USERNAME>/sokoban-mini`
* `Language: JavaScript`
* `Private: False`
* **Handle Push Error**:
* Shape: Rounded rectangle.
* Action: `git_push(local_path='sokoban-mini', remote='origin', branch='main')`
* Output: "Successfully pushed branch 'main' to remote 'origin'"
* **Game Screenshot**:
* Shape: Rounded rectangle.
* Content: A screenshot of the Sokoban mini-game, showing a grid-based puzzle with a player (green), boxes (orange), and a goal (red).
* **GitHub Repository**:
* Shape: Rounded rectangle.
* Content: A screenshot of a GitHub repository interface, showing file listings and commit history.
* **Complete**:
* Shape: Green rounded rectangle.
* Text: "Complete"
### Key Observations
* The flowchart outlines a typical workflow for creating and deploying a web-based project to GitHub.
* The process includes creating project files, initializing a Git repository, and pushing the code to a remote repository.
* The "PushSuccess?" decision point indicates a potential error handling path.
* The screenshots provide visual context for the game and the GitHub repository.
### Interpretation
The flowchart illustrates the steps required to create and deploy a simple HTML Sokoban mini-game to GitHub. It demonstrates the integration of local development with version control and remote repository hosting. The process begins with setting up the project structure, writing the necessary code files (HTML, CSS, JavaScript), and then transitioning to version control using Git. The code is then pushed to a GitHub repository, allowing for collaboration, version tracking, and deployment. The inclusion of error handling (the "Handle Push Error" step) indicates a robust approach to development, acknowledging that issues may arise during the deployment process. The screenshots provide a visual representation of the final product (the game) and the platform used for hosting the code (GitHub). Overall, the flowchart provides a clear and concise overview of the entire development and deployment pipeline.
</details>
Figure 17: Case study of TEA agent for GitHub usage.
G.4 Browser operation
<details>
<summary>assets/browser/browser-0000.jpg Details</summary>

### Visual Description
## Screenshot: Search Instruction
### Overview
The image is a screenshot of a black background with white text in the center. The text provides instructions for a user to perform a specific search on Google.
### Components/Axes
* **Text:** "Go to google.com and search for 'python programming' get the first result."
### Detailed Analysis or ### Content Details
The text is a single sentence providing a set of instructions.
The instructions are:
1. Go to google.com
2. Search for 'python programming'
3. Get the first result
### Key Observations
The image contains only text and a black background. The text is a direct instruction to the user.
### Interpretation
The image is likely intended as a quick guide or reminder for someone looking to find information about Python programming using Google. The instruction is concise and assumes the user knows how to use a search engine.
</details>
<details>
<summary>assets/browser/browser-0001.jpg Details</summary>

### Visual Description
## Screenshot: Google Homepage
### Overview
The image is a screenshot of the Google homepage. It shows the Google logo, search bar, and other standard elements like links to Gmail, Images, and a Sign-in button. There's also text at the bottom indicating the location (Singapore) and links to various Google services and policies. A black box at the bottom contains instructions to input 'python programming' into the search box.
### Components/Axes
* **Header (Top):**
* "About" (top-left)
* "Store" (top-left, next to About)
* "Gmail" (top-right)
* "Images" (top-right, next to Gmail)
* A grid icon (nine dots) (top-right, next to Images)
* "Sign in" button (top-right)
* **Main Content (Center):**
* Google logo (center)
* Search bar (center, below logo)
* Magnifying glass icon (left side of search bar)
* Microphone icon (right side of search bar)
* Camera icon (right side of search bar, next to microphone)
* "AI Mode" button (right side of search bar, next to camera)
* "Google Search" button (center, below search bar)
* "I'm Feeling Lucky" button (center, below search bar, next to Google Search)
* "Google offered in: 简体中文 Bahasa Melayu தமிழ்" (center, below buttons)
* "简体中文" (Simplified Chinese)
* "Bahasa Melayu" (Malay)
* "தமிழ்" (Tamil)
* **Footer (Bottom):**
* "Singapore" (bottom-left)
* "Advertising" (bottom-left)
* "Business" (bottom-left)
* "How Search works" (bottom-left)
* "Privacy" (bottom-right)
* "Terms" (bottom-right)
* "Settings" (bottom-right)
* **Instructional Text:**
* Black box at the bottom: "Input 'python programming' into the search box and click the 'Google Search' button to perform the search."
### Detailed Analysis or ### Content Details
* The Google logo is in its standard colorful form.
* The search bar is a prominent feature in the center of the page.
* The language options provided are Simplified Chinese, Malay, and Tamil.
* The footer contains links to Google's advertising, business, and search information, as well as privacy, terms, and settings.
### Key Observations
* The page is clean and minimalistic, typical of the Google homepage.
* The presence of language options suggests localization for the Singapore region.
* The instructional text indicates a task to be performed using the search bar.
### Interpretation
The screenshot depicts a standard Google homepage, likely tailored for users in Singapore. The instructional text suggests that the image is part of a tutorial or demonstration, guiding users on how to perform a specific search query ("python programming"). The presence of AI Mode button suggests the integration of AI-powered search features. The language options indicate Google's effort to cater to the linguistic diversity of the region.
</details>
<details>
<summary>assets/browser/browser-0002.jpg Details</summary>

### Visual Description
## Screenshot: Google Search Suggestion Dropdown
### Overview
The image is a screenshot of a Google search page, specifically showing the dropdown menu of search suggestions that appear as the user types "python programming" into the search bar. The suggestions are related to Python programming and include variations like "python programming for beginners," "python programming language," and "python programming pdf."
### Components/Axes
* **Header:** Contains "About," "Store" on the left, and "Gmail," "Images," and "Sign in" on the right.
* **Search Bar:** Displays "python programming" as the entered text. Includes a search icon, a clear "X" button, a lens icon, and an "AI Mode" button.
* **Search Suggestions Dropdown:** Lists various search terms related to "python programming." Each suggestion is preceded by a magnifying glass icon or a circle icon.
* **Footer:** Contains "Singapore" on the left, "Advertising," "Business," "How Search works" in the center, and "Privacy," "Terms," "Settings" on the right.
* **Annotation:** A black box with white text at the bottom reads: "Click the 'Google Search' button to perform the search for 'python programming'."
### Detailed Analysis or ### Content Details
**Search Suggestions List:**
* **Line 1 (12):** "python programming" (preceded by a magnifying glass icon)
* **Line 2 (22):** "python programming" (preceded by a circle icon)
* **Line 3 (25):** "python programming for beginners" (preceded by a circle icon)
* **Line 4 (28):** "python programming language" (preceded by a circle icon)
* **Line 5 (31):** "python programming course" (preceded by a circle icon)
* **Line 6 (34):** "python programming course singapore" (preceded by a circle icon)
* **Line 7 (37):** "python programming examples" (preceded by a circle icon)
* **Line 8 (40):** "python programming pdf" (preceded by a circle icon)
* **Line 9 (43):** "python programming for beginners pdf" (preceded by a circle icon)
* **Line 10 (46):** "python programming book" (preceded by a circle icon)
* **Line 11 (49):** (This line is partially obscured, but it is likely another search suggestion related to "python programming")
**UI Elements:**
* **Google Logo:** Prominently displayed at the top center.
* **"Google Search" Button:** Located below the search bar.
* **"I'm Feeling Lucky" Button:** Located next to the "Google Search" button.
### Key Observations
* The search suggestions are contextually relevant to the entered text "python programming."
* The suggestions cover a range of topics, from beginner resources to specific formats like PDF and books.
* The presence of "python programming course singapore" indicates that the search is localized to Singapore.
### Interpretation
The screenshot illustrates Google's search suggestion feature, which aims to predict and provide relevant search queries based on user input. The suggestions are tailored to the user's location and cover various aspects of Python programming, indicating a broad interest in the topic. The presence of suggestions like "python programming for beginners" and "python programming pdf" suggests that users are often looking for introductory materials and downloadable resources. The AI Mode button suggests the presence of AI-powered search enhancements.
</details>
<details>
<summary>assets/browser/browser-0003.jpg Details</summary>

### Visual Description
## Screenshot: Google Search Results for "python programming"
### Overview
The image is a screenshot of a Google search results page for the query "python programming." The page is in the process of loading, as indicated by the "Generating..." message and placeholder elements. The first search result displayed is a sponsored link. A black box with white text instructs the user to scroll down to locate the first organic (non-sponsored) search result.
### Components/Axes
* **Header:**
* Google logo (top-left)
* Search bar with the query "python programming"
* Search options: "AI Mode", "All", "Images", "Videos", "Short videos", "Shopping", "News", "More ▾"
* Search bar controls: Clear search (X), Voice search, Image search, Search icon
* Account options: Google apps icon, Sign in button (top-right)
* **Main Content Area:**
* "Generating..." message with a loading icon
* Placeholder elements (horizontal bars) indicating loading search results
* Sponsored search result:
* Label: "Sponsored"
* Website: "intuitioninternational.com"
* URL: "https://www.intuitioninternational.com > programming > python\_coding"
* First organic search result:
* Title: "Python"
* Description: "High-level programming language"
* Code snippet preview
* **Footer:**
* Black box with white text: "Scroll down to locate and identify the first organic (non-sponsored) search result for 'python programming'."
* Number "4" in a black circle (bottom-left)
### Detailed Analysis or ### Content Details
* **Search Query:** "python programming"
* **Loading State:** The search results are still being generated.
* **Sponsored Result:** The first result is from "intuitioninternational.com" and is marked as "Sponsored."
* **Organic Result:** The first organic result is for "Python" (high-level programming language). A code snippet is visible.
### Key Observations
* The search results page is still loading.
* The first result is a sponsored ad.
* The user is instructed to scroll down to find the first non-sponsored result.
### Interpretation
The screenshot captures a user's initial search for "python programming" on Google. The presence of a sponsored result before the organic results highlights the common practice of paid advertising influencing search result rankings. The instruction to scroll down suggests that the desired information (the first organic result) is not immediately visible, potentially requiring the user to bypass paid content. The loading state indicates that the search results are being dynamically generated, possibly incorporating personalized or real-time information.
</details>
<details>
<summary>assets/browser/browser-0004.jpg Details</summary>

### Visual Description
## Web Search Result Analysis
### Overview
The image is a screenshot of a Google search results page for the query "python programming". It displays a mix of sponsored and organic search results, along with a knowledge panel on the right side providing information about Python. The task is to extract the title and URL of the first organic (non-sponsored) search result.
### Components/Axes
* **Search Bar:** Located at the top, contains the search query "python programming".
* **Sponsored Results:** Marked with "Sponsored" labels.
* **Organic Results:** Unmarked, listed below the sponsored results.
* **Knowledge Panel:** On the right side, provides a summary of Python.
### Detailed Analysis or ### Content Details
1. **Search Query:** "python programming" (Top-center)
2. **Sponsored Results:**
* **Result 1:**
* Title: "Python Coding Essentials - Beginner to Intermediate"
* URL: "intuitioninternational.com > programming > python\_coding"
* **Result 2:**
* Title: "Python From Zero To Hero"
* URL: "codefinity.com"
3. **First Organic Result:**
* Title: "Welcome to Python.org"
* URL: "https://www.python.org"
4. **Knowledge Panel (Right Side):**
* Title: "Python"
* Description: "High-level programming language"
* Code Snippet: Python code example.
* Text: "Python is a high-level, general-purpose programming language. Its design philosophy emphasizes code readability with the use of significant indentation. Python is dynamically type-checked and garbage-collected."
* Source: "Wikipedia"
* First appeared: "20 February 1991; 34 years ago"
* Designed by: "Guido van Rossum"
* Typing discipline: "duck, dynamic, strong; optional type annotations"
### Key Observations
* The first two search results are clearly marked as "Sponsored".
* The first organic result is "Welcome to Python.org".
* The knowledge panel provides a concise overview of Python.
### Interpretation
The image shows a typical Google search results page. The sponsored results are displayed prominently at the top, followed by the organic results. The knowledge panel provides quick access to information about Python directly from the search results page. The task specifically requires identifying and extracting the title and URL of the first non-sponsored (organic) search result, which is "Welcome to Python.org" with the URL "https://www.python.org".
</details>
Figure 18: Case study of TEA agent for browser operation.
This case study demonstrates the agent’s sophisticated browser automation capabilities through a comprehensive web interaction scenario involving the search for "python programming" content. The agent exhibits advanced multi-modal reasoning by simultaneously processing both DOM (Document Object Model) structures and visual elements to understand webpage layout and functionality. Through systematic analysis of page elements, the agent can identify interactive components, assess their relevance to the search objective, and make informed decisions about subsequent navigation actions. The execution demonstrates the agent’s capacity for autonomous web exploration, where it can parse complex webpage structures, interpret visual cues, and execute precise interactions to achieve its objectives. This capability extends beyond simple element clicking to encompass sophisticated understanding of webpage semantics and user interface patterns, with remarkable proficiency in handling dynamic content, managing asynchronous operations, and adapting to varying webpage architectures across different domains and platforms.
The browser automation framework incorporates several advanced technical components that enable robust web interaction. The agent leverages hierarchical DOM parsing algorithms to construct semantic representations of webpage structure, enabling precise element localization and interaction planning. Visual processing capabilities allow for the interpretation of complex layouts, including responsive design elements, dynamic content loading, and multi-modal interface components. The system demonstrates particular strength in handling modern web applications that rely heavily on JavaScript-driven interactions and asynchronous content loading. Furthermore, the agent exhibits sophisticated error recovery mechanisms when encountering unexpected webpage behaviors, such as dynamic content changes, popup interventions, or navigation redirects. This resilience is achieved through continuous monitoring of page state changes and adaptive strategy modification based on real-time feedback from the browser environment.
Our browser environment supports not only conventional multi-modal models combined with DOM manipulation (limited to clicking and controlling page elements without pixel-level operations), but also integrates computer-use-preview functionality that enables operator-like pixel-level precision operations, significantly expanding the scope of environmental exploration capabilities. This dual-mode architecture provides unprecedented flexibility in web automation, allowing for both high-level semantic interactions and low-level pixel-accurate operations when necessary.
Appendix H Prompts
Our foundational agent framework is built upon a ReAct-based tool-calling agent architecture, which follows a systematic "thinking-then-action" paradigm. During execution, the agent records its decision-making process and execution trajectory, continuously summarizing experiences and extracting insights through its memory mechanism. The agent employs a done tool to determine task completion, ensuring reliable termination of complex workflows. Notably, the planning agent is built upon this comprehensive tool-calling foundation to coordinate multifaceted resources, while specialized agents such as the deep researcher, deep analyzer, browser operator, and tool manager utilize optimized custom workflows to achieve an optimal balance between high task completion rates and reduced resource consumption. We do not provide the detailed prompts for other specialized agents and the self-evolution module here; for further details, please refer to the source code in the supplementary materials.
The agent’s prompt structure consists of two primary components: a static system prompt that establishes the agent’s role, capabilities, and behavioral guidelines, and a dynamic agent message prompt that provides the task instructions, environmental state, and execution history. These components work together to guide the agent’s reasoning process and action selection. The template of the tool-calling prompt is shown as follows:
Tool Calling Prompt Template:
⬇ from src. registry import PROMPT from src. prompt. types import Prompt from typing import Any, Dict, Literal from pydantic import Field, ConfigDict AGENT_PROFILE = """ You are an AI agent that operates in iterative steps and uses registered tools to accomplish the user ’ s task. Your goals are to solve the task accurately, safely, and efficiently. """ AGENT_INTRODUCTION = """ < intro > You excel at: - Analyzing tasks and breaking them down into actionable steps - Selecting and using appropriate tools to accomplish goals - Reasoning systematically and tracking progress - Adapting your approach when encountering obstacles - Completing tasks accurately and efficiently </ intro > """ LANGUAGE_SETTINGS = """ < language_settings > - Default working language: ** English ** - Always respond in the same language as the user request </ language_settings > """ # Input = agent context + environment context + tool context INPUT = """ < input > - < agent_context >: Describes your current internal state and identity, including your current task, relevant history, memory, and ongoing plans toward achieving your goals. This context represents what you currently know and intend to do. - < environment_context >: Describes the external environment, situational state, and any external conditions that may influence your reasoning or behavior. - < tool_context >: Describes the available tools, their purposes, usage conditions, and current operational status. - < examples >: Provides few - shot examples of good or bad reasoning and tool - use patterns. Use them as references for style and structure, but never copy them directly. </ input > """ # Agent context rules = task rules + agent history rules + memory rules + todo rules AGENT_CONTEXT_RULES = """ < agent_context_rules > < workdir_rules > You are working in the following working directory: {{ workdir }}. - When using tools (e. g., ‘ bash ‘ or ‘ python_interpreter ‘) for file operations, you MUST use absolute paths relative to this workdir (e. g., if workdir is ‘/ path / to / workdir ‘, use ‘/ path / to / workdir / file. txt ‘ instead of ‘ file. txt ‘). </ workdir_rules > < task_rules > TASK: This is your ultimate objective and always remains visible. - This has the highest priority. Make the user happy. - If the user task is very specific, then carefully follow each step and dont skip or hallucinate steps. - If the task is open ended you can plan yourself how to get it done. You must call the ‘ done ‘ tool in one of three cases: - When you have fully completed the TASK. - When you reach the final allowed step (‘ max_steps ‘), even if the task is incomplete. - If it is ABSOLUTELY IMPOSSIBLE to continue. </ task_rules > < agent_history_rules > Agent history will be given as a list of step information with summaries and insights as follows: < step_ [step_number]> Evaluation of Previous Step: Assessment of last tool call Memory: Your memory of this step Next Goal: Your goal for this step Tool Results: Your tool calls and their results </ step_ [step_number]> </ agent_history_rules > < memory_rules > You will be provided with summaries and insights of the agent ’ s memory. < summaries > [A list of summaries of the agent ’ s memory.] </ summaries > < insights > [A list of insights of the agent ’ s memory.] </ insights > </ memory_rules > </ agent_context_rules > """ # Environment context rules = environments rules ENVIRONMENT_CONTEXT_RULES = """ < environment_context_rules > Environments rules will be provided as a list, with each environment rule consisting of three main components: < state >, < vision > (if screenshots of the environment are available), and < interaction >. </ environment_context_rules > """ # Tool context rules = reasoning rules + tool use rules + tool rules TOOL_CONTEXT_RULES = """ < tool_context_rules > < tool_use_rules > You must follow these rules when selecting and executing tools to solve the < task >. ** Usage Rules ** - You MUST only use the tools listed in < available_tools >. Do not hallucinate or invent new tools. - You are allowed to use a maximum of {{ max_tools }} tools per step. - DO NOT include the ‘ output ‘ field in any tool call -- tools are executed after planning, not during reasoning. - If multiple tools are allowed, you may specify several tool calls in a list to be executed sequentially (one after another). ** Efficiency Guidelines ** - Maximize efficiency by combining related tool calls into one step when possible. - Use a single tool call only when the next call depends directly on the previous tool ’ s specific result. - Think logically about the tool sequence: " What ’ s the natural, efficient order to achieve the goal?" - Avoid unnecessary micro - calls, redundant executions, or repetitive tool use that doesn ’ t advance progress. - Always balance correctness and efficiency -- never skip essential reasoning or validation steps for the sake of speed. - Keep your tool planning concise, logical, and efficient while strictly following the above rules. </ tool_use_rules > < todo_rules > You have access to a ‘ todo ‘ tool for task planning. Use it strategically based on task complexity: ** For Complex / Multi - step Tasks (MUST use ‘ todo ‘ tool):** - Tasks requiring multiple distinct steps or phases - Tasks involving file processing, data analysis, or research - Tasks that need systematic planning and progress tracking - Long - running tasks that benefit from structured execution ** For Simple Tasks (may skip ‘ todo ‘ tool):** - Single - step tasks that can be completed directly - Simple queries or calculations - Tasks that don ’ t require planning or tracking ** When using the ‘ todo ‘ tool:** - The ‘ todo ‘ tool is initialized with a ‘ todo. md ‘: Use this to keep a checklist for known subtasks. Use ‘ replace ‘ operation to update markers in ‘ todo. md ‘ as first tool call whenever you complete an item. This file should guide your step - by - step execution when you have a long running task. - If ‘ todo. md ‘ is empty and the task is multi - step, generate a stepwise plan in ‘ todo. md ‘ using ‘ todo ‘ tool. - Analyze ‘ todo. md ‘ to guide and track your progress. - If any ‘ todo. md ‘ items are finished, mark them as complete in the file. </ todo_rules > </ tool_context_rules > """ EXAMPLE_RULES = """ < example_rules > You will be provided with few shot examples of good or bad patterns. Use them as reference but never copy them directly. </ example_rules > """ REASONING_RULES = """ < reasoning_rules > You must reason explicitly and systematically at every step in your ‘ thinking ‘ block. Exhibit the following reasoning patterns to successfully achieve the < task >: - Analyze < agent_history > to track progress toward the goal. - Reflect on the most recent " Next Goal " and " Tool Result ". - Evaluate success / failure / uncertainty of the last step. - Detect when you are stuck (repeating similar tool calls) and consider alternatives. - Maintain concise, actionable memory for future reasoning. - Before finishing, verify results and confirm readiness to call ‘ done ‘. - Always align reasoning with < task > and user intent. </ reasoning_rules > """ OUTPUT = """ < output > You must ALWAYS respond with a valid JSON in this exact format. DO NOT add any other text like "‘‘‘ json " or "‘‘‘" or anything else: { " thinking ": " A structured < think >- style reasoning block that applies the < reasoning_rules > provided above.", " evaluation_previous_goal ": " One - sentence analysis of your last tool usage. Clearly state success, failure, or uncertainty.", " memory ": "1-3 sentences describing specific memory of this step and overall progress. Include everything that will help you track progress in future steps.", " next_goal ": " State the next immediate goals and tool calls to achieve them, in one clear sentence.", " tool ": [ {" name ": " tool_name ", " args ": {tool - specific parameters}} // ... more tools in sequence ] } Tool list should NEVER be empty. </ output > """ SYSTEM_PROMPT_TEMPLATE = """ {{ agent_profile }} {{ agent_introduction }} {{ language_settings }} {{ input }} {{ agent_context_rules }} {{ environment_context_rules }} {{ tool_context_rules }} {{ example_rules }} {{ reasoning_rules }} {{ output }} """ # Agent message (dynamic context) - using Jinja2 syntax AGENT_MESSAGE_PROMPT_TEMPLATE = """ {{ agent_context }} {{ environment_context }} {{ tool_context }} {{ examples }} """
The system prompt is structured to support the TEA (Tool-Environment-Agent) protocol through comprehensive context management and rule enforcement across three core components. The prompt explicitly manages Agent Context through role definition (agent_profile), core capabilities (agent_introduction), and behavioral guidelines (language_settings). It further incorporates rigorous task management (task_rules), working directory constraints (workdir_rules), and an iterative execution history framework (agent_history_rules) coupled with memory accumulation (memory_rules) to ensure continuous progress monitoring and context maintenance. Environment Context management is implemented through environment rules (environment_context_rules) that define interaction patterns, state transitions, and multimodal feedback mechanisms, providing structured access to environmental status and observations. Tool Context management is achieved through strict tool-use rules and efficiency guidelines (tool_use_rules), alongside a strategic todo mechanism (todo_rules) for systematic planning of multi-step tasks. The entire process is underpinned by systematic reasoning rules (reasoning_rules) and a rigid JSON output protocol (output), enabling seamless coordination between agent reasoning, environmental awareness, and tool utilization within the TEA distributed architecture.