# AgentOrchestra: Orchestrating Multi-Agent Intelligence with the Tool-Environment-Agent(TEA) Protocol
**Authors**:
- Wentao Zhang
- Liang Zeng
- Yuzhen Xiao
- Yongcong Li
- Ce Cui
- Yilei Zhao
- Rui Hu
- Yang Liu
- Yahui Zhou
- Bo An (Skywork AI Nanyang Technological University)
## Abstract
Recent advances in LLM-based agent systems have shown promise in tackling complex, long-horizon tasks. However, existing LLM-based agentprotocols (e.g., A2A and MCP) under-specify cross-entity lifecycle and context management, version tracking, and ad-hoc environment integration, which in turn encourages fixed, monolithic agent compositions and brittle glue code. To address these limitations, we introduce the ToolâEnvironmentâAgent (TEA) protocol, a unified abstraction that models environments, agents, and tools as first-class resources with explicit lifecycles and versioned interfaces. TEA provides a principled foundation for end-to-end lifecycle and version management, and for associating each run with its context and outputs across components, improving traceability and reproducibility. Moreover, TEA enables continual self-evolution of agent-associated components Unless otherwise specified, agent-associated components include prompts, memory/tool/agent/environment code, and agent outputs (solutions). through a closed feedback loop, producing improved versions while supporting version selection and rollback. Building on TEA, we present AgentOrchestra, a hierarchical multi-agent framework in which a central planner orchestrates specialized sub-agents for web navigation, data analysis, and file operations, and supports continual adaptation by dynamically instantiating, retrieving, and refining tools online during execution. We evaluate AgentOrchestra on three challenging benchmarks, where it consistently outperforms strong baselines and achieves 89.04% on GAIA, establishing state-of-the-art performance to the best of our knowledge. Overall, our results provide evidence that TEA and hierarchical orchestration improve scalability and generality in multi-agent systems.
<details>
<summary>x2.png Details</summary>

### Visual Description
Icon/Small Image (32x33)
</details>
AgentOrchestra: Orchestrating Multi-Agent Intelligence with the Tool-Environment-Agent(TEA) Protocol
## 1 Introduction
Recent advances in LLM-based agent systems have enabled strong performance on both general-purpose and complex, long-horizon tasks across diverse domains, including web navigation (OpenAI, 2025b; MĂźller and Ĺ˝uniÄ, 2024), computer use (Anthropic, 2024a; Qin et al., 2025), code execution (Wang et al., 2024a), game playing (Wang et al., 2023; Tan et al., 2024), and research assistance (OpenAI, 2024; DeepMind, 2024; xAI, 2025). Despite this progress, cross-environment generalization remains limited because context is scattered across prompts and logs, environment integration relies on brittle glue code, and agent-associated components are typically fixed rather than feedback-driven self-evolution.
Additionally, current agent protocols fall short of serving as a general substrate for scalable, general-purpose agents. As summarized in Table 1, representative protocols such as Googleâs A2A (Google, 2025) and Anthropicâs MCP (Anthropic, 2024b) provide important building blocks, including task-level collaboration and messaging in A2A, as well as tool and resource schemas, discovery, and invocation in MCP. However, three protocol-level gaps remain: i) Lifecycle and context management are fragmented, as neither standardizes unified primitives to manage lifecycles and maintain consistent, versioned execution context across agent-associated components; ii) Self-evolution is not supported at the protocol level, as both protocols largely treat prompts and resources as externally maintained assets, and do not define a closed loop to refine prompts or tools from execution feedback with traceable versioning; iii) Environments are not first-class, environments are delegated to application-specific runtimes instead of being managed components with clear boundaries and constraints. This makes it difficult to switch agents across environments, reuse environments, and isolate parallel runs, often reducing systems to glue-code orchestration.
Table 1: Comparison of TEA Protocol with A2A and MCP. Symbols: $\checkmark$ = Yes, $\triangle$ = Partial, $\times$ = No.
| Dimension | TEA | A2A | MCP |
| --- | --- | --- | --- |
| Core Entities | Tool, Env, Agent | Agent, Tool | Model |
| Lifecycle & Version | $\checkmark$ | $\times$ | $\times$ |
| Entity Transformations | $\checkmark$ | $\times$ | $\times$ |
| Self-Evolution Support | $\checkmark$ | $\times$ | $\times$ |
| Open Ecosystem | â | $\triangle$ | $\triangle$ |
To address these limitations, we propose the ToolâEnvironmentâAgent (TEA) protocol, which treats environments, agents, and tools as explicitly managed components under a unified protocol layer. Concretely, TEA standardizes component identifiers and version semantics, and binds each run to its context and execution state, so that artifacts remain traceable across iterations. Importantly, TEA goes beyond MCP by standardizing cross entity lifecycle semantics, explicit version semantics with stable entity identifiers, run-indexed context capture, explicit environment boundaries with constraints, and closed loop evolution hooks driven by execution feedback. As a result, execution state, artifacts, and context can be consistently persisted, reused, and traced across runs and iterations. TEA further enables self-evolution by defining a closed loop in which execution feedback can trigger agent-associated components during runtime, with updates recorded as new versions. Finally, TEA models environments as first-class components with explicit boundaries and constraints, for example web sandboxes, file systems, and code execution runtimes, improving reuse and isolation across heterogeneous domains and reducing context leakage in parallel executions. This also encourages consolidating functionally related tools into coherent environments; for example, discrete file operations can be organized as a managed file system, reducing context fragmentation and management overhead. Overall, TEA aims to make agent construction more composable and reproducible in practice. Detailed motivations for the TEA protocol and in-depth comparisons with existing protocols are provided in Appendix A, B.
Based on the TEA protocol, we develop AgentOrchestra, a hierarchical multi-agent framework for general-purpose task solving that integrates high-level planning with modular collaboration. AgentOrchestra uses a central planner to decompose a user objective and delegate sub-tasks to specialized agents for research, web navigation, analysis, tool synthesis, and reporting. Compared to flat coordination, where an orchestrator selects from a growing global pool of agents and tools and tends to accumulate irrelevant context, AgentOrchestra adopts hierarchical delegation with localized tool ownership. The planner routes each sub-task to a domain-specific sub-agent (or environment), which maintains and exposes only a curated toolset and context for its domain. This structure converts global coordination into a sequence of localized routing decisions, enabling tree-structured expansion as new capabilities are added while keeping the orchestratorâs decision scope and context footprint bounded. For example, the planner first selects a domain-level agent, which then supplies only the tools and context required for that domain. Furthermore, AgentOrchestra incorporates a self-evolution module that leverages TEAâs lifecycle and versioning mechanisms to refine agent- associated components based on execution feedback. Our contributions are threefold:
- We introduce the TEA protocol, which unifies environments, agents, and tools as first-class, versioned components with lifecycles to support context management and execution.
- We develop AgentOrchestra, a hierarchical multi-agent system built on TEA, demonstrating scalable orchestration through tree-structured routing and feedback-driven self-evolution.
- We conduct extensive evaluations on three challenging benchmarks, including ablations to isolate the effects of key components. AgentOrchestra consistently outperforms strong baselines and achieves 89.04% on GAIA, establishing state-of-the-art performance to the best of our knowledge.
## 2 Related Work
### 2.1 Tool and Agent Protocols
Recent protocols standardize tool interfaces and agent communication. For instance, MCP (Anthropic, 2024b) unifies tool integration for LLMs, while A2A (Google, 2025) enables agent-to-agent messaging and coordination. Other efforts, such as the Agent Network Protocol (ANP) (Ehtesham et al., 2025) and frameworks like SAFEFLOW (Li et al., 2025), enhance interoperability and safety in multi-agent systems. While these protocols provide essential building blocks, they primarily treat agents and tools as isolated service endpoints, often overlooking environments as dynamic, first-class components. TEA extends these existing standards rather than replacing them. By integrating tools, environments, and agents into a unified context-aware framework, TEA resolves protocol fragmentation with integrated lifecycle and version management missing in MCP or A2A.
### 2.2 General-Purpose Agents
Integrating tools with LLMs represents a paradigm shift, enabling agents to exhibit enhanced flexibility, cross-domain reasoning, and natural language interaction (Liang and Tong, 2025). Such systems have demonstrated efficacy across diverse domains, including web browsing (OpenAI, 2025b; MĂźller and Ĺ˝uniÄ, 2024), computer operation (Anthropic, 2024a; Qin et al., 2025), code execution (Wang et al., 2024a), and game playing (Wang et al., 2023; Tan et al., 2024). Standardized interfaces like OpenAIâs Function Calling and Anthropicâs MCP (OpenAI, 2023; Anthropic, 2024b), alongside frameworks such as ToolMaker (WĂślflein et al., 2025), have further streamlined the synthesis of LLM-compatible tools. Building upon these foundations, multi-agent architectures like MetaGPT (Hong et al., 2023) demonstrate the potential of specialized agent coordination for complex problem-solving. However, many current approaches still struggle with efficient communication, dynamic role allocation, and scalable teamwork. The emergence of generalist frameworks, including Manus (Shen and Yang, 2025), OpenHands (Wang et al., 2024b), and smolagents (Roucher et al., 2025), has advanced unified perception and tool-augmented action. While recent efforts like Alita (Qiu et al., 2025) explore minimal predefinition and maximal self-evolution, these systems often lack unified protocols for cross-layer resource management. This gap motivates our proposal of the TEA Protocol and AgentOrchestra.
## 3 The TEA Protocol
The TEA Protocol is fundamentally designed around coroutine-based asynchronous execution, enabling concurrent task processing and parallel multi-agent coordination. As illustrated in Figure 1, the protocol architecture comprises three primary layers: i) Basic Managers provide foundational services through six specialized components (model, prompt, memory, dynamic, version, and tracer); ii) Core Protocols define the Tool Context Protocol (TCP), Environment Context Protocol (ECP), and Agent Context Protocol (ACP), each implemented through a context manager for lifecycle engineering and a server for standardized orchestration; and iii) Protocol Transformations establish bidirectional conversion pathways (e.g., A2T, E2T, A2E) enabling dynamic role reconfiguration. Additionally, the protocol incorporates a Self-Evolution Module that wraps agent-associated components as evolvable variables for iterative optimization. Details and formalization can be found in Appendix C.
<details>
<summary>x3.png Details</summary>

### Visual Description
\n
## Diagram: Tool-Environment-Agent (TEA) Framework
### Overview
The image depicts a diagram illustrating the Tool-Environment-Agent (TEA) framework. It shows the interactions between Agents, Tools, and Environments, mediated by various context protocols. The diagram is structured with a central TEA component and surrounding elements representing the actors and communication channels.
### Components/Axes
The diagram consists of the following key components:
* **Agents:** Represented by a circular node labeled "Agents" in the center-top of the diagram.
* **Tools:** Represented by a circular node labeled "Tools" in the bottom-left of the diagram.
* **Environments (Envs):** Represented by a circular node labeled "Envs" in the bottom-right of the diagram.
* **TEA:** A central triangular shape labeled "Tool-Environment-Agent (TEA)" containing an image of a coffee cup.
* **Context Protocols:**
* Agent Context Protocol (ACP): A rectangular box above the "Agents" node.
* Tool Context Protocol (TCP): A rectangular box below the "Tools" node.
* Environment Context Protocol (ECP): A rectangular box below the "Envs" node.
* **Actors:** Four individual icons representing roles: Researcher, Analyzer, Browser, and Manager, positioned above the "Agent Context Protocol".
* **Tool Icons:** Icons representing specific tools: Bash, Python, Midify, Text, Web.
* **Environment Icons:** Icons representing environments: Browser, Computer, Mobile Files, Github.
* **Interaction Arrows:** Arrows indicating the flow of information between components, labeled as follows:
* A2T: Agent to Tool
* T2A: Tool to Agent
* E2T: Environment to Tool
* T2E: Tool to Environment
* A2E: Agent to Environment
* E2A: Environment to Agent
### Detailed Analysis or Content Details
The diagram illustrates a cyclical interaction between Agents, Tools, and Environments.
* **Agents:** The "Agents" node is connected to the TEA component via bidirectional arrows labeled A2T and A2E. This suggests agents can both utilize tools and interact directly with environments. The actors above the "Agent Context Protocol" represent different types of agents.
* **Tools:** The "Tools" node is connected to the TEA component via bidirectional arrows labeled T2A and T2E. This indicates tools receive input from agents and provide output to environments. The tool icons at the bottom represent specific tools available.
* **Environments:** The "Environments" node is connected to the TEA component via bidirectional arrows labeled E2T and E2A. This shows environments provide input to tools and receive output from agents. The environment icons at the bottom represent different environments.
* **TEA Component:** The central TEA component acts as a mediator, facilitating the interaction between Agents, Tools, and Environments. The coffee cup image within the TEA component may symbolize the framework's role in "brewing" solutions or facilitating complex processes.
* **Context Protocols:** The context protocols (ACP, TCP, ECP) likely define the rules and standards for communication between the respective components.
### Key Observations
* The bidirectional arrows suggest a dynamic and iterative process where information flows freely between Agents, Tools, and Environments.
* The central TEA component highlights the importance of integrating these three elements for effective operation.
* The specific tools and environments listed provide concrete examples of the framework's applicability.
* The diagram emphasizes the role of context protocols in ensuring seamless communication.
### Interpretation
The TEA framework appears to be a system designed to enable intelligent agents to interact with and utilize various tools and environments. The framework's strength lies in its ability to facilitate a closed-loop system where agents can leverage tools to manipulate environments and receive feedback, allowing for continuous learning and adaptation. The context protocols are crucial for ensuring that these interactions are meaningful and effective. The diagram suggests a flexible and modular architecture, where different agents, tools, and environments can be easily integrated into the framework. The inclusion of specific tools and environments (Bash, Python, Web, Github) indicates the framework's potential applications in areas such as software development, data analysis, and automation. The overall design emphasizes a holistic approach to agent-based systems, recognizing the importance of considering the interplay between agents, tools, and the surrounding environment.
</details>
Figure 1: Architecture of the TEA Protocol.
### 3.1 Basic Managers
The Basic Managers constitute the foundation of the TEA Protocol, providing essential services through six specialized managers: i) the model manager abstracts heterogeneous LLM backends through a unified interface; ii) the prompt manager handles prompt lifecycle and versioning; iii) the memory manager coordinates persistence via session-based concurrency control; iv) the dynamic manager enables runtime code execution and serialization; v) the version manager maintains evolution histories for all components; and vi) the tracer records comprehensive execution trajectories and system-wide telemetry, serving as a data collection engine for audit, debugging, and the synthesis of high-quality datasets for agent training.
### 3.2 Core Protocols
The TEA Protocol defines three core context protocols: the Tool Context Protocol (TCP), the Environment Context Protocol (ECP), and the Agent Context Protocol (ACP). These protocols share a unified architectural design, each implemented through two core components: a context manager for context engineering, lifecycle management, and semantic retrieval, and a server that exposes standardized interfaces to other system modules. Each protocol generates a unified contract document (analogous to Agent Skills (Anthropic, 2025)) that aggregates all registered componentsâ descriptions to facilitate resource discovery and usage.
Tool Context Protocol. TCP fundamentally extends MCP (Anthropic, 2024b) by introducing integrated context engineering and comprehensive lifecycle management. Implemented through a ToolContextManager and a TCPServer, it supports seamless tool loading from both local registries and persistent configurations. During registration, TCP automatically synthesizes multiple representation formats, including function-calling schemas for LLM interfaces, natural language descriptions for documentation, and type-safe argument schemas for validation, providing LLMs with rich semantic information for accurate parameter inference. Furthermore, TCP incorporates a robust versioning system and a semantic retrieval mechanism based on vector embeddings, ensuring that tools can evolve over time while remaining easily discoverable through similarity-based queries.
Environment Context Protocol. ECP addresses the lack of unified interfaces in current agent systems by formalizing computational environments as first-class components with distinct observation and action spaces. Following an architectural pattern similar to TCP, it employs an EnvironmentContextManager to maintain state coherence and manage the contextual execution environments required by tools. ECP automatically discovers and registers environment-specific actions, converting them into standardized interfaces that agents can invoke via action toolkits. This design enables agents to operate across heterogeneous domains, such as browsers or file systems, without bespoke adaptations, while leveraging versioning and semantic retrieval to manage environment-level capabilities.
Agent Context Protocol. ACP establishes a unified framework for the registration, representation, and orchestration of autonomous agents, overcoming the poor interoperability and fragmented attribute definitions in existing multi-agent systems. It utilizes an AgentContextManager to maintain agent states and execution contexts, providing a foundation for persistent coordination across tasks and sessions. ACP captures semantically enriched metadata regarding agentsâ roles, competencies, and objectives, and formalizes the modeling of complex inter-agent dynamics, including cooperative, competitive, and hierarchical configurations. By embedding structured contextual descriptions and maintaining relationship representations, ACP facilitates adaptive collaboration and systematic integration within the broader TEA ecosystem.
### 3.3 Protocol Transformations
While TCP, ECP, and ACP provide independent specifications for tools, environments, and agents, practical deployment requires seamless interoperability across these protocols. Well-defined transformation pathways are essential for enabling computational components to assume alternative roles and exchange contextual information in a principled manner. These transformations constitute the foundation for dynamic role reconfiguration, allowing components to flexibly adapt their functional scope in response to evolving task requirements and system constraints. We identify six fundamental categories of protocol transformations:
- Agent-to-Tool (A2T). Encapsulates an agentâs capabilities and reasoning into a standardized tool interface while preserving awareness. For example, a deep researcher workflow can be packaged as a general-purpose search tool.
- Tool-to-Agent (T2A). Treats tools as operational actuators by mapping an agentâs goals into parameterized tool invocations, aligning reasoning with tool constraints. For example, a data analysis agent may invoke SQL tools to query structured databases.
- Environment-to-Tool (E2T). Converts actions of environments into standardized tool interfaces, enabling agents to interact with environments through consistent tool calls. For example, browser actions such as Navigate and Click can be consolidated into a context-aware toolkit.
- Tool-to-Environment (T2E). Elevates a collection of tools into an environment abstraction where functions become actions within a coherent action space governed by shared state. For example, a development toolkit can be encapsulated as a programming environment for sequential code-edit-compile-debug workflows.
- Agent-to-Environment (A2E). Encapsulates an agent as an interactive environment by exposing its decision rules and state dynamics as an operational context for other agents. For example, a market agent can be represented as an environment that provides trading rules and dynamic responses for training.
- Environment-to-Agent (E2A). Embeds reasoning and adaptive decision-making into an environmentâs dynamics, transforming it into an autonomous agent that can initiate behaviors and enforce constraints. For example, a game environment can be elevated into an opponent agent that adapts its strategy to the playerâs actions.
### 3.4 Self-Evolution Module
The Self-Evolution Module enables agents to continuously improve performance by optimizing system components during task execution. It wraps evolvable components, including prompts, tool/agent/environment/memory code, and successful execution solutions, as variables for iterative optimization. The module employs two primary methods: textgrad (Yuksekgonul et al., 2025) for gradient-based refinement and self-reflection for strategic analysis. Optimized components are automatically registered as new versions via the version manager, ensuring that subsequent tasks leverage improved capabilities while maintaining access to historical records for analysis and rollback.
<details>
<summary>x4.png Details</summary>

### Visual Description
\n
## Diagram: Agent Architecture Overview
### Overview
This diagram illustrates the architecture of an agent system, likely a large language model (LLM) based agent, detailing its components, interactions, and underlying protocols. The diagram is structured around a central "Agent Context Protocol (ACP)" and shows how various agents, tools, and environments interact to achieve user objectives. The diagram is oriented from top to bottom, with user objectives at the top and the underlying infrastructure at the bottom.
### Components/Axes
The diagram is composed of several key sections:
* **Top Section (Planning):** Includes "User Objectives", "Actions", "Planning Agent", "Planning Tool", "Task", "Planner", "Researcher", "Analyzer", "Generator", "Reporter", and "Answer".
* **Middle Section (Agents):** Contains "Deep Researcher Agent", "Browser Use Agent", "Deep Analyzer Agent", and "Deep Generator Agent".
* **Central Section (Protocols):** Features "Tool Context Protocol (TCP)", "Agent Context Protocol (ACP)", and "Environment Context Protocol (ECP)".
* **Bottom Section (Infrastructure):** Includes "Tool-Environment-Agent (TEA)", "Envs", "General Tools", "Basic Managers", and "Self-Evolution Module".
* **Tools:** Icons representing various tools like search, browser, code execution, and writing.
* **Agents:** Represented by cartoon robot icons, each with a specific function.
* **Protocols:** Boxes outlining the communication protocols between components.
* **Arrows:** Indicate the flow of information and control between components.
### Detailed Analysis or Content Details
**Top Section (Planning):**
* **User Objectives:** Input to the "Actions" block.
* **Actions:** Contains icons for "Interpret user tasks", "Decompose into manageable sub-tasks", "Assign sub-agent to specialized sub-tasks", and "Feedback".
* **Planning Agent:** A box stating "Create, update, and manage plans for complex tasks simultaneously. Track execution states".
* **Planning Tool:** Checkmark icon.
* **Task:** Connected to "Planner".
* **Planner:** Connected to "Researcher".
* **Researcher:** Connected to "Analyzer".
* **Analyzer:** Connected to "Generator".
* **Generator:** Connected to "Reporter".
* **Reporter:** Connected to "Answer".
**Middle Section (Agents):**
* **Deep Researcher Agent:** Contains icons for "Optimize Search", "Refine Insight", and "Queries".
* **Browser Use Agent:** Contains icons for "Decide Actions", "Browser Results", "Record Actions", and "Reverse Results".
* **Deep Analyzer Agent:** Contains icons for "Organize Diverse Formats", "Reason and Summarize".
* **Deep Generator Agent:** Contains icons for "Tool Retrieval", "Tool Creation", "Reuse", "Add Content", "Expert Report".
**Central Section (Protocols):**
* **Tool Context Protocol (TCP):** Shows connections to "Bash", "Python", "Searcher", "Analyzer", "Agent Tools", "Web", "Tools", "Local", "Remote".
* **Agent Context Protocol (ACP):** Central hub connecting all agents.
* **Environment Context Protocol (ECP):** Shows connections to "File System", "Browser", "Computer".
**Bottom Section (Infrastructure):**
* **Tool-Environment-Agent (TEA):** Contains "Agents", "Tools", and "Envs".
* **General Tools:** Icons for "Web", "Tools", "Modify".
* **Basic Managers:** Includes "MCP Tool Manager", "Memory Manager", "Prompt Manager", "Dynamic Manager".
* **Self-Evolution Module:** Includes "Version Manager", "Tracer", "Reflection".
**Connections & Flows:**
* User Objectives flow down to Actions, then to the Planning Agent.
* The Planning Agent interacts with the Planning Tool and generates a Task.
* The Task is processed by the Planner, Researcher, Analyzer, Generator, and Reporter, ultimately leading to an Answer.
* The Deep Researcher Agent utilizes search tools.
* The Browser Use Agent interacts with a browser.
* The Deep Analyzer Agent processes information.
* The Deep Generator Agent creates content.
* The TCP manages interactions with tools and environments.
* The ACP facilitates communication between agents.
* The ECP manages interactions with the environment.
* The TEA provides the underlying infrastructure.
* Basic Managers handle core functionalities.
* The Self-Evolution Module enables learning and improvement.
### Key Observations
* The diagram emphasizes a modular and hierarchical architecture.
* The ACP serves as a central communication hub.
* The system incorporates multiple agents, each specializing in a specific task.
* The use of protocols (TCP, ACP, ECP) suggests a well-defined communication framework.
* The inclusion of a Self-Evolution Module indicates a focus on continuous learning and improvement.
* The diagram is visually complex, suggesting a sophisticated system.
* The use of cartoon robot icons for agents adds a friendly and approachable aesthetic.
### Interpretation
The diagram depicts a sophisticated agent system designed to handle complex tasks by breaking them down into smaller, manageable sub-tasks. The system leverages multiple specialized agents, each equipped with specific tools and capabilities. The ACP acts as a central nervous system, coordinating communication and collaboration between agents. The TCP and ECP provide standardized interfaces for interacting with tools and environments, respectively. The Self-Evolution Module suggests that the system is capable of learning from its experiences and improving its performance over time.
The diagram highlights the importance of modularity, communication, and adaptability in building intelligent agent systems. The use of protocols and specialized agents allows for greater flexibility and scalability. The inclusion of a Self-Evolution Module suggests a commitment to continuous improvement and long-term sustainability.
The diagram suggests a system capable of automating complex workflows, providing intelligent assistance, and adapting to changing circumstances. It represents a significant step towards building more powerful and versatile AI agents. The diagram is a high-level overview and does not provide specific details about the implementation of each component. However, it provides a clear and concise representation of the system's overall architecture and functionality.
</details>
Figure 2: Architecture of AgentOrchestra implemented based on TEA protocol.
## 4 AgentOrchestra
AgentOrchestra is a concrete instantiation of the TEA Protocol, designed as a hierarchical multi-agent framework that integrates high-level planning with modular agent collaboration. As illustrated in Figure 2, AgentOrchestra features a central planning agent that decomposes complex objectives and delegates sub-tasks to a team of specialized sub-agents. This section outlines our agent design principles and the architecture of both planning and specialized sub-agents. Details can be found in Appendix D.
### 4.1 Agent Design Principles
Within the TEA Protocol framework, agents are autonomous components that follow a structured interaction model with six core components. i) Agent: Managed via the ACP for registration and coordination. ii) Environment: External context and resources managed by the ECP, exposing unified interfaces for observation and action. iii) Model: LLM reasoning engines abstracted by the Basic Managers for model-agnostic interoperability and dynamic switching. iv) Memory: Session-based persistence that records trajectories and extracts reusable insights. v) Observation: The current context, including tasks, environment states, execution history, and available resources (tools and sub-agents). vi) Action: TCP-managed, executed via parameterized tool calls, where one tool may support multiple actions.
This architectural design facilitates a continuous perceptionâinterpretationâaction cycle. The agent first perceives the current observation and retrieves relevant context from memory. It then interprets this information through the unified model interface to determine the optimal action. The action is executed within the managed environment, and the resulting state transitions and insights are recorded back into memory to refine subsequent reasoning cycles. This iterative loop continues until the task objectives are satisfied or a termination condition is reached. Further details are provided in Appendix D.1.
### 4.2 Planning Agent
The planning agent is the central orchestrator of AgentOrchestra. It interprets the user goal, decomposes it into sub-tasks, and dispatches them to specialized sub-agents or TCP tools via ACP-mediated communication while tracking global progress and consolidating intermediate feedback. To enable principled orchestration, it leverages long-term memory to guide resource selection and dynamically constructs a unified invocation interface, including resources produced through E2T and A2T transformations. Execution follows an iterative loop of interpretation, allocation, and action, with automatic replanning under environment shifts or execution failures. Session management and tracer-based logging provide auditability and support robust long-horizon task completion.
### 4.3 Specialized Sub-Agents
To address diverse real-world challenges, AgentOrchestra instantiates specialized sub-agents tailored for task domains. These sub-agents are managed via the ACP and coordinate through the planning agent to execute complex workflows: i) Deep Researcher Agent: Specialized for comprehensive information gathering through multi-round research workflows. It performs parallel breadth-first searches across multiple engines and recursively issues follow-up queries until task objectives are satisfied, producing relevance-ranked, source-cited summaries. ii) Browser Use Agent: Provides automated, fine-grained web interaction by integrating both browser and computer environments under the ECP. It supports DOM-level and pixel-level operations (e.g., mouse movements), achieving unified control over interactive elements. iii) Deep Analyzer Agent: A workflow-oriented module designed for multi-step reasoning on heterogeneous multimodal data (e.g., text, PDFs, images, audio, video or zip). It applies type-specific analysis strategies and iterative refinement to synthesize insights into coherent conclusions. iv) Tool Generator Agent: Facilitates intelligent tool evolution through the automated creation, retrieval, and systematic reuse of TCP-compliant tools. It employs semantic search to identify tools and initiates a code synthesis process to develop new capabilities when gaps are identified. v) Reporter Agent: It aggregates and harmonizes evidence collected by upstream agents (e.g., the Deep Researcher Agent, Browser Use Agent, and Deep Analyzer Agent), then composes structured markdown with automatically deduplicated references and normalized URLs for consistent source attribution.
## 5 Empirical Studies
This section presents our experimental setup and results, including benchmark evaluations, baseline comparisons, and comprehensive analysis. Additional examples are provided in the Appendix F.
Experimental Settings. We evaluate our framework on three benchmarks: SimpleQA Wei et al. (2024), a 4,326-question factual accuracy benchmark; GAIA Mialon et al. (2023), assessing real-world reasoning, multimodal processing, and tool use with 301 test and 165 validation questions; and Humanityâs Last Exam (HLE) Phan et al. (2025), a 2,500-question multimodal benchmark for human-level reasoning and general intelligence. We report score (pass@1), which measures the proportion of questions for which the top prediction is fully correct. Specifically, the planning agent ( $m{=}50$ ), deep researcher ( $m{=}3$ ), tool generator ( $m{=}10$ ), deep analyzer ( $m{=}3$ ), and reporter are all built on gemini-3-flash-preview; the browser use agent employs gpt-4.1 ( $m{=}5$ ) and computer-use-preview(4o) ( $m{=}50$ ), where $m$ denotes the maximum steps.
### 5.1 Performance across Benchmarks
<details>
<summary>x5.png Details</summary>

### Visual Description
## Bar Chart: GAIA Test Scores
### Overview
The image presents a bar chart displaying scores from a "GAIA Test" for various models. The chart compares scores across three levels (Level1, Level2, Level3) and an average score. The x-axis represents different models, and the y-axis represents the score, ranging from approximately 40 to 100.
### Components/Axes
* **Title:** GAIA Test
* **Y-axis Label:** Score
* **X-axis Labels:** AgentOrchestrator, Halo, BokEQ-Arena, Avoria, h2oGPT-Arena, Desearch, of-Align, od-Align, AgentOrchestrator, Halo, h2oGPT-Arena, Avoria, Desearch, Llama-DR, Llama-UR, AgentOrchestrator, Halo, h2oGPT-Arena, Desearch, Llama-DR, Llama-UR
* **Legend:**
* Level1 (Green)
* Level2 (Blue)
* Level3 (Purple)
* Average (Orange)
### Detailed Analysis
The chart consists of grouped bars for each model, representing the scores for Level1, Level2, Level3, and the average.
**AgentOrchestrator:**
* Level1: Approximately 99.8
* Level2: Approximately 96.3
* Level3: Approximately 93.7
* Average: Approximately 88.4
**Halo:**
* Level1: Approximately 98.8
* Level2: Approximately 94.8
* Level3: Approximately 91.4
* Average: Approximately 86.9
**BokEQ-Arena:**
* Level1: Approximately 97.9
* Level2: Approximately 93.8
* Level3: Approximately 89.9
* Average: Approximately 85.3
**Avoria:**
* Level1: Approximately 97.0
* Level2: Approximately 92.9
* Level3: Approximately 88.9
* Average: Approximately 84.4
**h2oGPT-Arena:**
* Level1: Approximately 96.1
* Level2: Approximately 92.0
* Level3: Approximately 88.0
* Average: Approximately 83.5
**Desearch:**
* Level1: Approximately 95.2
* Level2: Approximately 91.1
* Level3: Approximately 87.1
* Average: Approximately 82.6
**of-Align:**
* Level1: Approximately 94.3
* Level2: Approximately 90.2
* Level3: Approximately 86.2
* Average: Approximately 81.7
**od-Align:**
* Level1: Approximately 93.4
* Level2: Approximately 89.3
* Level3: Approximately 85.3
* Average: Approximately 80.8
**Second AgentOrchestrator Group:**
* Level1: Approximately 88.4
* Level2: Approximately 84.4
* Level3: Approximately 80.4
* Average: Approximately 75.9
**Second Halo Group:**
* Level1: Approximately 87.5
* Level2: Approximately 83.5
* Level3: Approximately 79.5
* Average: Approximately 75.0
**Second h2oGPT-Arena Group:**
* Level1: Approximately 86.6
* Level2: Approximately 82.6
* Level3: Approximately 78.6
* Average: Approximately 74.1
**Second Avoria Group:**
* Level1: Approximately 85.7
* Level2: Approximately 81.7
* Level3: Approximately 77.7
* Average: Approximately 73.2
**Second Desearch Group:**
* Level1: Approximately 84.8
* Level2: Approximately 80.8
* Level3: Approximately 76.8
* Average: Approximately 72.3
**Llama-DR:**
* Level1: Approximately 83.9
* Level2: Approximately 79.9
* Level3: Approximately 75.9
* Average: Approximately 71.4
**Llama-UR:**
* Level1: Approximately 83.0
* Level2: Approximately 79.0
* Level3: Approximately 75.0
* Average: Approximately 70.5
**Third AgentOrchestrator Group:**
* Level1: Approximately 82.1
* Level2: Approximately 78.1
* Level3: Approximately 74.1
* Average: Approximately 69.6
**Third Halo Group:**
* Level1: Approximately 81.2
* Level2: Approximately 77.2
* Level3: Approximately 73.2
* Average: Approximately 68.7
**Third h2oGPT-Arena Group:**
* Level1: Approximately 80.3
* Level2: Approximately 76.3
* Level3: Approximately 72.3
* Average: Approximately 67.8
**Third Desearch Group:**
* Level1: Approximately 79.4
* Level2: Approximately 75.4
* Level3: Approximately 71.4
* Average: Approximately 66.9
**Third Llama-DR Group:**
* Level1: Approximately 78.5
* Level2: Approximately 74.5
* Level3: Approximately 70.5
* Average: Approximately 66.0
**Third Llama-UR Group:**
* Level1: Approximately 77.6
* Level2: Approximately 73.6
* Level3: Approximately 69.6
* Average: Approximately 65.1
### Key Observations
* The scores generally decrease as you move from Level1 to Level2 to Level3, and the average score is consistently lower than all three levels.
* AgentOrchestrator, Halo, BokEQ-Arena, and Avoria consistently achieve the highest scores across all levels.
* Llama-UR consistently achieves the lowest scores across all levels.
* There are three distinct groupings of models, with a noticeable drop in scores between each group.
### Interpretation
The chart demonstrates the performance of different models on the GAIA test across three levels of difficulty. The consistent ranking of models suggests inherent differences in their capabilities. The decreasing scores from Level1 to Level3 and the lower average scores indicate that the test becomes more challenging with each level, and the models' performance degrades accordingly. The grouping of models suggests that there are tiers of performance, with some models significantly outperforming others. The large gap between the first and last groups suggests a substantial difference in the underlying technology or training data used for these models. The average score provides a baseline for comparison, highlighting which models exceed or fall below the overall performance level. The data suggests that AgentOrchestrator, Halo, BokEQ-Arena, and Avoria are the most robust models tested, while Llama-UR requires further improvement.
</details>
Figure 3: GAIA Test Results.
GAIA. AgentOrchestra achieves state-of-the-art performance (89.04% avg.) by mitigating the dimensionality curse and semantic drift that arise in large-scale agentic planning. We attribute this success to two architectural properties enabled by the TEA Protocol. First, hierarchical decoupling of the action space reduces planning complexity: while methods (e.g., ToolOrchestra, AWorld) must map goals to a monolithic toolkit, our hierarchical routing decomposes the global task into locally tractable sub-problems, lowering cognitive entropy for the central orchestrator and preserving abstract reasoning under long horizons, even amid low-level sensorimotor noise (e.g., granular DOM events). Second, ECP formalizes epistemic environment boundaries: GAIAâs multi-domain tasks require temporal and cross-modal state coherence, and baselines often degrade during domain transitions, such as from browser retrieval to local python analysis. By treating environments as first-class managed components, TEA preserves and propagates session-critical state (e.g., authentication tokens and transient file-system mutations) across agent boundaries, reducing contextual forgetting and enabling compositional generalization on challenging Level 2 and Level 3 scenarios. Third, AgentOrchestra supports recursive refinement of reasoning trajectories. When faced with complex problems, the Planning Agent evaluates intermediate insights and, when necessary, invokes the Tool Generator Agent to synthesize context-specific functionalities on the fly. This on-demand tool evolution bypasses the fixed-capability bottleneck of static agent components.
Table 2: Performance on GAIA Validation.
| Agents | Level 1 | Level 2 | Level 3 | Average |
| --- | --- | --- | --- | --- |
| HF ODR (o1) (HuggingFace, 2024) | 67.92 | 53.49 | 34.62 | 55.15 |
| OpenAI DR (OpenAI, 2024) | 74.29 | 69.06 | 47.60 | 67.36 |
| Manus (Shen and Yang, 2025) | 86.50 | 70.10 | 57.69 | 73.90 |
| Langfun (Google, 2024) | 86.79 | 76.74 | 57.69 | 76.97 |
| AWorld (Yu et al., 2025) | 88.68 | 77.91 | 53.85 | 77.58 |
| AgentOrchestra | 92.45 | 83.72 | 57.69 | 82.42 |
Table 3: Performance on SimpleQA and HLE.
| Model and Agent | SimpleQA |
| --- | --- |
| Models | |
| o3 (w/o tools) | 49.4 |
| gemini-2.5-pro-preview-05-06 | 50.8 |
| Agents | |
| Perplexity DR (Perplexity, 2025) | 93.9 |
| AgentOrchestra | 95.3 |
| Model and Agent | HLE |
| Models | |
| o3 (w/o tools) | 20.3 |
| claude-3.7-sonnet (w/o tools) | 8.9 |
| gemini-2.5-pro-preview-05-06 | 17.8 |
| Agents | |
| OpenAI DR (OpenAI, 2024) | 26.6 |
| Perplexity DR (Perplexity, 2025) | 21.1 |
| AgentOrchestra | 37.46 |
SimpleQA. AgentOrchestra achieves SOTA performance (95.3% accuracy), significantly surpassing both monolithic LLMs (e.g., o3 at 49.4%) and specialized retrieval agents like Perplexity Deep Research (93.9%). We attribute this improvement to systematic reduction of epistemic uncertainty through our hierarchical verification pipeline. SimpleQA primarily targets short-form factuality, where hallucinations often arise from the modelâs inability to reconcile conflicting web-based evidence or its tendency to rely on internal parametric memory. AgentOrchestra mitigates these issues by enforcing cross-agent consensus: the Planning Agent orchestrates a retrieve-verify-synthesize cycle where the Deep Researcher performs multi-engine breadth-first searches while the Deep Analyzer evaluates evidence consistency across heterogeneous sources. By decoupling retrieval from analysis, the system prevents "confirmation bias" inherent in single-agent architectures, where the same model both proposes and validates a hypothesis. Furthermore, the integration with the Reporter Agent ensures traceable attribution, grounding every factual claim in a re-verified source, which effectively transforms the task from an open-domain generation problem into a structured evidence-synthesis process.
HLE. AgentOrchestra achieves 37.46% on the HLE benchmark, a substantial margin over leading baselines like o3 (20.3%) and Perplexity Deep Research (21.1%). This gain highlights the frameworkâs capacity for long-horizon analytical reasoning and adaptive capability expansion in expert-level domains. HLE demands more than simple retrieval; it requires synthesizing disparate, highly specialized knowledge. In this setting, the hierarchical structure enables strategic pruning of the hypothesis space, allowing the Planning Agent to maintain global objective coherence while delegating technical validation to specialized agents such as the Deep Analyzer. As a result, the final solution is both analytically rigorous and cross-verified against multimodal evidence, yielding robust performance on challenging expert-level tasks.
### 5.2 Ablation Studies
Table 4: Sub-agent effectiveness across GAIA Test.
| P | R | B | A | T | Level 1 | Level 2 | Level 3 | Average | Improvement |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
| â | | | | | 54.84 | 33.96 | 10.20 | 36.54 | â |
| â | â | | | | 86.02 | 47.17 | 34.69 | 57.14 | +56.40% |
| â | â | â | | | 89.25 | 71.07 | 46.94 | 72.76 | +27.33% |
| â | â | â | â | | 91.40 | 77.36 | 61.22 | 79.07 | +8.67% |
| â | â | â | â | â | 98.92 | 85.53 | 81.63 | 89.04 | +12.61% |
Effectiveness of the specialized sub-agents. Ablation studies on the GAIA Test demonstrate the synergistic effect inherent in our multi-agent coordination. Integrating coarse-grained exploratory retrieval (Researcher) with fine-grained operational interaction (Browser) nearly doubles performance (36.54% to 72.76%), proving that breadth of information and depth of interaction are mutually reinforcing. The Deep Analyzerâs 8% gain highlights the necessity of specialized reasoning pipelines for high-entropy multimodal tasks, while the Tool Generatorâs 12.61% boost validates the efficacy of on-demand capability synthesis in overcoming the limitations of static, predefined toolsets. These results suggest that complex problem-solving emerges not just from individual agent strength, but from the structured delegation of specialized roles.
Efficiency analysis. AgentOrchestra âs operational efficiency is evaluated across varying task complexities. Simple tasks typically complete within 30 seconds using approximately 5k tokens, while medium-complexity tasks average 3 minutes (25k tokens). Complex multimodal or long-horizon scenarios require approximately 10 minutes and 100k tokens. Compared to monolithic baselines, our hierarchical architecture optimizes resource allocation, maintaining operational costs comparable to commercial research agents while delivering significant performance gains.
Effectiveness of the self-evolution module. The TEA Protocol enables self-optimization by treating system components as evolvable variables, helping bridge the gap between base model capacity and task requirements. Evaluations on GPQA-Diamond and AIME benchmarks show that iterative refinement, including gradient-based (TextGrad) and symbolic (self-reflection) approaches, mitigates reasoning bottlenecks in foundation models. The improvement is exemplified by a 13.34% gain on AIME25 for gpt-4.1 under self-reflection, highlighting recursive trajectory refinement. Leveraging execution feedback via TEAâs versioning and tracer mechanisms, the system identifies and corrects logical inconsistencies in its planning. Overall, this shifts reasoning from one-shot inference to a managed optimization process, enabling AgentOrchestra to evolve problem-solving strategies for frontier-level tasks.
Table 5: Effectiveness of the self-evolution module. Direct means using the base model directly.
| Strategy | GPQA-Diamond | AIME24 | AIME25 |
| --- | --- | --- | --- |
| Base Model: gpt-4o | | | |
| Direct | 47.98% | 13.34% | 6.67% |
| w/ TextGrad | 54.04% | 10.00% | 10.00% |
| w/ Self-reflection | 55.05% | 20.00% | 6.67% |
| Base Model: gpt-4.1 | | | |
| Direct | 61.11% | 23.34% | 20.00% |
| w/ TextGrad | 65.15% | 26.67% | 23.34% |
| w/ Self-reflection | 68.18% | 33.34% | 33.34% |
Regarding tool evolution, the tool generator agent demonstrates efficient creation and reuse capabilities within the TCP framework. During our evaluation, the agent autonomously generated over 50 specialized tools, achieving a 30% reuse rate across subsequent tasks. This indicates an effective balance between tool specialization and generalization, ensuring that the systemâs capabilities expand adaptively while maintaining resource efficiency.
## 6 Conclusion
We introduced the TEA Protocol, unifying environments, agents, and tools to address fragmentation in existing standards. Building on this, we presented AgentOrchestra, a hierarchical multi-agent framework with specialized sub-agents for planning, research, web interaction, and multimodal analysis. Evaluations on three benchmarks show that AgentOrchestra achieves SOTA performance and scalable orchestration through dynamic resource transformations. Future work will extend TEA to support dynamic role allocation and autonomous agent reconfiguration. Building on tool and solution evolution, we will pursue deeper self-evolution, such as using RL to optimize agent components and decision policies without fine-tuning LLM parameters. We also aim to expand these mechanisms to agent structures and communication protocols, while enhancing multimodal capabilities for fine-grained real-time video analysis.
## 7 Limitations
### 7.1 Limitations of TEA Protocol and AgentOrchestra
Despite its strengths in orchestrating multi-agent systems, AgentOrchestra has several limitations that provide directions for future research:
First, System Complexity and Learning Curve. The TEA protocol introduces a structured abstraction layer for tools, environments, and agents to ensure interoperability. However, this structure may present a steeper learning curve for developers compared to simpler, ad-hoc scripting methods. To address this, we will provide extensive documentation, interactive tutorials, and a variety of pre-configured templates to simplify the onboarding process.
Second, Communication and Execution Overhead. Standardizing interactions through a formal protocol can introduce marginal computational and communication overhead, potentially increasing latency in real-time applications. We plan to optimize the serialization protocols and explore asynchronous execution models to minimize these effects in future versions.
Third, Dependence on Underlying Model Capabilities. The effectiveness of the orchestration is inherently limited by the reasoning and instruction-following performance of the foundation LLMs used. While TEA provides a robust framework, it cannot fully compensate for failures caused by model hallucinations or poor tool-use logic. Future work will focus on developing model-agnostic error recovery strategies and more sophisticated validation layers to enhance system-wide resilience.
### 7.2 Potential Risks
While AgentOrchestra and the TEA protocol aim to enhance multi-agent productivity, their capability to interact with local environments and web browsers introduces certain ethical and security risks.
One primary concern is the Misuse for Malicious Automation. The frameworkâs flexibility in controlling browser sessions and executing terminal commands could be repurposed to develop unauthorized "plugins" or "cheats" for online platforms, leading to unfair advantages or automated fraud. Furthermore, there are significant Privacy and Security Risks associated with granting autonomous agents access to personal data or sensitive system resources. If not properly sandboxed or governed by strict security policies, an agent could inadvertently leak private information or perform harmful, irreversible system actions. To mitigate these risks, we emphasize that AgentOrchestra should be used within isolated, monitored environments, and we advocate for the integration of robust human-in-the-loop verification mechanisms and strict access control policies in any real-world deployment.
## References
- Anthropic (2024a) Introducing Computer Use, a New Claude 3.5 Sonnet, and Claude 3.5 Haiku. Note: https://www.anthropic.com/news/3-5-models-and-computer-use Accessed: 2025-05-13 Cited by: §1, §2.2.
- Anthropic (2024b) Introducing the Model Context Protocol. Note: https://www.anthropic.com/news/model-context-protocol Cited by: §C.2.1, §D.1, §1, §2.1, §2.2, §3.2.
- Anthropic (2025) Equipping agents for the real world with Agent Skills. Note: https://www.anthropic.com/engineering/equipping-agents -for-the-real-world-with-agent-skills Cited by: §C.2, §3.2.
- K. Black, N. Brown, J. Darpinian, K. Dhabalia, D. Driess, A. Esmail, M. Equi, C. Finn, N. Fusai, M. Y. Galliker, et al. (2025) $\pi$ 0. 5: a vision-language-action model with open-world generalization. arXiv preprint arXiv:2504.16054. Cited by: §A.1.2.
- G. Brockman, V. Cheung, L. Pettersson, J. Schneider, J. Schulman, J. Tang, and W. Zaremba (2016) Openai gym. arXiv preprint arXiv:1606.01540. Cited by: §C.2.2.
- G. DeepMind (2024) Gemini Deep Research. Note: https://gemini.google/overview/deep-research/?hl=en Cited by: §1.
- A. Ehtesham, A. Singh, G. K. Gupta, and S. Kumar (2025) A survey of agent interoperability protocols: Model context protocol (mcp), agent communication protocol (acp), agent-to-agent protocol (a2a), and agent network protocol (anp). arXiv preprint arXiv:2505.02279. Cited by: §2.1.
- Google (2024) LangFun Agent. Note: https://github.com/google/langfun Cited by: Table 2.
- Google (2025) Announcing the Agent2Agent Protocol (A2A). Note: https://developers.googleblog.com/en/a2a-a-new-era-of-agent-interoperability/ Cited by: §C.2.3, §1, §2.1.
- S. Hong, X. Zheng, J. Chen, Y. Cheng, J. Wang, C. Zhang, Z. Wang, S. K. S. Yau, Z. Lin, L. Zhou, et al. (2023) MetaGPT: Meta Programming for Multi-agent Collaborative Framework. arXiv preprint arXiv:2308.00352 3 (4), pp. 6. Cited by: §2.2.
- HuggingFace (2024) Open-source DeepResearch - Freeing Our Search Agents. Note: https://huggingface.co/blog/open-deep-research Cited by: Table 2.
- P. Li, X. Zou, Z. Wu, R. Li, S. Xing, H. Zheng, Z. Hu, Y. Wang, H. Li, Q. Yuan, et al. (2025) Safeflow: A principled protocol for trustworthy and transactional autonomous agent systems. arXiv preprint arXiv:2506.07564. Cited by: §2.1.
- G. Liang and Q. Tong (2025) LLM-Powered AI Agent Systems and Their Applications in Industry. arXiv preprint arXiv:2505.16120. Cited by: §2.2.
- X. Liang, J. Xiang, Z. Yu, J. Zhang, S. Hong, S. Fan, and X. Tang (2025) OpenManus: An Open-Source Framework for Building General AI Agents. Zenodo. External Links: Document, Link Cited by: §D.1.
- G. Mialon, C. Fourrier, C. Swift, T. Wolf, Y. LeCun, and T. Scialom (2023) GAIA: A Benchmark for General AI Assistants. External Links: 2311.12983, Link Cited by: §5.
- M. MĂźller and G. Ĺ˝uniÄ (2024) Browser Use: Enable AI to Control Your Browser External Links: Link Cited by: §1, §2.2.
- OpenAI (2023) Function Calling. Note: https://platform.openai.com/docs/guides/function-calling Cited by: §D.1, §2.2.
- OpenAI (2024) Introducing Deep Research. Note: https://openai.com/index/introducing-deep-research Cited by: §1, Table 2, Table 3.
- OpenAI (2025a) Context-Free Grammar. Note: https://platform.openai.com/docs/guides/function-calling#page-top Cited by: §A.1.2.
- OpenAI (2025b) Introducing Operator. Note: https://openai.com/blog/operator Cited by: §1, §2.2.
- Perplexity (2025) Introducing Perplexity Deep Research. Note: https://www.perplexity.ai/hub/blog/introducing-perplexity-deep-research Cited by: Table 3, Table 3.
- L. Phan, A. Gatti, Z. Han, N. Li, J. Hu, H. Zhang, C. B. C. Zhang, M. Shaaban, J. Ling, S. Shi, et al. (2025) Humanityâs Last Exam. arXiv preprint arXiv:2501.14249. Cited by: §5.
- Y. Qin, Y. Ye, J. Fang, H. Wang, S. Liang, S. Tian, J. Zhang, J. Li, Y. Li, S. Huang, et al. (2025) UI-TARS: Pioneering Automated GUI Interaction with Native Agents. arXiv preprint arXiv:2501.12326. External Links: Link Cited by: §1, §2.2.
- J. Qiu, X. Qi, T. Zhang, X. Juan, J. Guo, Y. Lu, Y. Wang, Z. Yao, Q. Ren, X. Jiang, X. Zhou, D. Liu, L. Yang, Y. Wu, K. Huang, S. Liu, H. Wang, and M. Wang (2025) Alita: generalist agent enabling scalable agentic reasoning with minimal predefinition and maximal self-evolution. External Links: 2505.20286, Link Cited by: §2.2.
- A. Roucher, A. V. del Moral, T. Wolf, L. von Werra, and E. Kaunismäki (2025) smolagents: A Smol Library to Build Great Agentic Systems. Note: https://github.com/huggingface/smolagents Cited by: §D.1, §2.2.
- M. Shen and Q. Yang (2025) From Mind to Machine: The Rise of Manus AI as a Fully Autonomous Digital Agent. External Links: 2505.02024, Link Cited by: §2.2, Table 2.
- W. Tan, W. Zhang, X. Xu, H. Xia, Z. Ding, B. Li, B. Zhou, J. Yue, J. Jiang, Y. Li, et al. (2024) Cradle: Empowering Foundation Agents toward General Computer Control. arXiv preprint arXiv:2403.03186. Cited by: §1, §2.2.
- G. Wang, Y. Xie, Y. Jiang, A. Mandlekar, C. Xiao, Y. Zhu, L. Fan, and A. Anandkumar (2023) Voyager: An Open-Ended Embodied Agent with Large Language Models. arXiv preprint arXiv:2305.16291. Cited by: §1, §2.2.
- X. Wang, Y. Chen, L. Yuan, Y. Zhang, Y. Li, H. Peng, and H. Ji (2024a) Executable Code Actions Elicit Better LLM Agents. External Links: 2402.01030, Link Cited by: §1, §2.2.
- X. Wang, B. Li, Y. Song, F. F. Xu, X. Tang, M. Zhuge, J. Pan, Y. Song, B. Li, J. Singh, et al. (2024b) OpenHands: An Open Platform for AI Software Developers as Generalist Agents. In The Thirteenth International Conference on Learning Representations, Cited by: §D.1, §2.2.
- J. Wei, N. Karina, H. W. Chung, Y. J. Jiao, S. Papay, A. Glaese, J. Schulman, and W. Fedus (2024) Measuring Short-Form Factuality in Large Language Models. External Links: 2411.04368, Link Cited by: §5.
- G. WĂślflein, D. Ferber, D. Truhn, O. ArandjeloviÄ, and J. N. Kather (2025) LLM Agents Making Agent Tools. arXiv preprint arXiv:2502.11705. Cited by: §2.2.
- xAI (2025) Grok 3 Beta â The Age of Reasoning Agents. Note: https://x.ai/news/grok-3 Cited by: §1.
- C. Yu, S. Lu, C. Zhuang, D. Wang, Q. Wu, Z. Li, R. Gan, C. Wang, S. Hou, G. Huang, W. Yan, L. Hong, A. Xue, Y. Wang, J. Gu, D. Tsai, and T. Lin (2025) AWorld: orchestrating the training recipe for agentic ai. External Links: 2508.20404, Link Cited by: Table 2.
- M. Yuksekgonul, F. Bianchi, J. Boen, S. Liu, P. Lu, Z. Huang, C. Guestrin, and J. Zou (2025) Optimizing generative AI by backpropagating language model feedback. Nature 639 (8055), pp. 609â616. Cited by: §C.4, §C.4, §3.4.
## Appendix A Comprehensive Motivation for TEA Protocol
This section provides a comprehensive motivation for the TEA Protocol by examining the fundamental relationships and transformations between agents, environments, and tools in multi-agent systems. The discussion is organized into two main parts: first, we explore the conceptual relationships between agents, environments, and tools, examining how these three fundamental components interact and complement each other in modern AI systems; second, we analyze why transformation relationships between these components are necessary, demonstrating the need for their conversion and integration through the TEA Protocol to create a unified, flexible framework for general-purpose task solving.
### A.1 Conceptual Relationships
#### A.1.1 Environment
The environment constitutes one of the fundamental components of multi-agent systems, providing the external stage upon which agents perceive, act, and accomplish tasks. Within the context of the TEA Protocol, highlighting the role of environments is crucial, since environments not only define the operational boundaries of agents but also exhibit complex structural and evolutionary properties. In what follows, we outline the motivation for explicitly modeling environments in the TEA framework from several perspectives.
Classification of environments. From a broad perspective, environments can be divided into two categories: the real world and the virtual world. The real world is concrete and directly perceivable by humans, such as kitchens, offices, or factories. By contrast, the virtual world cannot be directly perceived or objectively described by humans, including domains such as the network world, simulation platforms, and game worlds. Importantly, these two types of environments are not independent. Rather, they are tightly coupled through physical carriers, such as computers, displays, keyboards, mice, and sensors, which act as mediators that enable the bidirectional flow of information between the real and virtual domains. Hence, environments should be regarded not as isolated domains but as interdependent layers connected through mediating carriers.
Nested and expandable properties. Environments are inherently nested and expandable. For example, when an individual is situated in a kitchen, their observable range and available tools are restricted to kitchen-related objects such as faucets, knives, and microwaves, all governed by the local rules of that sub-environment. When the activity range extends to the living room, new objects such as televisions, remote controls, and chairs become accessible, while the kitchen remains embedded as a sub-environment within a broader space. Furthermore, environments can interact with one another, as when a bottle of milk is taken from the kitchen to the living room. This demonstrates that enlarged environments can be conceptualized not merely as simple unions, but rather as structured integrations of the state and action spaces of smaller constituent environments, where local rules and affordances are preserved while new forms of interaction emerge from their composition.
Relationship with stateâaction spaces. In reinforcement learning, environments are formalized in terms of state and action spaces. The state space comprises the set of possible environmental states, represented in modalities such as numerical values, text, images, or video. The action space denotes the set of operations available to agents, generally divided into continuous and discrete spaces. Real and virtual environments are naturally continuous, but discrete abstractions are often extracted for the sake of tractability, forming the basis of most reinforcement learning systems. However, this discretization constrains the richness of interaction. In contrast, large language models (LLMs) enable a new paradigm: instead of selecting from a discrete set, LLMs can generate natural language descriptions that encode complex action sequences. These outputs can be understood as an intermediate representation between continuous and discrete action spaces, richer and more expressive than discrete actions, yet still mappable to concrete operations in continuous environments. To realize this mapping, intermediate actions are required as bridges. For instance, the natural language command âboil waterâ can be decomposed into executable steps such as turning on the kettle, filling it with water, powering it on, and waiting until boiling. This property indicates that LLM-driven interaction expands the definition of action representations and broadens the scope of environmental engagement.
Mediation and interaction. The notion of mediation highlights that environments are not static backdrops but relative constructs whose boundaries depend on available carriers and interfaces. In hybrid physicalâvirtual systems, for example, Internet-of-Things (IoT) devices serve as mediators: a smart refrigerator in the physical world can be controlled through a mobile application in the virtual world, while the application itself is subject to network protocols. Consequently, the definition of an environment is dynamic and conditioned by interactional means. In the TEA Protocol, this mediation must be explicitly modeled, since it determines accessibility and interoperability across environments.
Toward intelligent environments. Traditionally, environments are passive components that provide states and respond to actions. However, as embedded simulators, interfaces, and actuators grow more sophisticated, environments may gradually acquire semi-agentic properties. For instance, a smart home environment may not only respond to the low-level command âturn on the lightâ but also understand and execute a high-level instruction such as âcreate a comfortable atmosphere for reading,â by autonomously adjusting lighting, curtains, and background music. This trend suggests that environments are evolving from passive contexts into adaptive and cooperative components.
In conclusion, the environment should not be regarded as a passive backdrop for agent activity, but as a dynamic and evolving component that fundamentally shapes the scope and feasibility of interaction. Its dual nature across real and virtual domains, its nested and compositional structure, and its formalization through stateâaction spaces all demonstrate that environments provide both the constraints and the affordances within which agents operate. At the same time, the rise of LLM-based agents introduces new forms of action representation that require environments to support more flexible, language-driven interfaces. Looking ahead, as environments increasingly incorporate adaptive and semi-agentic features, their role in task execution will only become more central. Within the TEA Protocol, this motivates treating environments as a co-equal pillar alongside agents and tools, ensuring that general-purpose task solving remains both grounded in environmental constraints and empowered by environmental possibilities.
#### A.1.2 Agent
Within the TEA Protocol, the motivation for treating agents as a core component alongside environments and tools extends beyond mere terminological convenience. Agents represent the indispensable connective tissue between the generative capabilities of LLMs, the operational affordances of tools, and the structural dynamics of environments. While environments provide the stage on which tasks unfold and tools extend the range of possible actions, it is agents that unify perception, reasoning, and execution into coherent task-solving processes. Without explicitly recognizing agents as an independent pillar, the TEA Protocol would lack a systematic way to explain how abstract linguistic outputs can be transformed into grounded operations, how tools can be selected and orchestrated, and how autonomy, memory, and adaptivity emerge in multi-agent systems. The following dimensions illustrate why agents must be elevated to a core component of the framework.
Necessity of environment interaction. Unlike large language models (LLMs), which only produce textual descriptions that require conversion into executable actions, agents are fundamentally characterized by their ability to directly interact with environments. While LLMs can generate detailed plans, instructions, or hypotheses, such outputs remain inert unless they are translated into concrete operations that affect the state of an environment. This gap between symbolic reasoning and actionable execution highlights the necessity of an intermediate entity capable of grounding abstract instructions into domain-specific actions. Agents fulfill precisely this role: they map language-level reasoning to executable steps, whether in physical settings, such as controlling robotic arms or sensors, or in virtual contexts, such as interacting with databases, APIs, or software systems.
By serving as this mapping layer, agents enable the closure of full task loops, where perception leads to reasoning, reasoning produces plans, and plans culminate in actions that in turn modify the environment. Without explicitly modeling agents, the process would remain incomplete, as LLMs alone cannot guarantee the translation of reasoning into operational change. Within the TEA Protocol, this necessity justifies the elevation of agents to a core component: they provide the indispensable interface that connects the generative capacities of LLMs with the affordances and constraints of environments, ensuring that tasks are not only conceived but also carried through to completion.
The decisive role of non-internalizable tools. The fundamental distinction between LLMs and agents lies in whether they can effectively employ tools that cannot be internalized into model parameters. Some tools can indeed be absorbed into LLMs, particularly those whose logic can be fully simulated in symbolic space, whose inputs and outputs are representable in language or code, and whose patterns fall within the training distribution (for example, mathematical reasoning, structured text formatting, code generation, and debugging). For example, early LLMs struggled with JSON output formatting and code reasoning, often requiring external correction or checking tools, but reinforcement learning (RL) and supervised fine-tuning (SFT) have progressively enabled such capabilities to be internalized.
In contrast, many tools remain non-internalizable because they are intrinsically tied to environmental properties. These include tools that depend on physical devices such as keyboards, mice, and robotic arms, external infrastructures such as databases and APIs, or proprietary software governed by rigid protocols. Two recent approaches further illustrate this limitation. Vision-language-action (VLA) (Black et al., 2025) models map perceptual inputs directly into actions, which may appear to bypass intermediate symbolic descriptions, yet the resulting actions must still be aligned with the discrete action spaces of environments. This alignment represents not a fundamental internalization but a compromise, adapting model outputs to the constraints of environmental action structures. Similarly, the upgraded function calling mechanism introduced after GPT-5, which incorporates context-free grammar (CFG) (OpenAI, 2025a), allows LLMs to output structured and rule-based actions that conform to external system requirements. However, this remains a syntactic constraint on model outputs, effectively providing a standardized interface to external systems rather than a truly internalized ability of the model.
Agents therefore play a decisive role in mediating this boundary. They allow LLMs to internalize symbolic tools, thereby enhancing reasoning and self-correction, while also orchestrating access to non-internalizable tools through external mechanisms. This dual pathway ensures that LLMs are not confined to their parameterized capabilities alone but can extend into broader operational domains. In this way, agents transform the tension between internalizable and non-internalizable tools from a limitation into an opportunity, enabling robust problem solving in multimodal, embodied, and real-world contexts.
Memory and learning extension. Another crucial motivation for agents lies in their capacity to overcome the intrinsic memory limitations of LLMs. Due to restricted context windows, LLMs struggle to maintain continuity across extended interactions or to accumulate knowledge over multiple sessions. Agents address this shortcoming by incorporating external memory systems capable of storing, retrieving, and contextualizing past experiences. Such systems simulate long-term memory and enable experiential learning, allowing agents to refine strategies based on historical outcomes rather than treating each interaction as isolated. However, in the TEA Protocol, memory is not defined as a core protocol component but is instead positioned at the infrastructure layer. This design choice reflects the anticipation that future LLMs may gradually internalize memory mechanisms into their parameters, thereby reducing or even eliminating the need for external memory systems. In other words, while memory expansion is indispensable for todayâs agents, it may represent a transitional solution rather than a permanent defining element of agency.
Bridging virtual and external worlds. It has been suggested that LLMs encode within their parameters a kind of âvirtual world,â enabling them to simulate reasoning and predict outcomes internally. However, without an external interface, such simulations remain trapped in closed loops of self-referential inference, disconnected from the contingencies of real-world environments. Agents play a critical role in bridging this gap: they translate the abstract reasoning of LLMs into concrete actions, validate outcomes against environmental feedback, and close the loop between perception, reasoning, and execution. This bridging function transforms LLMs from purely linguistic engines into operationally grounded components whose outputs can be tested, refined, and extended within real or simulated environments.
Autonomy and goal-directedness. Beyond reactivity, agents are motivated by their capacity for autonomy. While LLMs typically operate in a reactive fashion, producing outputs in response to explicit prompts, agents can adopt proactive behaviors. They are capable of formulating subgoals, planning action sequences, and dynamically adapting strategies in light of environmental changes or task progress. This goal-directedness is what elevates agents from passive tools into active participants in problem solving. Autonomy ensures that agents are not merely executing instructions but are able to pursue objectives, adjust course when facing uncertainty, and coordinate with other agents. Such properties are essential for multi-agent collaboration and for tackling open-ended, general-purpose tasks that require initiative as well as adaptability.
Taken together, these motivations highlight why agents must be modeled as a core pillar of the TEA Protocol. Environments provide the stage for interaction, tools expand the operational scope, but it is agents that integrate reasoning, memory, tool usage, and autonomy into cohesive systems of action. By serving as mediators between LLMs and their environments, agents ensure that abstract reasoning is translated into grounded execution, enabling robust and scalable task solving across domains. In this sense, agents represent the crucial entity that transforms language models from passive predictors into active problem solvers within a unified multi-agent framework.
#### A.1.3 Tool
Within the TEA Protocol, the decision to treat tools as a core component alongside environments and agents extends far beyond a matter of convenience in terminology. Tools represent the crucial mediating constructs that encapsulate and operationalize the action spaces of environments, while simultaneously serving as the primary extension layer of agent capabilities. Environments provide the structural stage on which interactions occur, and agents embody the reasoning and decision-making mechanisms that drive behavior, but it is through tools that such reasoning becomes executable and scalable. Without tools, agents would be confined to abstract planning or primitive environmental actions, and environments would remain underutilized as passive backdrops rather than dynamic arenas of transformation.
Moreover, tools play a unique role in bridging symbolic reasoning and concrete execution, providing the abstraction layers necessary to decompose complex tasks into manageable units, and enabling cross-domain transfer through their modularity and portability. They also reveal the shifting boundary between what can be internalized into an agentâs parameters and what must remain external, highlighting the evolving interplay between intelligence and embodiment. In this sense, tools are not merely auxiliary aids but indispensable pillars that shape the architecture of multi-agent systems. The following dimensions illustrate the motivations for elevating tools to a core component of the TEA.
Extending the operational boundary. The primary function of tools is to expand the operational scope of agents beyond what is directly encoded in model parameters or supported by immediate environment interactions. Environments by themselves typically offer only primitive actions, and LLMs by themselves are limited to symbolic reasoning. Tools bridge this gap by furnishing additional pathways for action, allowing agents to manipulate physical artifacts or virtual systems in ways that exceed the direct expressive capacity of the model. From physical devices such as hammers, keyboards, and robotic arms to virtual infrastructures such as databases, APIs, and code execution engines, tools multiply the modes through which agents can influence their environments. Without tools, agents would be confined to intrinsic reasoning and the primitive action space of environments, leaving them incapable of executing tasks that require domain-specific operations. With tools, however, complex objectives can be decomposed into modular operations that are both tractable and reusable. This decomposition makes problem solving significantly more efficient, while also enhancing adaptability across domains. In this way, tools act as multipliers of agency, transforming abstract reasoning into a wider range of tangible interventions.
Hierarchy and abstraction. Tools are not flat or uniform components but exhibit a hierarchical and abstract structure. At the lowest level, tools correspond to atomic environmental actions, such as âclicking a buttonâ or âmoving one step.â These atomic units can then be combined into higher-level compound tools such as âopening a fileâ or âconducting a search.â At an even higher level, compound tools may evolve into strategy-like constructs, such as âwriting a report,â âplanning a trip,â or âcompleting a financial transaction.â Each level builds upon the previous, creating a hierarchy of reusable capabilities. This hierarchical structure is not only efficient but also central to interpretability. Higher-level tools inherently carry semantic labels that communicate their function, which in turn makes agent behavior more transparent to human observers and more predictable to other agents. Such abstraction layers reduce the cognitive and computational load on the agent when planning, since invoking a high-level tool can encapsulate dozens or hundreds of low-level steps. Moreover, in multi-agent systems, the semantic richness of high-level tools serves as a lingua franca, facilitating coordination and collaboration.
Boundary between tools and agent capabilities. The relationship between tools and agents is dynamic rather than static. As LLM reasoning and learning capabilities improve, certain tools can be gradually internalized into model parameters, effectively transforming into latent agent abilities. Examples include logical inference, grammar correction, structured text formatting, and code generation, which once required external support but have increasingly been subsumed into the modelâs intrinsic skills. In this sense, the boundary between what is a âtoolâ and what is an âabilityâ is fluid and shaped by the trajectory of model development. By contrast, many tools remain non-internalizable because they are tightly coupled with environmental properties or external infrastructures. These include robotic arm manipulation, database queries, API interactions, and other operations that inherently depend on external systems or physical substrates. This duality creates a layered conception of agency: a âcore capability layerâ composed of skills internalized within the model, and an âextended layerâ realized through external tool use. The shifting line between these two layers reflects the ongoing negotiation between intelligence and embodiment, highlighting why tools must be explicitly recognized as a structural component.
Evolution and portability. Tools are not static constructs but evolve alongside environments and agent requirements. In programming contexts, for instance, an initial tool may simply execute code. Over time, as demands increase, this basic function evolves into more advanced utilities such as âstatic code analysis,â âautomated test generation,â and âcontinuous deployment.â A similar trajectory occurs in other domains, where rudimentary tools gradually give rise to sophisticated pipelines capable of handling more complex and specialized tasks. In addition to evolution, tools are inherently portable. A well-designed summarization tool, for example, can be reused across very different contexts, from condensing news articles to producing academic literature reviews. This reusability makes tools a natural vehicle for cross-domain generalization, enabling knowledge and functionality to transfer without retraining the underlying model. For these reasons, the TEA Protocol emphasizes modularization and standardization of tools, ensuring that they can evolve flexibly while maintaining interoperability across agents and environments.
Toward intelligent tools. Traditional tools are passive, executing predefined functions only when invoked by an agent. They wait for explicit instructions and do not adapt to context or anticipate needs. However, the trajectory of tool development points toward increasing intelligence, where tools exhibit perception, analysis, and even limited decision-making capabilities. For example, an advanced debugging tool may not only check code upon request but also proactively scan for hidden vulnerabilities, propose optimizations, and even prioritize issues based on estimated risk. Such capabilities blur the line between tools and agents, effectively creating semi-agentic components. Intelligent tools can share responsibility for decision making, reduce the supervisory burden on agents, and participate in distributed problem-solving processes. In this way, tools transition from being passive executors to collaborative partners, altering the topology of multi-agent systems and reshaping the balance between reasoning and execution. Recognizing this trend is critical for designing flexible architectures, as it ensures that the TEA Protocol remains relevant in scenarios where tools are no longer inert extensions but active contributors to system intelligence.
In summary, tools serve as both encapsulations of environmental action spaces and as extensions of agent capabilities. They reduce task complexity through hierarchical abstraction, extend applicability through the balance of internalization and externalization, and foster scalability through evolution, portability, and intelligent design. By transforming the interaction between environments and agents into a modular and expandable architecture, tools anchor the adaptability and generality of multi-agent systems. For these reasons, the TEA Protocol must model tools as a core pillar, providing standardized interfaces that ensure flexible invocation and sharing across contexts, thereby supporting the overarching goal of general-purpose task solving.
### A.2 Transformation Relationships
While agents, environments, and tools are modeled as distinct pillars within the TEA Protocol, their boundaries are not fixed but fluid. Practical systems often demand that one entity temporarily assume the role of another in order to achieve modularity, scalability, and seamless collaboration. These transformation relationships are therefore indispensable, as they provide the mechanisms by which reasoning can be encapsulated into standardized functions, tools can be elevated into autonomous actors, and environments can acquire adaptive properties. In what follows, we examine the motivations for such transformations, beginning with the bidirectional conversions between agents and tools.
Agent-to-Tool (A2T). The motivation for the A2T transformation lies in compressing the complex reasoning and interaction capabilities of agents into reusable tool interfaces. Instead of remaining as fully autonomous components, some agents can be abstracted into functional modules, thereby enhancing modularity, interoperability, and scalability within multi-agent systems. This transformation can be explained from three perspectives:
- Modularization and encapsulation of complex autonomous systems. Although an agent possesses the complete perceptionâreasoningâexecution chain, a single autonomous agent is often too complex to be directly reused in large-scale systems. Through A2T transformation, the internal logic of the agent is âfoldedâ into a black-box tool interface, whose external manifestation is reduced to a clear input and output. In this way, it no longer exists as an âindependent autonomous entity,â but as a âfunctional moduleâ that can provide services to other agents or workflows. This encapsulation emphasizes the reduction of collaboration complexity, enabling higher-level systems to focus solely on results without interfering in or interpreting the agentâs internal reasoning process.
- Difference in role semantics: autonomous entity vs. functional unit. As an agent, it must perceive its environment, set goals, and dynamically adjust strategies. As a tool, however, it merely performs a specified function when invoked. In many multi-agent scenarios, it is unnecessary for all agents to maintain high degrees of autonomy, as this would create excessive interaction overhead and conflict management. Downgrading certain agents into tools (A2T) means relinquishing their goal-setting and decision-making functions while retaining only their reusable capabilities. This role shift ensures that the system contains both âautonomous coresâ and âfunctional components,â thereby forming a layered structure of collaboration.
- Enhancing composability and ecological reusability. Once encapsulated as a tool, an agent can be reused across diverse systems and contexts like a modular building block. For instance, a âdeep research agentâ operates autonomously by dynamically planning search strategies, iteratively analyzing data, and summarizing insights. After A2T encapsulation, however, it becomes a âresearch toolâ that simply receives a query request and returns results, ready for invocation by higher-level agents. This transformation greatly enhances interoperability and composability, enabling agents to be reused in different workflows without incurring integration costs due to their autonomous identity.
Tool-to-Agent (T2A). Within the TEA Protocol, the essence of T2A transformation is to incorporate tools into the callable interface layer of agents, making them the âoperational actuatorsâ through which abstract plans are executed in real environments. Agents are primarily responsible for setting goals and performing high-level reasoning, while tools handle concrete operations and interactions with environments. This division of labor not only optimizes system architecture but also ensures that complex tasks can be accomplished through layered collaboration. The necessity of T2A can be articulated along three key dimensions:
- Bridging reasoning and execution to close the task loop. The outputs of agents are often high-level plans or symbolic descriptions, but without executable mappings, these outputs remain inert and fail to alter the environment. T2A provides the crucial mechanism for grounding abstract reasoning into concrete actions. For example, a planning agent may generate the instruction âanalyze the database and generate a report,â while database query and visualization tools carry out the corresponding SQL queries and chart rendering. Without T2A, agent reasoning would remain disconnected from environmental change, leaving the perceptionâreasoningâexecutionâfeedback loop incomplete. Thus, T2A is indispensable for ensuring that agents can translate reasoning into operational impact.
- Reducing cognitive and computational burden of core agents. If every low-level operation were to be handled directly by an agent, it would be overloaded with detail management, increasing computational costs and undermining strategic reasoning efficiency. Through T2A, agents can delegate domain-specific or low-level tasks to specialized tools and concentrate on higher-level planning and adaptation. For instance, a data analysis agent need not implement SQL parsing, execution, and optimization itself, but instead invokes SQL tools that encapsulate these functions. This separation prevents agents from being âtrapped in detailsâ and ensures that their resources remain dedicated to abstract reasoning. The necessity here lies in maintaining agents at the right level of abstraction to maximize efficiency and scalability.
- Enhancing modularity and ecological extensibility. Tools are inherently modular and portable across domains, whereas agent reasoning mechanisms evolve more gradually. With T2A, agents can flexibly incorporate new tools through standardized interfaces without retraining or structural modification, thereby rapidly expanding their functional boundaries. For example, a writing agent can seamlessly integrate grammar checkers, translation tools, or image generators to support multimodal authoring, all without altering its core reasoning logic. This modularity and extensibility ensure that agents remain adaptive as environments and ecosystems evolve, allowing the system to sustain long-term scalability and cross-domain applicability.
Environment-to-Tool (E2T). The core motivation of E2T lies in abstracting the raw action space of environments into a structured and standardized toolkit, where individual actions are no longer isolated calls but interconnected components sharing contextual information and causal constraints. This transformation enables agents to operate environments at a higher level of planning rather than dealing with fragmented primitives. Its necessity can be articulated in three main dimensions:
- Enhancing interaction consistency and planability. Raw environment actions are often fragmented and tightly coupled to implementation details, making strategies hard to generalize or reproduce. Through E2T, these actions are typed and explicitly annotated with preconditions and postconditions, forming a âplannable interface layerâ that supports sequential decision-making. Agents thus gain a consistent and reusable structure for reasoning across complex environments.
- Strengthening semantic alignment and composability. Toolkits enforce standardized input-output patterns, error-handling semantics, and shared invariants. This allows individual tools to be reliably composed into macro-tools and reused across structurally similar environments. As a result, agents can align semantics across heterogeneous domains, improving transferability and reducing the engineering cost of adaptation.
- Ensuring unified security and operability. An E2T toolkit not only abstracts actions but also integrates mechanisms such as permission control, compliance boundaries, execution logs, and performance optimization. Compared with direct manipulation of raw actions, this design guarantees governability and observability of interactions, providing a stable operational foundation for scalable intelligent systems.
Tool-to-Environment (T2E). The essence of T2E lies in elevating a set of originally independent tools into an environment abstraction, transforming them from isolated callable interfaces into a unified action space governed by shared state and contextual rules. This transformation means that tools are no longer merely passive functions but are organized into a coherent environment where sequential decision-making, long-term planning, and adaptive control become possible. For example, in a programming scenario, tools for code editing, compilation, and debugging are scattered when invoked independently, but under T2E they are encapsulated as a programming environment that maintains code state consistency and contextual continuity, thereby enabling agents to execute complete development workflows. The necessity of T2E is reflected in three key aspects:
- From function calls to stateful spaces. Tools used in isolation are often stateless or weakly stateful, with limited causal connections between invocations. Through T2E, tools are embedded within a shared state space, ensuring historical dependencies and preconditionâpostcondition constraints are preserved. This upgrade supports sequential reasoning and long-horizon planning. For instance, code editing must remain consistent with compilation and debugging, which is only guaranteed within a stateful environment abstraction.
- Enhanced compositionality and planning. T2E organizes tools into a structured environment with explicit transition rules, enabling agents to combine primitive tool actions into higher-level strategies. Instead of treating each tool as a standalone utility, agents can now treat the toolset as an interconnected action space, allowing for the construction of complex workflows such as âdesignâimplementâtestâdeployâ pipelines.
- Unified governance and scalability. By encapsulating tools into an environment, T2E makes it possible to enforce system-wide policies such as access control, compliance constraints, execution logging, and performance monitoring. This ensures that agent interactions remain safe, auditable, and scalable, even as the toolset grows in size and complexity.
Agent-to-Environment (A2E). The A2E transformation redefines an agent not merely as an autonomous decision-maker but as an interactive environment that exposes state spaces, interaction rules, and feedback mechanisms for other agents. In this view, an agent is abstracted into a contextual substrate upon which other agents can act, thereby turning its internal reasoning and behavioral logic into the operational constraints of an environment. This design highlights the interchangeability of agents and environments and provides a principled pathway for hierarchical modeling and scalable system integration. The necessity of this transformation can be articulated across three dimensions:
- Layered and modular system design. In complex tasks, if all agents directly interact with the base environment, the system quickly becomes unmanageable and difficult to extend. Through A2E, high-level agents can be abstracted as environments, exposing simplified interaction interfaces for lower-level agents. For example, a âmarket agentâ can be abstracted as an environment that maintains trading rules, asset states, and dynamic pricing, while individual trader agents perform buying and selling actions within it. This establishes a clear hierarchical structure in which low-level agents focus on local optimization and high-level agents (as environments) coordinate global dynamics, thereby improving scalability and maintainability.
- Facilitating multi-agent training and transfer learning. A2E also provides a practical framework for training and simulation in multi-agent systems. A well-trained agent can be transformed into an environment that offers stable yet challenging dynamics for other agents to learn from. For instance, a navigation agent can be redefined as an environment, exposing route planning and obstacle feedback to new agents, thus eliminating the need to remap complex dynamics. This approach accelerates training, supports transfer of task knowledge, and improves generalization under limited data and computational resources.
- Human-in-the-loop interaction and rule modeling. In many collaborative scenarios, humans themselves can be viewed as special agents. However, treating them as fully autonomous components complicates the adaptation of artificial agents to human constraints. Through A2E, humans can instead be modeled as environments, where their preferences, behaviors, and constraints are expressed as environmental feedback. For example, in an interactive writing system, human edits and suggestions can be treated as feedback signals, guiding an artificial agent to iteratively refine its outputs. This modeling offers a unified interface that allows agents to better align with human intentions, thereby improving efficiency and user experience in human-AI collaboration.
Environment-to-Agent (E2A). The E2A transformation elevates environments from passive containers of state and action spaces into autonomous components capable of reasoning, decision-making, and proactive interaction. Traditionally, environments only provide state transitions in response to external actions, but in dynamic and open-ended scenarios, this passivity often becomes a limitation. By embedding reasoning mechanisms and adaptive policies into environments, E2A enables them to operate as agents in their own right, expanding the functional landscape of multi-agent systems. The necessity of this transformation can be articulated across three dimensions:
- Enhancing realism and challenge in training. Passive environments often fail to capture the richness of real-world dynamics, where external systems and actors are not static but actively adaptive. Through E2A, an environment can be transformed into an adversarial or cooperative agent, thereby offering dynamic strategies and responses that better approximate real-world complexity. For example, in reinforcement learning for autonomous driving, an environment that passively simulates traffic can be upgraded into an opponent agent that actively generates unpredictable vehicle behaviors, thus creating more robust and realistic training conditions.
- Facilitating adaptive coordination and cooperation. In multi-agent systems, agents often need to adapt to evolving contexts, but purely passive environments cannot provide the necessary adaptive feedback loops. By converting environments into agents, they can participate in coordination, negotiation, and joint planning. For instance, a smart city simulation environment can be redefined as an agent that dynamically manages traffic flows, energy distribution, and environmental policies, actively engaging with other agents (e.g., transportation or energy management agents). This transformation ensures that system-level goals are co-constructed rather than imposed unilaterally.
- Expanding the functional scope of environments. Beyond training and coordination, E2A extends environments into autonomous participants in computational ecosystems. A passive environment can only define possibilities, but as an agent, it can proactively initiate actions, enforce constraints, and even set goals that shape the trajectory of interaction. For example, in gaming, a dungeon environment that passively defines maps and rewards can be transformed into an opponent agent that actively strategizes, adapts difficulty levels, and tailors interaction to player behavior. This shift not only increases engagement but also makes environments integral contributors to task execution and system evolution.
### A.3 Motivation for the Self-Evolution Module
General purpose agents operate under shifting task distributions, evolving environments, and expanding tool ecosystems. In this setting, treating prompts, tools, and coordination policies as static assets can lead to accumulated brittleness, where small interface changes, unseen task patterns, or environment specific constraints cause cascading failures. This motivates a protocol level self-evolution mechanism that allows agent-associated components to be refined from execution feedback while remaining governed. In TEA, self-evolution is coupled with version management and tracing so that each update is recorded with explicit version lineage, enabling reproducibility, audit, and rollback when an update degrades performance. Moreover, modeling environments with explicit boundaries and constraints provides a natural safety and permission layer for evolution, preventing uncontrolled side effects during online updates. Finally, although refinement introduces additional computation, TEA encourages reuse of evolved components and synthesized tools across tasks, amortizing one time refinement cost over subsequent runs.
### A.4 Other Relationships
Tool typology and roles. In the design of agentâtool interactions, tools can be categorized according to their functional roles and structural properties. Different types of tools vary in their degree of statefulness, contextual awareness, adaptivity, and autonomy. This typology highlights how tools evolve from simple callable functions to more adaptive and contextually grounded components, shaping how agents can reason, coordinate, and act through them.
- Ordinary tools (MCP-style). Stateless callable functions with weak or implicit inter-tool relations. They typically lack environment-bound context and do not adapt their behavior to evolving task states beyond provided parameters.
- Agent-to-Tool (A2T). An agent is exposed as a callable tool while preserving internal policies, memory, and coordination capabilities. Compared with ordinary tools, A2T exhibits task adaptivity and limited autonomy, enabling on-the-fly decomposition and parameter refinement.
- Environment-to-Tool (E2T). An environmentâs action space is lifted into a context-aware toolkit. Tools within the toolkit are explicitly related via shared state, pre/post-conditions, and constraints, yielding stronger intra-tool structure than standalone MCP tools.
Scaling selection via hierarchical management. As tool ecosystems grow, selecting appropriate candidates becomes a major bottleneck. TCP supports delegating coherent tool families (or toolkits) to agent or environment managers, inducing a tree-structured index (category $\rightarrow$ toolkit $\rightarrow$ primitive tool). This hierarchical routing substantially reduces search cost and aligns with TEA transformations (A2T/E2T/T2E) by allowing managers to prune branches and surface only context-relevant subsets.
Embedding-based retrieval. Each tool is assigned a vector embedding derived from its name, description, schema, and usage signals. Vector similarity enables rapid shortlist generation for candidate tools and can be combined with keyword filtering and hierarchical routing (tree walk + ANN search). This hybrid retrieval pipeline improves recall under tool proliferation while reducing latency and cognitive load for agent planners.
## Appendix B Comparison with Other Protocols
Table 6: Protocol-level comparison: TEA Protocol vs. A2A vs. MCP across fundamental dimensions including entity management, lifecycle, version control, self-evolution, and ecosystem support. Symbols: $\checkmark$ = Yes/Supported, $\triangle$ = Partial, $\times$ = No/Not supported. Highlighted rows (blue background) indicate key distinguishing features.
| Dimension | TEA | A2A | MCP |
| --- | --- | --- | --- |
| Basic Information | | | |
| Proposer | Our work | Google | Anthropic |
| Core Entity | Tool, Environment, Agent | Agent, Tool | Model |
| Protocol Focus | Tool, Environment, Agent | Agent, Tool | Tool/Resource |
| Agent & System Features | | | |
| Agent First-Class | $\checkmark$ | $\triangle$ | $\times$ |
| Multi-Agent | $\checkmark$ | $\triangle$ | $\times$ |
| Tracer | $\checkmark$ | $\times$ | $\times$ |
| Memory | $\checkmark$ | $\times$ | $\times$ |
| Entity Lifecycle | $\checkmark$ | $\times$ | $\times$ |
| Version Management | $\checkmark$ | $\times$ | $\times$ |
| Self-Evolution Support | $\checkmark$ | $\times$ | $\times$ |
| Context Management | $\checkmark$ | $\triangle$ | $\times$ |
| Entity Transformations | $\checkmark$ | $\times$ | $\times$ |
| Scalability | $O(\log n)$ | $O(n^{2})$ | $O(n)$ |
| General & Ecosystem | | | |
| Model-Agnostic | $\checkmark$ | $\checkmark$ | $\checkmark$ |
| Framework-Agnostic | $\checkmark$ | $\checkmark$ | $\checkmark$ |
| Key Strength | Unified | Interop. | Standard. |
| Open Ecosystem | $\checkmark$ | $\triangle$ | $\triangle$ |
Table 6 provides a systematic comparison across fundamental protocol dimensions. We explain each dimension in detail as follows:
### B.1 Basic Information
Proposer: This dimension identifies the originating organization for each protocol. Googleâs A2A protocol was introduced as part of their agent communication framework, focusing on enabling agents to communicate with each other. Anthropicâs MCP (Model Context Protocol) was designed to standardize how LLMs interact with tools and resources. TEA Protocol is proposed in this work as a unified framework that extends beyond these existing approaches by integrating tools, environments, and agents into a cohesive system.
Core Component: This dimension defines the fundamental building blocks treated as first-class protocol components. The TEA Protocol uniquely unifies Tools, Environments, and Agents as co-equal, first-class components, each governed by dedicated context protocols (TCP, ECP, ACP) that provide comprehensive lifecycle and version management. This unified abstraction is critical for enabling self-evolution, where components can dynamically adapt their implementations (e.g., code evolution or prompt refinement). In contrast, existing protocols lack a unified first-class component abstraction. Googleâs A2A protocol centers primarily on agent-to-agent communication, and does not establish tools, environments, context, or tasks as independent, managed components. This architectural limitation results in state dispersion across heterogeneous agents, complicates global lifecycle management, and leads to a tight coupling between reasoning and execution, which significantly hinders system refactorability. Anthropicâs MCP treats tools as passive, stateless interfaces rather than evolvable and composable components. Within this framework, tools lack internal state semantics, versioning and dependency models, and mechanisms for context inheritance. Ultimately, while existing protocols facilitate the invocation of resources, they fail to provide unified mechanisms for systematic management and structural evolution.
Protocol Focus: This dimension describes the primary communication and interaction patterns each protocol addresses. TEA provides three unified protocols: TCP (Tool Context Protocol) for tool management, ECP (Environment Context Protocol) for environment abstraction, and ACP (Agent Context Protocol) for agent orchestration. These protocols work together to enable seamless interoperability across all three component types, with each protocol maintaining comprehensive lifecycle tracking, version histories, and evolution support. This enables dynamic adaptation scenarios such as tool evolution (where tools can be updated, refined, or replaced while maintaining backward compatibility), prompt evolution (where agent prompts can be versioned and improved over time), and agent capability evolution (where agents can learn and adapt their behaviors). A2A focuses specifically on agent-to-agent messaging and coordination, providing communication primitives but not addressing tools or environments directly, and lacks any version or evolution management. MCP handles tool and resource integration for LLMs, standardizing how models invoke tools and access resources, but treats tools as static components without lifecycle or version management, making it impossible to support tool evolution or prompt refinement workflows.
### B.2 Agent & System Features
Agent First-Class: First-class support signifies that agents are modeled as independent, managed protocol components with their own semantic schemas, state metadata, and lifecycle mechanisms. TEAâs ACP provides full first-class status to agents, capturing their roles, competencies, and objectives within a unified schema that enables seamless registration, discovery, and orchestration. A2A provides only partial support; although it enables communication, it treats agents more as opaque RPC endpoints with service-level identifiers rather than semantically rich components with managed internal states. MCP does not define agents as protocol components at all, focusing instead on model-to-tool interactions, thereby overlooking the agent as a primary unit of orchestration and management.
Multi-Agent: Multi-agent support refers to mechanisms for coordinating multiple agents in collaborative, competitive, or hierarchical configurations. TEAâs ACP formalizes multi-agent dynamics through structured relationship representations, supporting hierarchical organization (where high-level agents coordinate low-level agents), cooperative configurations (where agents collaborate toward shared goals), and competitive scenarios (where agents may have conflicting objectives). A2A enables call-level agent interactions, allowing agents to invoke each other as services, but lacks structured collaboration patterns, or negotiation mechanisms. MCP does not address multi-agent scenarios at all, as it focuses on model-tool interactions rather than agent coordination.
Tracer: Tracer refers to mechanisms for recording and tracking the complete execution process of agents, capturing detailed execution traces, decision points, tool invocations, state transitions, and intermediate results throughout task execution. TEA provides comprehensive tracing capabilities through its tracer system, which meticulously records the agent execution process for each task, enabling persistent task tracking, progress monitoring, error handling, and post-execution analysis. This allows developers to understand how agents reason, act, and evolve throughout task completion, facilitating debugging, optimization, and continuous improvement of agent behaviors. A2A and MCP lack tracing mechanisms, meaning execution tracking must be implemented ad-hoc in each application, leading to inconsistent logging and difficulty in understanding agent decision-making processes and debugging complex workflows.
Memory: Memory interfaces provide mechanisms for storing, retrieving, and managing information across agent interactions and sessions. TEA provides a dedicated memory manager that coordinates different manager components (tool managers, environment managers, agent managers) through session-based management. The memory manager operates as a workflow agent that records complete execution histories, automatically determines when to summarize information, and extracts task insights to assist future task completion. Critically, the session-based management ensures that concurrent calls do not result in resource conflicts, maintaining data consistency and preventing race conditions across multiple agent interactions. This enables agents to build upon past experiences and maintain long-term knowledge while ensuring reliable concurrent access. A2A and MCP do not define memory management protocols, leaving memory concerns to be handled entirely at the application layer, which can lead to inconsistent memory management, difficulty in sharing knowledge across agents, and potential resource conflicts in concurrent scenarios.
Component Lifecycle: Component lifecycle management refers to comprehensive lifecycle tracking and management for all component types (tools, environments, and agents) throughout their operational lifetime. TEA provides unified component lifecycle management through its context protocols (TCP, ECP, ACP), handling creation, registration, state tracking, execution monitoring, and controlled decommissioning for all three component types. This enables dynamic maintenance of instance code, proper resource allocation, state coherence, and graceful termination. Critically, TEAâs lifecycle management supports self-evolution scenarios where components can be updated, refined, or replaced while maintaining operational continuity. A2A and MCP lack comprehensive lifecycle management at this level: A2A only provides basic agent communication without lifecycle tracking for tools or environments, while MCP treats tools as static resources with no lifecycle management, making it impossible to support dynamic updates or evolution.
Version Management: Version management refers to mechanisms for tracking, maintaining, and managing multiple versions of components (tools, environments, and agents) including their code, prompts, and capabilities. TEA provides comprehensive version management through lifecycle and version systems embedded in TCP, ECP, and ACP. This enables critical self-evolution scenarios: code evolution where tool and environment implementations can be versioned, updated, and maintained with backward compatibility; prompt evolution where agent prompts can be versioned, A/B tested, and incrementally improved based on performance feedback; and capability evolution where agents can maintain multiple capability versions and gradually deploy improvements. Each component maintains version metadata, change histories, and evolution trajectories, enabling rollback, comparison, and gradual deployment of improvements. This is essential for building adaptive systems that improve over time. A2A and MCP completely lack version management: A2A treats agents as static service endpoints without versioning support, while MCP treats tools as immutable resources with no version control, making it impossible to support tool evolution, prompt refinement, or adaptive capability development.
Self-Evolution Support: Self-evolution support refers to comprehensive mechanisms that enable components (tools, environments, and agents) to evolve, adapt, and improve over time. TEA provides full self-evolution support by combining component lifecycle management and version management systems, enabling components to dynamically update, refine, and evolve while maintaining operational continuity and backward compatibility. This enables critical self-evolution scenarios: tool evolution where tools can be dynamically updated, refined, or replaced while maintaining version histories; prompt evolution where agent prompts can be versioned, A/B tested, and incrementally improved based on performance feedback; and agent capability evolution where agents can learn from experiences, adapt their behaviors, and maintain multiple capability versions. The combination of lifecycle and version management enables rollback, comparison, gradual deployment, and continuous improvement workflows that are essential for building adaptive systems that improve over time. A2A and MCP completely lack self-evolution support: A2A treats agents as static service endpoints without lifecycle or versioning mechanisms, while MCP treats tools as immutable resources with no lifecycle or version management, making it impossible to support any form of evolution, refinement, or adaptive capability development.
### B.3 Context & System Capabilities
Context Management: Context management refers to mechanisms for capturing, organizing, and retrieving contextual information about tools, environments, agents, and their relationships. TEA offers comprehensive context management through its three context protocols: TCP maintains tool context with embedding-based retrieval and semantic relationship modeling, ECP manages environment state and execution context, and ACP tracks agent states and coordination context. This enables intelligent tool selection, environment-aware execution, and context-aware agent orchestration. A2A provides limited context sharing between agents through message passing, but lacks structured context management or relationship modeling. MCP uses flat tool descriptions without modeling inter-tool relationships, toolkits, or contextual execution environments, making it difficult to select appropriate tools in large-scale systems.
Component Transformations: Component transformations enable components (tools, environments, and agents) to dynamically change their roles (e.g., an agent becoming a tool, or an environment becoming an agent). TEA uniquely supports six transformation types: Agent-to-Tool (A2T) encapsulates agent capabilities as reusable tools, Tool-to-Agent (T2A) designates tools as agent actuators, Environment-to-Tool (E2T) converts environment actions into toolkits, Tool-to-Environment (T2E) elevates tool sets into environment abstractions, Agent-to-Environment (A2E) encapsulates agents as interactive environments for hierarchical modeling, and Environment-to-Agent (E2A) infuses reasoning into environments. These transformations enable dynamic role reconfiguration and flexible system architectures. A2A and MCP do not support component transformations, meaning components have fixed roles that cannot be dynamically adapted to changing task requirements.
Scalability: In an open ecosystem with $n$ coordinatable resources, the fundamental difference in coordination overhead stems from the presence or absence of hierarchical component abstraction and routing mechanisms. A2A adopts a flat multi-agent peer-to-peer collaboration model, where coordination can grow quickly with system scale due to dense pairwise interactions and state alignment. MCP reduces tool integration costs through unified interfaces, but still relies on traversing a large candidate pool or explicit application-level orchestration during resource discovery and capability matching, which can make coordination grow with the number of resources. In contrast, TEA unifies agents, tools, and environments as managed components through transformations (e.g., A2E), and utilizes tree-structured indexing and hierarchical routing for resource localization and task distribution. Under hierarchical capability organization, this can reduce coordination to logarithmic-depth routing, since each decision only considers a small, context-relevant subset at each level.
### B.4 General & Ecosystem
Model-Agnostic and Framework-Agnostic: Model-agnostic means protocols work with diverse LLM backends (GPT, Claude, Gemini, etc.), while framework-agnostic means they can be integrated into different application frameworks. All three protocols are designed with these properties: TEA provides a unified LLM interface at the infrastructure layer that abstracts model heterogeneity, A2Aâs agent communication is independent of the underlying models, and MCPâs tool interface works with any LLM that supports function calling. This ensures broad compatibility and allows developers to choose models and frameworks based on their specific needs rather than protocol constraints.
Key Strength: This dimension highlights each protocolâs primary advantage. TEAâs strength lies in its unified integration of Tools, Environments, and Agents into a single cohesive framework, enabling seamless interoperability and dynamic transformations between component types. A2A excels at agent interoperability, providing efficient mechanisms for agents to communicate and coordinate. MCP provides robust tool standardization, making it easy to integrate diverse tools with LLMs through a consistent interface.
Open Ecosystem Support: Open ecosystem support refers to whether a protocol can independently enable a thriving ecosystem of interoperable agents, tools, and environments without requiring additional frameworks. TEA provides a complete protocol stack with all necessary components (tool management, environment abstraction, agent orchestration, transformations, context management, etc.) to support an open ecosystem where different developers can create compatible agents, tools, and environments that seamlessly interoperate. A2A and MCP provide partial ecosystem support: A2A enables agent-to-agent interoperability but lacks tool and environment management, requiring additional frameworks for complete ecosystem support; MCP enables tool integration and standardization but lacks agent coordination and environment management, also requiring additional frameworks to achieve full ecosystem capabilities.
## Appendix C Details of TEA Protocol
We provide a detailed presentation of the TEA Protocol in this section, as illustrated in Figure 1. The protocol architecture is fundamentally designed around coroutine-based asynchronous execution, enabling concurrent and parallel execution across all system components. This design supports multiple execution patterns: a single agent can concurrently execute multiple independent tasks without state interference, multiple agents can coordinate on shared tasks through collaborative mechanisms, and multiple agents can operate on distinct tasks in parallel.
The TEA Protocol comprises three architectural layers: i) Basic Managers provide foundational services through six specialized managers: the model manager abstracts heterogeneous LLM backends through a unified interface, ensuring model-agnostic interoperability; the prompt manager handles prompt lifecycle management, versioning, and retrieval for agent systems; the memory manager coordinates memory operations across different component managers via session-based concurrency control, preventing resource conflicts in concurrent scenarios; the dynamic manager implements serialization and deserialization mechanisms, converting components (prompts, memory, agents, tools, environments) and their associated code into JSON representations for persistence and restoration; the version manager maintains version histories for all components, where modifications generate new versions while preserving backward compatibility, and component access by identifier retrieves the most recent version by default; and the tracer captures comprehensive execution traces, recording decision points, tool invocations, state transitions, and intermediate results for post-execution analysis and debugging. ii) Core Protocols define three context protocols: the Tool Context Protocol (TCP), Environment Context Protocol (ECP), and Agent Context Protocol (ACP), each managing their respective component types with dedicated schemas, metadata registries, and lifecycle management. iii) Protocol Transformations establish bidirectional conversion relationships among TCP, ECP, and ACP, enabling dynamic role reconfiguration and seamless resource orchestration across component boundaries.
Additionally, the protocol incorporates a Self-Evolution Module that addresses the critical requirement for adaptive agent capabilities by encapsulating evolvable components, including prompts, tool implementations, agent architectures, memory strategies, environment actions, and successful execution solutions, as differentiable variables. The module integrates textgrad optimization and self-reflection mechanisms, allowing agents to iteratively refine these components during task execution. Optimized components are automatically registered as new versions through the version manager, ensuring that subsequent tasks leverage improved capabilities while maintaining access to historical versions for comparative analysis and rollback.
### C.1 Basic Managers
The Basic Managers constitute the foundation of the TEA Protocol, providing essential services that enable higher-level functionalities. These managers include:
- Model Manager provides a unified interface for diverse large language models across multiple providers (OpenAI, Anthropic, Google, OpenRouter, etc.), supporting various model types including chat/completions, responses API, embeddings, and transcriptions. The manager maintains a centralized registry of model configurations, each encapsulating provider-specific parameters, capabilities (streaming, function calling, vision), and fallback mechanisms. It abstracts provider heterogeneity through a standardized invocation interface, enabling seamless model switching and ensuring consistent interaction patterns regardless of the underlying API. The manager supports asynchronous execution, tool/function calling, structured output formats, and automatic fallback to alternative models upon failures, ensuring robust and reliable model access across the system.
- Prompt Manager manages the complete lifecycle of prompts for agents, providing comprehensive version control, template rendering, and dynamic updates. The manager maintains a centralized registry of prompt configurations, each encapsulating system prompts, agent message templates, metadata, and version histories. It supports modular template rendering with dynamic variable substitution, enabling flexible prompt composition through configurable modules. The manager implements automatic versioning where prompt updates create new versions while preserving historical versions, enabling rollback and comparative analysis. It provides asynchronous registration, retrieval, and update operations with concurrent initialization support, ensuring efficient prompt management across multiple agents. The manager integrates with the self-evolution module by exposing trainable variables within prompts, allowing optimization algorithms to refine prompt content while maintaining version consistency. Prompts are persisted as JSON configurations and can be exported as contract documents, ensuring reproducibility and documentation of prompt evolution.
- Memory Manager provides comprehensive memory support to agents, managing the complete lifecycle of memory systems through registration, initialization, and session coordination. The manager implements session-based concurrency control, where each agent task operates within isolated memory sessions identified by session IDs, agent names, and task IDs. This session isolation ensures that concurrent calls from multiple agents or tasks do not result in resource conflicts or data corruption. The manager supports event-based memory operations, allowing agents to record execution events, step information, and contextual data throughout task execution. Memory systems are registered with configurations and can be dynamically retrieved, updated, and versioned, enabling agents to maintain persistent state and learn from historical interactions while ensuring thread-safe concurrent access.
- Dynamic Manager provides runtime code execution and serialization capabilities for all components (prompts, memory, agents, tools, environments) and their associated code. The manager enables dynamic loading of Python classes and functions from source code strings, creating virtual modules in memory without requiring disk-based files. It implements intelligent code analysis to automatically detect and inject necessary imports based on symbol usage, supporting context-aware import injection for different component types. The manager provides serialization and deserialization mechanisms for parameter schemas, converting Pydantic models to JSON representations and reconstructing them when needed. This enables components and their code to be stored as JSON configurations, loaded dynamically at runtime, and shared across different execution contexts, facilitating code evolution, version management, and dynamic component instantiation.
- Version Manager provides unified version management for all component types (tools, environments, agents, prompts, memory, etc.), maintaining comprehensive version histories with metadata, descriptions, and timestamps. The manager implements semantic versioning with automatic version generation, supporting major, minor, and patch version increments based on the nature of changes. It maintains version histories for each component, tracking the evolution trajectory and enabling access to any historical version for rollback, comparison, or analysis. The manager supports version lifecycle operations including deprecation and archiving, allowing controlled phase-out of older versions while preserving historical records. Version information is persisted as JSON, and component access by name automatically retrieves the latest version by default, while explicit version specification enables precise version control. This unified versioning system ensures consistent evolution tracking across all component types and enables seamless rollback capabilities when needed.
- Tracer provides comprehensive execution tracing capabilities for recording and analyzing agent execution processes throughout task completion. The tracer maintains session-based record management, where each execution step is captured as a structured record containing observation data, tool invocations, session identifiers, task identifiers, timestamps, and unique record IDs. Records are organized by session ID, enabling isolation of execution traces across different agent sessions and tasks while supporting cross-session analysis. The tracer implements flexible query mechanisms, allowing retrieval of records by session ID, task ID, record index, or record ID, facilitating both real-time monitoring and post-execution analysis. It supports persistent storage through JSON serialization with file locking mechanisms to ensure thread-safe concurrent access, enabling execution traces to be saved, loaded, and shared across different execution contexts. The tracer captures the complete execution trajectory including decision points, state transitions, tool call sequences, and intermediate results, providing a comprehensive audit trail for debugging, performance analysis, behavior understanding, and continuous improvement of agent capabilities.
These components work together to support the coroutine-based asynchronous framework, enabling parallel execution and concurrent task handling.
### C.2 Core Protocols
The TEA Protocol defines three core context protocols: the Tool Context Protocol (TCP), the Environment Context Protocol (ECP), and the Agent Context Protocol (ACP). These protocols share a unified architectural design, each implemented through two core components: a context manager and a server. The context manager serves as the central orchestrator, responsible for context engineering (maintaining contextual information and relationships between components), lifecycle management (handling component registration, versioning, state tracking, and resource allocation), and semantic retrieval (enabling efficient component discovery through vector embeddings). The server component encapsulates the context manager and exposes a unified interface, providing operations for component registration, retrieval, execution, version management, and lifecycle control to other system modules. Each protocol generates a unified contract document (similar to Anthropicâs Agent Skills (Anthropic, 2025)) that aggregates all registered componentsâ descriptions, providing a comprehensive overview of available tools, environments, and agents with their capabilities, parameters, and usage guidelines. This architectural pattern ensures consistent access patterns across tools, environments, and agents while maintaining separation of concerns between internal management logic and external service interfaces.
#### C.2.1 Tool Context Protocol
MCP (Anthropic, 2024b) is the most widely adopted tool protocol and is defined by three components: tools, prompts, and resources, corresponding respectively to model-controlled functions, user-initiated interactive templates, and client-managed data. However, despite its widespread adoption, MCP suffers from several fundamental limitations that hinder its effectiveness in complex multi-agent systems (see Table 6). First, MCP lacks context management capabilities, meaning that tool execution environments cannot be adaptively provided to agents, constraining the systemâs ability to maintain coherent context across tool invocations. Second, MCP provides no version management system, preventing tools from evolving over time while maintaining backward compatibility and version history. Third, MCP lacks component lifecycle management, meaning that tools cannot be dynamically registered, updated, or retired with proper lifecycle control, limiting the systemâs ability to manage tool resources effectively.
To address these limitations, we propose the Tool Context Protocol (TCP), a comprehensive framework that fundamentally extends MCPâs capabilities through several key innovations. TCP is implemented through two core components: the ToolContextManager and the TCPServer. The ToolContextManager serves as the central orchestrator for tool lifecycle management, supporting tool loading from both local registries (via the TOOL registry system) and persistent JSON configurations, enabling seamless integration of tools across different deployment scenarios. During tool registration, TCP automatically generates multiple representation formats for each tool: function-calling schemas for LLM function calling interfaces, natural language text descriptions for human-readable documentation, and structured argument schemas (Pydantic BaseModel types) for type-safe parameter validation, providing LLMs with rich semantic information for accurate parameter inference. TCP incorporates comprehensive version management, maintaining complete version history for each tool and supporting version restoration, enabling tools to evolve while preserving backward compatibility. The protocol employs a semantic retrieval mechanism that stores each toolâs description and metadata as vector embeddings using FAISS, enabling efficient similarity-based tool discovery through queryâembedding comparisons. Additionally, TCP generates tool contracts that aggregate all registered tools into unified documentation, facilitating tool discovery and usage. The TCPServer provides a unified API interface that encapsulates the ToolContextManager, exposing operations for tool registration, retrieval, execution, version management, and lifecycle control, ensuring consistent tool access patterns across the system.
#### C.2.2 Environment Context Protocol
In reinforcement learning, frameworks such as Gym (Brockman et al., 2016) provide standardized interfaces for training and testing environments, where each environment specifies its own observation and action spaces. The core abstraction of an environment consists of two fundamental components: observation (the current state of the environment, accessible through state queries) and action (operations that agents can perform to interact with and modify the environment state). However, most existing research on general-purpose agent systems either focuses on single environments or relies on ad-hoc adaptations to independent environments, seldom addressing the need for unified environment interfaces. Recent attempts to encapsulate environments as MCP tools allow agents to interact with them, but this approach lacks mechanisms to capture inter-tool dependencies and to manage the contextual execution environments required by tools.
To overcome these limitations, we introduce the Environment Context Protocol (ECP), a comprehensive framework that establishes unified interfaces and contextual management across diverse computational environments. ECP follows a similar architecture to TCP, implemented through two core components: the EnvironmentContextManager and the ECPServer. At its core, ECP recognizes that each environment provides a set of actions that agents can invoke, where each action represents an operation that agents can perform to interact with the environment. Each environment maintains its own state (observation) accessible through state queries, while actions provide the means for agents to interact with and modify this state. Similar to TCP, ECP supports environment loading from both local registries and persistent configurations, automatically discovers and registers all actions defined within each environment, and incorporates comprehensive version management, semantic retrieval mechanisms, and contract generation. The key distinction is that ECP manages environments (which encapsulate observation and action spaces) rather than standalone tools, enabling agents to interact with computational environments through standardized action interfaces while maintaining environment state coherence.
#### C.2.3 Agent Context Protocol
Existing agent frameworks or protocols, such as A2A (Google, 2025), typically rely on ad-hoc strategies for defining and managing agents, where each agent is associated with specific roles, capabilities, and policies. However, despite their utility, such systems suffer from several fundamental limitations that hinder their effectiveness in complex multi-agent systems (see Table 6). First, existing frameworks lack standardized representations of agent attributes, making it difficult to systematically capture and reason about agentsâ roles, competencies, and objectives, leading to poor interoperability across different agent implementations. Second, existing approaches provide insufficient means to capture and formalize inter-agent interactions, such as delegation, collaboration, or hierarchical organization, limiting the systemâs ability to support structured multi-agent coordination patterns. Third, existing frameworks fail to explicitly encode the contextual relationships between agents and the environments or tools they operate with, thereby complicating consistent state maintenance and coordination in multi-agent scenarios.
To overcome these shortcomings, we introduce the Agent Context Protocol (ACP), which establishes a unified schema for registering, representing, and coordinating agents within the TEA Protocol. ACP follows a similar architecture to TCP and ECP, implemented through two core components: the AgentContextManager and the ACPServer. Similar to TCP and ECP, ACP supports agent loading from both local registries and persistent configurations, and incorporates comprehensive version management, semantic retrieval mechanisms, and contract generation. The key distinction is that ACP manages agents (autonomous components with reasoning capabilities) rather than tools or environments, enabling agents to be registered, orchestrated, and coordinated through standardized interfaces. ACP establishes a unified schema for representing agents through semantically enriched metadata that captures agentsâ roles, competencies, and objectives. The protocol formalizes the modeling of inter-agent dynamics, allowing for cooperative, competitive, and hierarchical configurations through structured relationship representations. ACP enables persistent state tracking across tasks and sessions, ensuring continuity and context preservation in multi-agent interactions. By embedding contextualized descriptions of agents and their interactions, ACP facilitates flexible orchestration, adaptive collaboration, and systematic integration with TCP and ECP, laying the groundwork for scalable and extensible multi-agent architectures.
### C.3 Protocol Transformations
While TCP, ECP, and ACP provide independent specifications for tools, environments, and agents, practical deployment requires interoperability across these protocols. Thus, communication mechanisms and well-defined transformation pathways are indispensable for enabling components to assume alternative roles and exchange contextual information in a principled manner. For instance, when an agent must operate as a tool within a larger workflow, an explicit agent-to-tool transformation becomes necessary. More generally, we identify six fundamental categories of protocol transformations: Agent-to-Tool (A2T), Environment-to-Tool (E2T), Agent-to-Environment (A2E), Tool-to-Environment (T2E), Tool-to-Agent (T2A), and Environment-to-Agent (E2A). Together, these transformations constitute the foundation for dynamic role reconfiguration, enabling computational components to flexibly adapt their functional scope in response to task requirements and system constraints. This design not only ensures seamless interoperability across heterogeneous contexts but also enhances the adaptability and scalability of multi-entity systems.
- Agent-to-Tool (A2T). The A2T transformation encapsulates an agentâs capabilities and reasoning into a standardized tool interface, preserving contextual awareness while enabling seamless integration with existing tool ecosystems. For example, it can instantiate a deep researcher workflow that first generates queries, then extracts insights, and finally produces summaries, thereby providing a general-purpose tool for internet-scale retrieval tasks.
- Tool-to-Agent (T2A). The T2A transformation designates tools as the operational actuators of an agent, mapping the agentâs goals or policies into parameterized tool invocations. In this view, the agent reasons at a higher level while delegating concrete execution steps to tools, ensuring alignment between the agentâs decision space and the toolâs functional constraints. For example, a data analysis agent may employ SQL tools to query structured databases, or a design agent may invoke image editing tools to implement creative modifications. This separation allows agents to focus on strategic reasoning while relying on tools as reliable execution mechanisms.
- Environment-to-Tool (E2T). The E2T transformation converts environment-specific actions and capabilities into standardized tool interfaces, enabling agents to interact with environments through consistent tool calls. It maintains environment state coherence and exposes contextual information about available actions, allowing agents to operate across heterogeneous environments without bespoke adaptations. For example, in a browser environment, actions such as Navigate, GoBack, and Click can be consolidated into a context-aware toolkit that is directly accessible to agents.
- Tool-to-Environment (T2E). The T2E transformation elevates a collection of tools into an environment abstraction, where individual tool functions are treated as actions within a coherent action space governed by shared state and contextual rules. This conversion allows agents to interact with toolkits not merely as isolated functions but as structured environments, thereby supporting sequential decision-making, context preservation, and adaptive control. For example, a software development toolkit comprising tools for code editing, compilation, and debugging can be encapsulated as a programming environment, enabling agents to plan and execute development tasks while maintaining consistent state across tool invocations.
- Agent-to-Environment (A2E). The A2E transformation encapsulates an agent as an interactive environment, exposing its decision rules, behaviors, and state dynamics as an operational context for other agents. This conversion enables agents to function not only as autonomous components but also as adaptable environments in which other agents can act, thereby supporting multi-agent training, hierarchical control, and interactive simulations. For example, in a multi-agent simulation, a market agent can be represented as an environment that provides trading rules and dynamic market responses, allowing other agents to engage in transactions and learn adaptive strategies. Similarly, in human-in-the-loop interaction, a human agent can be modeled as an environment, enabling artificial agents to interpret user feedback and constraints as contextual signals for decision-making.
- Environment-to-Agent (E2A). The E2A transformation embeds reasoning and adaptive decision-making into the state dynamics and contextual rules of an environment, thereby elevating it into an autonomous agent. In this way, the environment is no longer a passive setting for action execution but becomes an active participant capable of initiating behaviors, coordinating with other agents, and enforcing constraints. For example, in adversarial gaming scenarios, an environment that originally only defines the state and action spaces can be transformed into an opponent agent that not only formulates strategies and responds proactively to player actions but also dynamically adjusts difficulty and interaction patterns, providing a more challenging training and evaluation platform. This transformation expands the functional role of environments within agent systems and offers a more dynamic and realistic testbed for multi-agent cooperation and competition research.
These six transformation categories establish a comprehensive framework for dynamic resource orchestration within the TEA Protocol. By enabling seamless transitions between tools, environments, and agents, the protocol transformations support adaptive architectures that reconfigure functional components in response to task requirements and contextual constraints.
### C.4 Self-Evolution Module
The Self-Evolution Module addresses the growing need for agent evolution capabilities in modern AI systems. This module enables agents to continuously improve their performance by optimizing various components during task execution. The module wraps evolvable components as evolvable variables, including: prompts that guide agent behavior and reasoning; tool code that implements agent capabilities; agent code that defines agent architectures and decision-making logic; memory code that manages information storage and retrieval; environment code that defines interaction spaces; and agent execution solutions that represent successful task completion strategies. The module employs two key algorithms for optimization: textgrad (Yuksekgonul et al., 2025) provides gradient-based optimization for text-based components, enabling fine-grained improvements through iterative refinement; and self-reflection enables agents to analyze their own performance, identify weaknesses, and propose improvements. When components are optimized during task execution, the optimized versions are automatically registered as new versions through the version manager, ensuring that subsequent tasks can leverage the improved components while maintaining access to previous versions for rollback and comparison. This self-evolution capability enables agents to adapt and improve over time, learning from experience and continuously refining their capabilities without manual intervention.
TextGrad. TextGrad (Yuksekgonul et al., 2025) treats a target component (e.g., a prompt template or a code snippet) as an optimizable variable and uses feedback from execution to drive iterative updates. In our setting, the feedback signal can be defined from task outcomes and trace data, such as success or failure, constraint violations, tool error messages, intermediate correctness checks, and any available scalar scores. Given a current variable state, the system first runs the component in a controlled setting and collects a run trace via the tracer. It then constructs a differentiable style supervision signal by prompting an LLM to attribute errors to specific spans of the variable and to produce gradient-like edit directions. The optimizer applies the suggested edits to obtain an updated variable, reruns a lightweight validation on held-out traces or the current task, and keeps the update only if it improves the chosen criteria. This loop repeats for a small number of iterations, after which the final variant is registered as a new component version with its lineage and associated trace.
Self-reflection. Self-reflection treats agent-associated components as optimizable variables and improves them through structured critique and revision rather than gradient-style updates. Concretely, after a run, the system summarizes the trace into a compact diagnosis that highlights failure points, missing information, incorrect assumptions, or unsafe actions, and then selects which variables to optimize based on their causal contribution to the observed failures. A reflection prompt then guides the model to propose targeted changes to the selected variables, such as rewriting a prompt instruction, refining a tool description or schema, adjusting a planning heuristic, or generating a patch to a tool implementation. Candidate changes are evaluated through re-execution under the same environment boundaries and constraints, using the tracer to verify that the revised component improves task outcomes and does not introduce new violations. Accepted changes are committed as new versions with rollback support, enabling future runs to select improved variants while preserving historical baselines.
### C.5 Formalization
In this subsection, we present a formal definition of the TEA protocol and its basic properties.
**Definition 1 (TEA Protocol)**
*Let $\mathcal{T},\mathcal{E},\mathcal{A}$ denote the sets of tools, environments, and agents; let TCP/ECP/ACP be the context protocols defined in this appendix; and let $\mathcal{M}$ denote the set of basic managers, including the model manager, prompt manager, memory manager, dynamic manager, version manager, and tracer, which provide foundational services for the protocol. The TEA Protocol is defined as the tuple
$$
\mathrm{TEA}\;=\;\langle\mathrm{TCP},\,\mathrm{ECP},\,\mathrm{ACP},\,\mathcal{M},\,\mathcal{P}_{\mathrm{TEA}}\rangle,
$$
where $\mathcal{P}_{\mathrm{TEA}}$ is a family of typed transformations over $\mathcal{T}\cup\mathcal{E}\cup\mathcal{A}$
$$
\{\mathrm{A2T},\,\mathrm{E2T},\,\mathrm{T2E},\,\mathrm{T2A},\,\mathrm{A2E},\,\mathrm{E2A}\}\subseteq\mathcal{P}_{\mathrm{TEA}}
$$
that satisfy: (i) interface consistency (exposed I/O signatures remain well-typed under the target protocol), and (ii) closure/compositionality (the composition of valid transformations is again an element of $\mathcal{P}_{\mathrm{TEA}}$ whenever domains and codomains match).*
**Definition 2 (Tool)**
*A tool is defined as a tuple
$$
T=\langle n_{T},d_{T},m_{T},g_{T},\phi_{T}\rangle,
$$
where $n_{T}$ is the tool name, $d_{T}$ is the description, $m_{T}$ is the metadata dictionary, $g_{T}\in\{\mathrm{True},\mathrm{False}\}$ indicates whether the tool supports self-evolution (i.e., whether its code can be optimized during task execution), and $\phi_{T}:\mathcal{I}_{T}\to\mathcal{O}_{T}$ is the functional mapping from input space $\mathcal{I}_{T}$ to output space $\mathcal{O}_{T}$ that implements the toolâs behavior.*
**Definition 3 (Tool Configuration)**
*A tool configuration is defined as
$$
\mathrm{ToolConfig}=\langle T,v_{T},C_{T},\mathcal{F}_{T}\rangle,
$$
where $T=\langle n_{T},d_{T},m_{T},g_{T},\phi_{T}\rangle$ is the tool definition, $v_{T}$ is the version string, $C_{T}$ is the source code string, and $\mathcal{F}_{T}=\{F_{\mathrm{fc},T},F_{\mathrm{text},T},F_{\mathrm{schema},T}\}$ is the set of tool representations (function-calling schema, natural language text, and structured argument schema).*
**Definition 4 (Tool Context Protocol (TCP))**
*We formalize TCP as the tuple
$$
\mathrm{TCP}=\langle\mathcal{T},\mathcal{C},\mathcal{S},\mathcal{I}\rangle,
$$
where:
- $\mathcal{T}$ is the set of registered tools, each $T\in\mathcal{T}$ defined as $\langle n_{T},d_{T},m_{T},g_{T},\phi_{T}\rangle$ and associated with a $\mathrm{ToolConfig}$ that maintains version history $\mathcal{H}_{T}:\mathbb{V}\rightharpoonup\mathrm{ToolConfig}$ (a partial function mapping version strings to configurations).
- $\mathcal{C}$ is the tool context manager that maintains state and implements all core functionalities: (i) state mappings $\rho:\mathbb{S}\rightharpoonup\mathrm{ToolConfig}$ (active registry) and $\eta:\mathbb{S}\times\mathbb{V}\rightharpoonup\mathrm{ToolConfig}$ (version history), (ii) embedding service $\xi:(d_{T},m_{T})\to\mathbb{R}^{d}$ with semantic retrieval via vector database, and (iii) lifecycle operations including loading from registries and code, building instances, version management, and contract generation.
- $\mathcal{S}$ is the TCP server that encapsulates $\mathcal{C}$ and exposes a unified interface, delegating all operations to the context manager while providing consistent access patterns.
- $\mathcal{I}$ is the set of interfaces exposed by $\mathcal{S}$ :
- $\mathtt{init}$ - initialize tools from registry and code, build instances, initialize vector database
- $\mathtt{register}$ - create instance, build ToolConfig, store in registry
- $\mathtt{get}$ - get tool instance by name from active registry
- $\mathtt{info}$ - get tool configuration by name from active registry
- $\mathtt{retrieve}$ - retrieve similar tools via semantic search using vector database
- $\mathtt{list}$ - list all registered tool names
- $\mathtt{update}$ - update existing tool with new implementation, generate new version
- $\mathtt{copy}$ - duplicate existing tool with optional new name and version
- $\mathtt{unregister}$ - remove tool from active registry and version history
- $\mathtt{restore}$ - restore specific historical version of tool by name and version
- $\mathtt{vars}$ - extract tool source code as Variable objects for self-evolution
- $\mathtt{setvars}$ - update tool code variables for self-evolution, generate new version
- $\mathtt{invoke}$ - execute tool by name with structured input, return ToolResponse
- $\mathtt{contract}$ - generate unified documentation by aggregating all toolsâ descriptions
- $\mathtt{save}$ - serialize tool configurations and version history to JSON file
- $\mathtt{load}$ - deserialize tool configurations and version history from JSON file
Given a request $r=(\mathtt{tool\_name},\mathtt{tool\_args})$ , $\mathcal{S}$ delegates to $\mathcal{C}$ , which uses $\mathtt{get}$ to obtain the tool instance from $\rho$ using $\mathtt{tool\_name}$ , and then invokes it with $\mathtt{tool\_args}$ via the $\mathtt{invoke}$ operation, returning a ToolResponse with execution results.*
Note. TCP explicitly supports the TEA transformations A2T via an exposure operator $\iota_{A}:A\mapsto T$ and E2T via a lifting operator $\Lambda:E\mapsto(\mathcal{S}_{E},K_{E})$ .
**Definition 5 (Environment)**
*An environment is defined as a tuple
$$
E=\langle n_{E},d_{E},m_{E},g_{E},\mathcal{A}_{E},\sigma_{E},\tau_{E}\rangle,
$$
where $n_{E}$ is the environment name, $d_{E}$ is the description, $m_{E}$ is the metadata dictionary, $g_{E}\in\{\mathrm{True},\mathrm{False}\}$ indicates whether the environment supports self-evolution, $\mathcal{A}_{E}$ is the action space (a dictionary mapping action names to action configurations), $\sigma_{E}:\bot\to\mathcal{S}_{E}$ is the state retrieval function that returns the current state $\mathcal{S}_{E}$ of the environment, and $\tau_{E}:\mathbb{S}\times\mathcal{D}\to\mathcal{O}_{a}$ is the action execution function that takes an action name and input dictionary and returns the action result.*
**Definition 6 (Environment Configuration)**
*An environment configuration is defined as
$$
\mathrm{EnvironmentConfig}=\langle E,v_{E},C_{E},\mathcal{A}_{E},R_{E}\rangle,
$$
where $E=\langle n_{E},d_{E},m_{E},g_{E},\mathcal{A}_{E},\sigma_{E},\tau_{E}\rangle$ is the environment definition, $v_{E}$ is the version string, $C_{E}$ is the source code string, $\mathcal{A}_{E}$ is the action space (dictionary of action configurations with multi-format representations), and $R_{E}$ is the rules string (generated environment rules for interaction).*
**Definition 7 (Environment Context Protocol (ECP))**
*We formalize ECP as the tuple
$$
\mathrm{ECP}=\langle\mathcal{E},\mathcal{C},\mathcal{S},\mathcal{I}\rangle,
$$
where:
- $\mathcal{E}$ is the set of registered environments, each $E\in\mathcal{E}$ defined as $\langle n_{E},d_{E},m_{E},g_{E},\mathcal{A}_{E},\sigma_{E},\tau_{E}\rangle$ and associated with an $\mathrm{EnvironmentConfig}$ that maintains version history $\mathcal{H}_{E}:\mathbb{V}\rightharpoonup\mathrm{EnvironmentConfig}$ (a partial function mapping version strings to configurations).
- $\mathcal{C}$ is the environment context manager that maintains state and implements all core functionalities: (i) state mappings $\rho:\mathbb{S}\rightharpoonup\mathrm{EnvironmentConfig}$ (active registry) and $\eta:\mathbb{S}\times\mathbb{V}\rightharpoonup\mathrm{EnvironmentConfig}$ (version history), (ii) embedding service $\xi:(d_{E},m_{E},\mathcal{A}_{E})\to\mathbb{R}^{d}$ with semantic retrieval via vector database, and (iii) lifecycle operations including loading from registries and code, building instances, action discovery, version management, and contract generation.
- $\mathcal{S}$ is the ECP server that encapsulates $\mathcal{C}$ and exposes a unified interface, delegating all operations to the context manager while providing consistent access patterns.
- $\mathcal{I}$ is the set of interfaces exposed by $\mathcal{S}$ :
- $\mathtt{init}$ - initialize environments from registry and code, build instances, initialize vector database
- $\mathtt{register}$ - create instance, discover actions, build EnvironmentConfig, store in registry
- $\mathtt{get}$ - get environment instance by name from active registry
- $\mathtt{info}$ - get environment configuration by name from active registry
- $\mathtt{state}$ - get current state of environment by name via get_state method
- $\mathtt{retrieve}$ - retrieve similar environments via semantic search using vector database
- $\mathtt{list}$ - list all registered environment names
- $\mathtt{update}$ - update existing environment with new implementation, generate new version
- $\mathtt{copy}$ - duplicate existing environment with optional new name and version
- $\mathtt{unregister}$ - remove environment from active registry and version history
- $\mathtt{restore}$ - restore specific historical version of environment by name and version
- $\mathtt{vars}$ - extract environment source code as Variable objects for self-evolution
- $\mathtt{setvars}$ - update environment code variables for self-evolution, generate new version
- $\mathtt{invoke}$ - execute environment action by name and action name with structured input
- $\mathtt{contract}$ - generate unified documentation by aggregating all environmentsâ rules
- $\mathtt{save}$ - serialize environment configurations and version history to JSON file
- $\mathtt{load}$ - deserialize environment configurations and version history from JSON file
Given a request $r=(\mathtt{env\_name},\mathtt{action\_name},\mathtt{action\_args})$ , $\mathcal{S}$ delegates to $\mathcal{C}$ , which uses $\mathtt{get}$ to obtain the environment instance from $\rho$ using $\mathtt{env\_name}$ , and then invokes the action with $\mathtt{action\_name}$ and $\mathtt{action\_args}$ via the $\mathtt{invoke}$ operation, returning an action result.*
Note. ECP explicitly supports the TEA transformations A2E via an encapsulation operator $\Omega_{A}:A\mapsto\widehat{E}$ that presents an agent as an interactive environment, and T2E via an abstraction operator $\Gamma:(\mathcal{S},K)\mapsto\widehat{E}$ that consolidates a toolkit into an environment abstraction.
**Definition 8 (Agent)**
*An agent is defined as a tuple
$$
A=\langle n_{A},d_{A},m_{A},g_{A}\rangle,
$$
where $n_{A}$ is the agent name, $d_{A}$ is the description, $m_{A}$ is the metadata dictionary, and $g_{A}\in\{\mathrm{True},\mathrm{False}\}$ indicates whether the agent supports self-evolution.*
**Definition 9 (Agent Configuration)**
*An agent configuration is defined as
$$
\mathrm{AgentConfig}=\langle A,v_{A},C_{A},\mathcal{F}_{A}\rangle,
$$
where $A=\langle n_{A},d_{A},m_{A},g_{A}\rangle$ is the agent definition, $v_{A}$ is the version string, $C_{A}$ is the source code string, and $\mathcal{F}_{A}$ is the set of agent representations (function-calling schemas, natural language descriptions, and Pydantic BaseModel argument schemas).*
**Definition 10 (Agent Context Protocol (ACP))**
*We formalize ACP as the tuple
$$
\mathrm{ACP}=\langle\mathcal{A},\mathcal{C},\mathcal{S},\mathcal{I}\rangle,
$$
where:
- $\mathcal{A}$ is the set of registered agents, each $A\in\mathcal{A}$ defined as $\langle n_{A},d_{A},m_{A},g_{A}\rangle$ and associated with an $\mathrm{AgentConfig}$ that maintains version history $\mathcal{H}_{A}:\mathbb{V}\rightharpoonup\mathrm{AgentConfig}$ (a partial function mapping version strings to configurations).
- $\mathcal{C}$ is the agent context manager that maintains state and implements all core functionalities: (i) state mappings $\rho:\mathbb{S}\rightharpoonup\mathrm{AgentConfig}$ (active registry) and $\eta:\mathbb{S}\times\mathbb{V}\rightharpoonup\mathrm{AgentConfig}$ (version history), (ii) embedding service $\xi:(d_{A},m_{A})\to\mathbb{R}^{d}$ with semantic retrieval via vector database, and (iii) lifecycle operations including loading from registries and code, building instances, version management, and contract generation.
- $\mathcal{S}$ is the ACP server that encapsulates $\mathcal{C}$ and exposes a unified interface, delegating all operations to the context manager while providing consistent access patterns.
- $\mathcal{I}$ is the set of interfaces exposed by $\mathcal{S}$ :
- $\mathtt{init}$ - initialize agents from registry and code, build instances, initialize vector database
- $\mathtt{register}$ - create instance, build AgentConfig, store in registry
- $\mathtt{get}$ - get agent instance by name from active registry
- $\mathtt{info}$ - get agent configuration by name from active registry
- $\mathtt{retrieve}$ - retrieve similar agents via semantic search using vector database
- $\mathtt{list}$ - list all registered agent names
- $\mathtt{update}$ - update existing agent with new implementation, generate new version
- $\mathtt{copy}$ - duplicate existing agent with optional new name and version
- $\mathtt{unregister}$ - remove agent from active registry and version history
- $\mathtt{restore}$ - restore specific historical version of agent by name and version
- $\mathtt{vars}$ - extract agent source code as Variable objects for self-evolution
- $\mathtt{setvars}$ - update agent code variables for self-evolution, generate new version
- $\mathtt{invoke}$ - execute agent method by name with structured input, return agent response
- $\mathtt{contract}$ - generate unified documentation by aggregating all agentsâ descriptions
- $\mathtt{save}$ - serialize agent configurations and version history to JSON file
- $\mathtt{load}$ - deserialize agent configurations and version history from JSON file
Given a request $r=(\mathtt{agent\_name},\mathtt{input})$ , $\mathcal{S}$ delegates to $\mathcal{C}$ , which uses $\mathtt{get}$ to obtain the agent instance from $\rho$ using $\mathtt{agent\_name}$ , and then invokes it with $\mathtt{input}$ via the $\mathtt{invoke}$ operation, returning an agent response.*
Note. ACP explicitly supports the TEA transformations T2A via a designation operator $\kappa_{T}:T\mapsto\widehat{A}$ and E2A via an elevation operator $\Psi_{E}:\widehat{E}\mapsto\widehat{A}$ that embeds reasoning/decision capabilities into an environment to obtain an agent abstraction.
## Appendix D The AgentOrchestra Implemented by TEA Protocol
AgentOrchestra is a concrete instantiation and implementation of the TEA Protocol, demonstrating how the protocolâs core principles and transformations can be applied to build a practical hierarchical multi-agent system. This section first introduces the fundamental design principles that govern agent behavior and interaction within the framework, including the definitions of key components such as agents, environments, models, memory, observations, and actions. We then present the specific agents and tools that constitute AgentOrchestra, including the planning agent for task decomposition and coordination, the deep researcher agent for comprehensive information gathering, the deep analyzer agent for complex reasoning tasks, the browser use agent for automated web interaction, the tool generator agent for intelligent tool evolution and management, and the reporter agent for automated report generation and citation management.
### D.1 Agent Design Principles
Agent. An agent is an autonomous computational entity that perceives and interprets the environment, maintains a history of actions and observations, and flexibly generates actions to accomplish a wide variety of user-specified tasks across diverse domains. Within the TEA Protocol framework, agents are managed through the ACP, which provides standardized registration, representation, and coordination mechanisms.
Environment. The environment represents the external context and resources within which the agent operates, providing the interface for action execution and information access. Within the TEA Protocol framework, environments are managed through the ECP, which provides unified inputs, outputs, and environment rules across multiple environments.
Model. LLMs are the core drivers of this framework, providing the reasoning and decision-making capabilities for agents. Within the TEA Protocol framework, models are managed through the Infrastructure Layer, which provides a unified interface for diverse LLMs. This design enables agents to dynamically select and switch between different LLMs during task execution, aligning each modelâs unique strengths with specific requirements.
Memory. Memory serves as a fundamental component of the agent, persistently recording the complete history of agent execution. Within the TEA Protocol framework, memory is managed through the Infrastructure Layer as a workflow agent that operates based on sessions, automatically recording agent execution paths across multiple tasks. This memory system automatically determines when to summarize and extract task insights to assist in task completion.
Observation. An observation primarily consists of the task description, attached files, the agentâs execution history, the environment state, and the set of available tools and sub-agents, providing the agent with a comprehensive view of the ongoing process.
Action. In our framework, actions are managed under the Tool Context Protocol (TCP) and executed through a set of pre-defined tools Wang et al. (2024b); Liang et al. (2025); Roucher et al. (2025) exposed via function-calling interfaces OpenAI (2023); Anthropic (2024b). Actions are not equivalent to tools. A single tool can support multiple actions by accepting different parameters. For example, a planning tool may support create, update and delete through a unified interface.
An agent operates in a perceptionâinterpretationâaction cycle. It observes the environment and stores information in memory, interprets context with the unified LLMs interface, and determines an action. The action is executed in a sandbox, with results recorded back to memory to refine reasoning and adaptation. This loop continues until objectives are achieved or a termination condition is met.
### D.2 Planning Agent
The planning agent serves as the central orchestrator in our hierarchical framework, dedicated to high-level reasoning, task decomposition, and adaptive planning. The planning agent utilizes structured thinking and unified invocation mechanisms to coordinate specialized sub-agents and tools for complex task completion. As illustrated in Figure 4, the planning agent implements a systematic iterative workflow that integrates structured reasoning, context management, and execution coordination with continuous monitoring and adaptive adjustments.
<details>
<summary>x6.png Details</summary>

### Visual Description
\n
## Diagram: Planning Agent Architecture
### Overview
This diagram illustrates the architecture of a "Planning Agent," outlining its components, actions, and pipeline for task management. The agent handles tasks, decomposes them into sub-tasks, assigns them to specialized sub-agents, and manages execution while adapting to objective shifts and unexpected errors.
### Components/Axes
The diagram is structured into three main sections: a "Todo" section, an "Execute" section, and a "Pipeline" section. There are also input and output elements representing "Task", "Objective Shifts", and "Unexpected Errors".
* **Todo:** Contains actions: "add", "update", "clear", "complete".
* **Execute:** Describes functionalities: "Create, update, and manage plans for complex tasks simultaneously" and "Track execution states".
* **Pipeline:** Consists of three stages: "Interpret user tasks", "Decompose into manageable sub-tasks", and "Assign to specialized sub-agents".
* **Input:** "Task" (left side)
* **Outputs:** "Objective Shifts (Update Plans)" and "Unexpected Errors" (bottom side).
* **Sub-agents:** "sub-agent A", "sub-agent B", "tool C", and "......" (right side).
### Detailed Analysis or Content Details
The diagram depicts a workflow.
1. **Task Input:** A "Task" enters the system from the left.
2. **Pipeline - Interpret:** The task is first "Interpreted" (represented by a rounded rectangle with a magnifying glass icon).
3. **Pipeline - Decompose:** The interpreted task is then "Decomposed" into smaller, manageable sub-tasks (represented by a rounded rectangle with a grid icon).
4. **Pipeline - Assign:** These sub-tasks are "Assigned" to specialized sub-agents (represented by a rounded rectangle with a number "2" inside a square icon).
5. **Sub-agents:** The assigned sub-tasks are handled by various sub-agents and tools, including "sub-agent A", "sub-agent B", and "tool C", with an ellipsis indicating more possibilities.
6. **Feedback Loop:** A bidirectional arrow labeled "Feedback" connects the sub-agents back to the "Assign" stage, indicating a feedback mechanism.
7. **Error & Shift Handling:** "Objective Shifts" (requiring plan updates) and "Unexpected Errors" feed back into the pipeline, suggesting a dynamic adaptation process.
8. **Todo Actions:** The "Todo" section lists actions that can be performed on the plan:
* "add": Add a new plan.
* "update": Update the plan.
* "clear": Delete the plan.
* "complete": Mark step as completed.
9. **Execute Functionality:** The "Execute" section describes the agent's capabilities:
* "Create, update, and manage plans for complex tasks simultaneously".
* "Track execution states".
### Key Observations
The diagram emphasizes a cyclical process of task interpretation, decomposition, assignment, execution, and adaptation. The feedback loop and error/shift handling mechanisms highlight the agent's robustness and ability to respond to changing conditions. The "Todo" section suggests a user interface or control panel for managing plans.
### Interpretation
This diagram represents a sophisticated agent-based system for task management. The architecture is designed to handle complex tasks by breaking them down into smaller, manageable units and distributing them to specialized agents. The feedback loop and error handling mechanisms suggest a learning and adaptive system. The "Planning Agent" is not a static planner but a dynamic system that can adjust to changing objectives and unexpected events. The inclusion of "sub-agent A", "sub-agent B", and "tool C" suggests a modular design, allowing for easy integration of new capabilities. The diagram is a high-level overview and doesn't delve into the specifics of the algorithms or data structures used within each component. The cat icon in the top-left corner is likely a branding element and doesn't contribute to the technical understanding of the system.
</details>
Figure 4: Planning Agent Workflow.
Structured Reasoning. The planning agent employs a structured thinking framework that guides each execution step, capturing reasoning processes, evaluation of previous goals, memory insights, next objectives, and tool/agent selections. This structured approach ensures systematic reasoning, explicit progress tracking, and transparent decision-making. The agent dynamically builds a unified interface that combines sub-agents from ACP and tools from TCP (including those transformed from environments via E2T and from agents via A2T), enabling seamless coordination of both specialized agents and domain-specific tools within a single execution framework.
Pipeline Workflow. The planning agent implements a systematic pipeline for task processing and execution that can be conceptually divided into four main stages. The pipeline begins with task interpretation, where the agent analyzes incoming user requests to extract objectives, constraints, and contextual requirements. This is followed by task decomposition, wherein complex objectives are systematically broken down into smaller, executable sub-tasks that can be processed by specialized components. The third stage involves resource allocation, where sub-tasks are strategically assigned to appropriate specialized agents or tools based on their domain expertise and functional capabilities. Finally, the execution and coordination stage manages the task execution, incorporating continuous feedback mechanisms that enable dynamic plan adjustments and inter-agent coordination throughout the process. The implementation incorporates session management for maintaining context across multiple interactions, memory storage and retrieval systems for learning from past experiences, and execution tracking for observability and debugging.
Adaptive Planning and Error Handling. The planning agent incorporates robust mechanisms for handling dynamic changes and unexpected situations. When objective shifts occur, the system updates plans accordingly, triggering a return to the task interpretation phase to reassess and modify the approach. Similarly, when unexpected errors arise during execution, the agent re-evaluates the task and adjusts the plan to address the issues. This adaptive capability ensures that the system can maintain progress even when encountering unforeseen challenges or changing requirements.
The planning agentâs design emphasizes modularity and scalability, interacting with sub-agents through the ACP and utilizing tools from the TCP, thereby concealing domain-specific details and facilitating the integration of new agent types and resources. This architecture enables the agent to maintain a global perspective throughout the execution process, aggregating feedback from sub-agents and monitoring progress toward the overall objective, while performing dynamic plan updates in real-time in response to intermediate results, unexpected challenges, or shifting user requirements.
### D.3 Deep Researcher Agent
The deep researcher agent is a specialized component designed for comprehensive information gathering through multi-round research workflows with multimodal capabilities. As illustrated in Figure 5, the agent implements a systematic pipeline workflow for research execution that begins with task analysis and query generation, followed by multi-engine web search across various platforms, result evaluation and completeness assessment, and iterative refinement through follow-up queries until comprehensive information is gathered. The agent leverages the Reporter Agent (detailed in Section D.7) to generate structured research reports with proper citations and references.
<details>
<summary>x7.png Details</summary>

### Visual Description
\n
## Diagram: Deep Researcher Agent Workflow
### Overview
This diagram illustrates the workflow of a "Deep Researcher Agent," outlining the process from task input to search execution and insight summarization. The diagram is divided into two main sections: "Search" and "Pipeline," with feedback loops indicating iterative refinement.
### Components/Axes
The diagram features the following components:
* **Task:** Represented by a rectangular box on the left, serving as the input to the system.
* **Engines:** A vertical label indicating the search engines used.
* **Search Engines:** baidu, bing, firecrawl, google. Each is represented by a logo within a rounded rectangle.
* **Execute:** A light blue rounded rectangle on the right, representing the execution phase. Contains two checkmark icons with the text: "Iteratively gather insights across multiple rounds" and "Obtain and summarize the final search results".
* **Pipeline:** A horizontal section detailing the processing steps.
* **Analyze:** A rounded rectangle with a person icon, labeled "generate fitting query".
* **Search:** A rounded rectangle with a "G" icon, labeled "query-based web search".
* **Insight:** A rounded rectangle with a lightbulb icon, labeled "extract insights about task".
* **Summarize:** A rounded rectangle with a diamond icon, labeled "summarize the insights".
* **Next Round (Update Query):** An arrow pointing from "Summarize" back to "Analyze".
* **Check Results:** An arrow pointing from "Summarize" to "Execute".
* **Arrows:** Connecting the components, indicating the flow of information.
### Detailed Analysis or Content Details
The diagram depicts a sequential process:
1. A "Task" is initiated.
2. The "Pipeline" begins with "Analyze" to generate a fitting query.
3. The query is then used in a "Search" phase.
4. "Insight" is extracted from the search results.
5. The insights are "Summarize"d.
6. The summarized insights are used to "Check Results" and potentially initiate a "Next Round" by updating the query and restarting the "Analyze" phase.
7. The "Search" engines (baidu, bing, firecrawl, google) feed into the "Execute" phase, which iteratively gathers insights and summarizes the final results.
The arrows indicate a clear flow of information from left to right, with feedback loops for iterative refinement. The "Execute" phase appears to be driven by the output of the search engines.
### Key Observations
The diagram emphasizes an iterative approach to research, with the "Next Round" loop suggesting continuous query refinement based on the insights gathered. The inclusion of multiple search engines indicates a comprehensive search strategy. The "Execute" phase is positioned as the final stage, responsible for synthesizing the information gathered throughout the process.
### Interpretation
The diagram represents a closed-loop system for automated research. The "Deep Researcher Agent" is designed to take a task as input, formulate queries, search the web using multiple engines, extract relevant insights, summarize those insights, and then use the summary to refine the search process. This iterative approach suggests a goal of maximizing the quality and relevance of the research results. The system is designed to be self-improving, learning from each iteration to generate more effective queries and extract more valuable insights. The diagram highlights the importance of both breadth (using multiple search engines) and depth (iterative refinement) in the research process. The system is likely intended to automate tasks that would traditionally require significant human effort, such as literature reviews or competitive analysis.
</details>
Figure 5: Deep Researcher Agent Workflow.
Search Engines. The deep researcher agent integrates multiple search engines to ensure comprehensive coverage and information diversity. The system supports six primary search engines: Baidu for Chinese-language content, Bing, Brave and DuckDuckGoSearch for general web search, Firecrawl for comprehensive web crawling and content extraction with full webpage content retrieval, and Google for comprehensive global search. Additionally, the agent can utilize specialized LLM-based search models for enhanced information retrieval. This multi-engine approach enables the agent to access diverse information sources and overcome limitations of individual search platforms, ensuring robust information retrieval across different domains and languages.
Pipeline Workflow. The core pipeline implements a systematic multi-stage process for research execution. The workflow begins with task analysis and query generation, where the agent generates optimized search queries based on the research objectives, contextual requirements, and previous search history. This initial analysis transforms vague research requests into specific, actionable search queries that can effectively target relevant information sources. This is followed by parallel web search, wherein the agent performs targeted searches across multiple engines and LLM-based search models simultaneously using the generated queries. The multi-engine approach is essential because different search platforms have varying coverage, indexing strategies, and content biases, ensuring comprehensive information retrieval while mitigating the limitations of individual search engines. The third stage involves result merging and evaluation, where the agent consolidates search results from multiple sources and evaluates whether the gathered information provides a complete answer to the research task. This evaluation step is necessary because it determines whether additional research rounds are needed or if sufficient information has been collected. Finally, the report generation stage uses the Reporter Agent to consolidate all research rounds into a structured markdown report with proper citations and references, and generates a comprehensive summary from the final report content.
Iterative Research Process. The deep researcher agent incorporates a sophisticated iterative mechanism for comprehensive research. After each roundâs evaluation, the system checks whether the gathered information provides a complete answer. When additional research is required, the agent enters the next round, where it updates and refines search queries based on previous findings and identified knowledge gaps. Each roundâs content, including queries, search results, and evaluations, is systematically added to the Reporter Agent, which maintains proper citation tracking throughout the research process. This iterative process continues until a complete answer is found or predefined research limits (maximum rounds) are reached. Upon completion, the Reporter Agent generates a final structured report with all citations properly numbered and referenced, ensuring not only comprehensive coverage of complex research topics but also proper source attribution and balanced control over exploration depth, efficiency, and resource consumption.
The deep researcher agentâs design emphasizes adaptability and comprehensiveness, enabling it to handle diverse research tasks ranging from factual inquiries to complex analytical investigations. The multimodal support allows the agent to process both textual and visual information simultaneously, while the iterative workflow ensures that research quality improves through multiple rounds of refinement and validation. The integration with the Reporter Agent ensures that all research findings are properly documented with citations, making the research process transparent and verifiable.
### D.4 Deep Analyzer Agent
The deep analyzer agent is a specialized component designed for complex reasoning tasks involving diverse data sources through a workflow-oriented approach with multimodal data support. As illustrated in Figure 6, the agent implements a systematic pipeline workflow for complex reasoning and analysis that begins with file classification and validation, followed by overall file summary assessment, type-specific analysis strategies, and iterative multi-round refinement until answers are found or analysis limits are reached. The agent leverages the Reporter Agent (detailed in Section D.7) to generate structured analysis reports with proper citations and references.
<details>
<summary>x8.png Details</summary>

### Visual Description
\n
## Diagram: Deep Analyzer Agent Workflow
### Overview
The image depicts a diagram illustrating the workflow of a "Deep Analyzer Agent". The agent processes various file types (image, text, audio, zip) through a pipeline, culminating in task execution, insight extraction, and summarization. The diagram is organized into two main sections: "Mdify" (file processing) and "Pipeline" (task enhancement and analysis), with feedback loops for iterative refinement.
### Components/Axes
The diagram consists of several rectangular blocks representing different stages or components. These are connected by arrows indicating the flow of data and control. Key components include:
* **Task & Files (Input):** Located on the left, representing the initial input to the agent.
* **Mdify (File Processing):** A larger block containing sub-blocks for different file types:
* **Image:** "Caption Image"
* **Text:** "Read the text file"
* **Audio:** "Transcribe Audio"
* **Zip:** "Read the contents of zip"
* **Execute:** A block on the right with two bullet points:
* "Generate an enhanced task from the input text and files"
* "Obtain and summarize the final analysis results"
* **Pipeline:** A horizontal sequence of blocks:
* **Mdify:** "convert files to markdown"
* **Enhance:** "get the enhanced task" (with a pencil icon)
* **Insight:** "extract insights about task" (with a lightbulb icon)
* **Summarize:** "summarize the insights"
* **Next Round (Task & Summaries):** A block at the bottom, representing the output of the pipeline for the next iteration.
* **Check Results:** A block connected to "Next Round" via an arrow.
### Detailed Analysis or Content Details
The diagram illustrates a cyclical process.
1. **Input:** The process begins with "Task & Files" entering the system.
2. **File Processing (Mdify):** The input files are processed based on their type: images are captioned, text files are read, audio is transcribed, and zip files are unpacked.
3. **Task Enhancement (Pipeline):** The processed files are converted to markdown, then enhanced to create a refined task.
4. **Insight Extraction & Summarization:** Insights are extracted from the enhanced task, and these insights are summarized.
5. **Iteration:** The summarized insights and the original task are fed back into the system as "Next Round (Task & Summaries)" for further analysis and refinement.
6. **Result Verification:** The "Check Results" block suggests a quality control step before proceeding to the next round.
7. **Execution:** The enhanced task is executed, generating results that are then summarized.
The arrows indicate a clear flow of information from left to right and a feedback loop from the bottom back to the pipeline. The "Mdify" block is duplicated in the pipeline, suggesting a re-processing step.
### Key Observations
* The diagram emphasizes an iterative process, with feedback loops for continuous improvement.
* The agent handles multiple file types, demonstrating versatility.
* The pipeline is clearly defined, with distinct stages for enhancement, insight extraction, and summarization.
* The inclusion of "Check Results" suggests a focus on accuracy and reliability.
* The diagram is visually clean and easy to follow, using icons to represent different actions.
### Interpretation
The diagram represents a sophisticated data analysis workflow. The "Deep Analyzer Agent" is designed to take raw input (tasks and files), process it through a series of steps to extract meaningful insights, and then refine those insights through iterative analysis. The agent's ability to handle multiple file types and its emphasis on feedback loops suggest a robust and adaptable system. The diagram highlights the importance of converting data into a standardized format (markdown) for efficient processing. The "Execute" block indicates that the agent is not just about analysis but also about taking action based on the insights gained. The overall design suggests a system aimed at automating complex analytical tasks and providing actionable intelligence. The presence of a cat icon in the top-left corner is purely decorative and does not contribute to the technical understanding of the diagram.
</details>
Figure 6: Deep Analyzer Agent Workflow.
File Classification and Support. The deep analyzer agent supports comprehensive file formats including text files, PDFs, images, audio, video, and compressed archives, with support for both local files and URLs. The system employs LLM-based file type classification to determine the appropriate analysis strategy for each file. For URLs, the system automatically detects file types based on URL patterns and extensions, while for local files, it uses both LLM classification and extension-based fallback mechanisms. This classification stage is essential because different file types require different analysis approaches: text and PDF files benefit from chunk-based markdown analysis, images and audio require direct multimodal LLM analysis, and videos may need both direct analysis and transcript-based processing.
Pipeline Workflow. The core pipeline implements a systematic multi-stage process for complex reasoning and analysis. The workflow begins with file validation and classification, where the system validates file accessibility and size constraints, then classifies each file by type (text, PDF, image, audio, video) to determine appropriate analysis strategies. This is followed by overall file summary, where the agent generates a preliminary summary based on file metadata (sizes, types, timestamps) to determine if the task can be answered from file information alone, enabling early termination when sufficient information is available. The third stage involves type-specific analysis, where the agent processes each file according to its type: text files are converted to markdown and analyzed in chunks; PDF files first attempt direct LLM analysis, then fall back to markdown conversion and chunk-based analysis if needed; images first attempt direct multimodal LLM analysis, then proceed to multi-step analysis if the answer is not found; audio files are analyzed directly through multimodal LLM; and video files first attempt direct LLM analysis, then convert to markdown transcripts for chunk-based analysis if needed. Each analysis step checks whether the answer has been found, enabling early stopping when sufficient information is obtained. Finally, the report generation stage uses the Reporter Agent to consolidate all analysis rounds into a structured markdown report with proper citations and references, and generates comprehensive summaries from the final report content.
Iterative Multi-Round Analysis. The deep analyzer agent incorporates a sophisticated iterative mechanism for comprehensive analysis refinement. The system executes multiple analysis rounds, with each round processing all files according to their classified types. After each round, the system synthesizes summaries from all file analyses and evaluates whether a complete answer has been found. When additional analysis is required, the agent enters the next round, where it continues processing files with refined strategies based on previous findings. Each roundâs content, including file classifications, analysis results, and answer evaluations, is systematically added to the Reporter Agent, which maintains proper citation tracking throughout the analysis process. This iterative process continues until a complete answer is found or predefined analysis limits (maximum rounds) are reached. Upon completion, the Reporter Agent generates a final structured report with all citations properly numbered and referenced, ensuring not only comprehensive coverage of complex reasoning tasks but also proper source attribution and balanced control over analysis depth, efficiency, and resource consumption.
Task-Only Analysis. When no files are provided, the deep analyzer agent can directly analyze tasks such as text games, math problems, logic puzzles, or reasoning challenges. The system performs multi-round analysis where each round applies step-by-step reasoning, breaks down task components, identifies key information and constraints, and generates insights and partial solutions. This capability enables the agent to handle diverse reasoning tasks that do not require file-based information, making it a versatile tool for both file-based and file-free analysis scenarios.
The deep analyzer agentâs design emphasizes workflow-oriented processing and multimodal data support, enabling it to handle diverse reasoning tasks ranging from document analysis to complex multi-step problem solving. The adaptive file type handling ensures optimal analysis strategies for each data source, while the iterative workflow guarantees that analysis quality improves through multiple rounds of refinement and validation. The integration with the Reporter Agent ensures that all analysis findings are properly documented with citations, making the analysis process transparent and verifiable.
### D.5 Browser Use Agent
The browser use agent is a specialized component designed for automated web interaction and task completion through the browser_use library. As illustrated in Figure 7, the agent implements a systematic workflow for web interaction and task execution that begins with task initialization and report setup, followed by browser agent execution with intelligent web navigation and interaction, result extraction and evaluation, and comprehensive report generation with execution records. The agent leverages the Reporter Agent (detailed in Section D.7) to generate structured browser task reports with proper documentation.
<details>
<summary>x9.png Details</summary>

### Visual Description
\n
## Diagram: Browser Use Agent Pipeline
### Overview
The image depicts a diagram illustrating the pipeline of a "Browser Use Agent". It shows the flow of a task through several stages, from preparation to recording, with feedback loops for checking results and updating the goal. The diagram is divided into two main sections: "Browser & Computer" and "Pipeline".
### Components/Axes
The diagram consists of the following components:
* **Header:** "Browser Use Agent"
* **Browser & Computer Section:** Contains "Actions" with four options: "goto", "input", "scroll", and "click".
* **Pipeline Section:** Contains five stages: "Prepare", "Generate", "Execute", "Evaluate", and "Record".
* **Task Input:** A rectangular box labeled "Task" on the left side, feeding into the "Prepare" stage.
* **Feedback Loops:** Arrows indicating feedback from "Check Results" to "Next Step" and from "Evaluate" to "Check Results".
* **Output:** "Check Results" and "Next Step (Update Next Goal)".
### Detailed Analysis or Content Details
The diagram illustrates a cyclical process.
1. **Task:** A task is initiated and fed into the "Prepare" stage.
2. **Prepare:** The browser environment is prepared.
3. **Generate:** A list of next actions is generated.
4. **Execute:** The actions list is executed.
5. **Evaluate:** The answer is checked.
6. **Record:** The execution state is recorded.
7. **Check Results:** The results are checked.
8. **Next Step:** The next goal is updated.
The "Browser & Computer" section details the available actions:
* **goto:** Go to the URL.
* **input:** Input a text.
* **scroll:** Scroll down or up.
* **click:** Click a button or position.
The "Execute" section in the upper-right corner provides further details:
* Iteratively generate, execute, and summarize actions.
* Generate next goal until task completion.
The arrows indicate the flow of information and control. The arrow from "Check Results" to "Next Step" suggests an iterative refinement of the goal based on the results. The arrow from "Evaluate" to "Check Results" indicates a feedback loop for verifying the answer.
### Key Observations
The diagram emphasizes an iterative and goal-oriented approach to browser automation. The feedback loops suggest a continuous process of evaluation and refinement. The "Browser & Computer" section highlights the basic actions that the agent can perform. The "Execute" section emphasizes the iterative nature of the process.
### Interpretation
This diagram represents a closed-loop system for automating browser tasks. The agent takes a task as input, prepares the environment, generates a sequence of actions, executes those actions, evaluates the results, and records the execution state. The feedback loops allow the agent to adapt its strategy based on the results, ultimately aiming to achieve the task goal. The system is designed to be iterative, continuously refining its actions until the task is completed. The inclusion of "Record" suggests a logging or debugging capability, allowing for analysis of the agent's behavior. The diagram suggests a sophisticated AI agent capable of complex browser interactions. The agent is not simply executing pre-defined scripts, but rather dynamically generating and adapting its actions based on the observed results.
</details>
Figure 7: Browser Use Agent Workflow.
Browser Agent Integration. The browser use agent leverages the browser_use library, which provides an intelligent browser automation framework with LLM-driven decision-making capabilities. The system integrates ChatOpenAI as the underlying language model for both task planning and page content extraction, enabling the agent to understand web page structures, generate appropriate actions, and extract relevant information. The browser agent supports comprehensive web interactions including URL navigation, form filling, element clicking, scrolling, and content extraction. The integration addresses the complexity of modern web applications by providing semantic understanding of page content and intelligent action selection, enabling the agent to handle dynamic web pages, JavaScript-rendered content, and complex user interfaces that require contextual understanding.
Pipeline Workflow. The core pipeline implements a systematic multi-stage process for web interaction and task execution. The workflow begins with task initialization and report setup, where the agent initializes a Report instance to track the browser task execution, records the task description, and prepares for result documentation. This initialization stage is essential because it establishes a structured framework for capturing execution details, enabling comprehensive documentation and post-execution analysis. This is followed by browser agent execution, wherein the browser_use Agent is instantiated with the specified task and LLM configuration, then executes the task through intelligent web navigation and interaction. The browser agent operates with a maximum step limit (typically 50 steps) to ensure task completion within reasonable bounds, and employs sophisticated page understanding mechanisms to extract content and generate appropriate actions. During execution, the agent generates visual execution records (GIF animations) and conversation logs, providing detailed traces of the interaction process. The third stage involves result extraction, where the agent extracts the final results from the browser agentâs execution history. The system attempts multiple extraction strategies: first checking for extracted content summaries, then falling back to final results, and finally extracting from the last stepâs action results if available. This multi-strategy approach ensures robust result extraction even when the browser agentâs output format varies. Finally, the report generation stage uses the Reporter Agent to consolidate the task description and execution results into a structured markdown report with proper formatting. The report includes the original task, execution results, and references to generated execution records (GIF files and logs), ensuring comprehensive documentation of the browser interaction process.
Concurrent Execution Support. The browser use agent incorporates robust mechanisms for handling concurrent task execution. Each browser task execution is assigned a unique call identifier (call_id), which is used to create isolated subdirectories for execution artifacts, preventing file conflicts when multiple browser tasks run simultaneously. The system generates unique paths for GIF animations, conversation logs, and report files based on the call_id, ensuring that concurrent executions do not interfere with each other. This concurrent execution support is essential for multi-agent scenarios where multiple browser tasks may be initiated simultaneously, enabling scalable and reliable browser automation in distributed agent systems.
Execution Record Generation. The browser use agent automatically generates comprehensive execution records during task execution. The system creates visual execution traces in GIF format, capturing the sequence of browser interactions and page states throughout the task execution. Additionally, the agent saves detailed conversation logs that record all LLM interactions, action decisions, and page content extractions. These execution records provide valuable debugging information, enable post-execution analysis, and support transparency in browser automation tasks. The records are organized in per-call subdirectories, making it easy to trace specific task executions and analyze browser interaction patterns.
The browser use agentâs design emphasizes intelligent web automation and comprehensive documentation, enabling it to handle diverse web-based tasks ranging from simple information retrieval to complex multi-step interactions. The integration with browser_use library provides sophisticated web understanding capabilities, while the Reporter Agent ensures that all browser interactions are properly documented with execution traces, making the automation process transparent and verifiable.
### D.6 Tool Generator Agent
The tool generator agent is a specialized component designed for intelligent tool evolution through automated creation, dynamic retrieval, and systematic reuse mechanisms under the TCP. As illustrated in Figure 8, the agent implements a systematic pipeline workflow for intelligent tool lifecycle management that begins with task analysis and tool retrieval, followed by tool creation and evaluation, and tool registration in TCP.
<details>
<summary>x10.png Details</summary>

### Visual Description
\n
## Diagram: Tool Generator Agent Pipeline
### Overview
This diagram illustrates the pipeline of a Tool Generator Agent, detailing how user tasks are analyzed, tools are retrieved or created, and then executed and evaluated. The diagram shows a flow from a "Task" input through a series of processing steps, interacting with a "TCP" (Tool Communication Protocol) component.
### Components/Axes
The diagram is structured around two main sections: "TCP" (Tools) and "Pipeline". The "Pipeline" is a sequence of steps: "Analyze user tasks", "Retrieve TCP tools retrieval", "Create generate tool code", "Evaluate tool evaluation", and "Discard discard the tool". The "TCP" section lists available tools: "Read a URL", "Run a code", "Query a wiki page", and "Locate a position". There is also an "Execute" section with two sub-steps: "Get tool context information" and "Get parameters and execute tool". Arrows indicate the flow of control and data.
### Detailed Analysis or Content Details
The process begins with a "Task" (represented by a rounded rectangle on the left).
1. **Analyze user tasks:** The task is first analyzed. An arrow leads from this step to the "Retrieve TCP tools retrieval" step.
2. **Retrieve TCP tools retrieval:** This step checks if existing tools can fulfill the task. A "Yes" branch leads to the "Register" arrow, which connects to the "Execute" section. A "No" branch leads to the "Create generate tool code" step.
3. **Create generate tool code:** If no existing tool is suitable, a new tool is generated. This step leads to the "Evaluate tool evaluation" step.
4. **Evaluate tool evaluation:** The generated tool is evaluated. This step leads to the "Discard discard the tool" step.
5. **Discard discard the tool:** If the tool is not satisfactory, it is discarded.
6. **Execute:** If a tool is retrieved or created and evaluated successfully, it is executed. The "Execute" section has two steps:
* "Get tool context information"
* "Get parameters and execute tool"
7. **TCP (Tools):** The "TCP" section lists the available tools:
* "read" - Read a URL
* "code" - Run a code
* "query" - Query a wiki page
* "locate" - Locate a position
The "Reuse" arrow connects the "TCP" tools to the "Execute" section.
### Key Observations
The diagram highlights a closed-loop process where the agent attempts to reuse existing tools first. If reuse is not possible, it generates new tools, evaluates them, and discards those that are not suitable. The "Execute" section is a separate component that handles the actual execution of the selected tool. The diagram emphasizes the iterative nature of tool selection and generation.
### Interpretation
This diagram represents a sophisticated agent architecture designed for dynamic task completion. The agent doesn't rely on a fixed set of tools but can adapt by creating new tools when necessary. The evaluation and discard steps suggest a quality control mechanism to ensure that only effective tools are used. The separation of the "TCP" and "Pipeline" components indicates a modular design, allowing for easy addition or modification of tools without affecting the core pipeline logic. The diagram suggests a system capable of handling a wide range of tasks by leveraging both pre-existing and dynamically generated tools. The "Reuse" arrow is critical, indicating a preference for efficiency and avoiding redundant tool creation. The entire process is driven by the initial "Task" input, suggesting a goal-oriented approach to problem-solving.
</details>
Figure 8: Tool Generator Agent Workflow.
Pipeline Workflow. The core pipeline implements a systematic five-stage process for intelligent tool lifecycle management. The workflow begins with task analysis, where the agent analyzes task requirements and extracts tool specifications including tool name, description, parameter schema, and implementation plan. This is followed by tool retrieval, wherein the agent uses TCPâs semantic search to retrieve similar tools from the registry. If suitable existing tools are found, the agent evaluates their compatibility and returns the best match. The third stage involves tool creation, where the agent generates new tool implementations using LLM-based code generation when no suitable existing tools are found. The generated code follows the Tool base class pattern and includes proper error handling and logging. The fourth stage is tool evaluation, where the agent validates newly created tools by loading the tool class, checking for required attributes (name, description, __call__ method), and verifying structural correctness. Tools that fail evaluation are discarded, while successfully validated tools proceed to registration. Finally, the tool registration stage registers validated tools in TCP, which automatically handles version management, contract generation, and persistence to JSON manifests, making the tools immediately available to all agents through the unified TCP interface.
TCP Integration. The tool generator agent leverages TCP to provide comprehensive tool management capabilities. Through TCPâs semantic retrieval mechanism, the agent can efficiently search for existing tools based on functional similarity, avoiding redundant tool creation. When new tools are generated, TCPâs registration process automatically handles version tracking, contract documentation, and persistence, ensuring that all tools are properly managed and accessible across the multi-agent system. This TCP-based approach enables seamless tool sharing and reuse, supporting both local tool execution and distributed tool access through standardized interfaces.
The tool generator agentâs design emphasizes TCP-based tool management, enabling it to handle diverse tool requirements ranging from simple utility functions to complex domain-specific operations. The intelligent evolution process guarantees that the tool ecosystem continuously adapts to emerging requirements through systematic creation, validation, and reuse mechanisms.
### D.7 Reporter Agent
The Reporter Agent is a specialized component designed for managing and generating structured markdown reports with proper citation and reference management. As illustrated in Figure 9, the agent implements a systematic workflow for report management that encompasses two primary phases: (i) the Pipeline for incremental content addition and processing, and (ii) the Report generation process with automated citation management. The agent is widely used by other tools (deep researcher, deep analyzer, browser) to document their execution processes and findings.
<details>
<summary>x11.png Details</summary>

### Visual Description
\n
## Diagram: Reporter Agent Workflow
### Overview
This diagram illustrates the workflow of a "Reporter Agent," outlining the steps involved in completing a task, from adding research contexts to generating a final report. The diagram is segmented into a "Report" section, a "Pipeline" section, and feedback loops for "Next Round" and "Check Results." The overall structure is a flow diagram, showing the sequential progression of actions.
### Components/Axes
The diagram consists of the following key components:
* **Reporter Agent Header:** Top center, labeled "Reporter Agent".
* **Report Section:** Located on the left side, enclosed in a yellow box with the title "Report". Contains action buttons: "Add content", "Add citations", "Extract ref", "Generate report". Each button has a corresponding color: light blue, pink, light green, and orange respectively.
* **Pipeline Section:** Located in the center, enclosed in a yellow box with the title "Pipeline". Contains sequential steps: "Add research contexts", "Extract references", "Cite the references", "Generate report". Each step has a corresponding icon: a speech bubble with an equals sign, a document with a magnifying glass, a document with a quotation mark, and an upward-pointing arrow.
* **Execute Section:** Located on the right side, enclosed in a light blue box with the title "Execute". Contains two checkmark-labeled actions: "Add content to list" and "If check completed, generate report".
* **Feedback Loops:** Two sections at the bottom: "Next Round (Task & Summaries)" and "Check Results", connected by arrows indicating a cyclical process.
* **Task Input:** A rectangular box labeled "Task" on the far left, serving as the initial input to the system.
* **Cat Icon:** A small cat icon in the top-left corner.
### Detailed Analysis or Content Details
The diagram depicts a four-step pipeline:
1. **Add:** "Add research contexts" - represented by a speech bubble with an equals sign icon. This step receives input from the "Task" box.
2. **Extract:** "Extract references" - represented by a document with a magnifying glass icon. This step receives output from the "Add" step.
3. **Cite:** "Cite the references" - represented by a document with a quotation mark icon. This step receives output from the "Extract" step.
4. **Export:** "Generate report" - represented by an upward-pointing arrow icon. This step receives output from the "Cite" step.
The "Report" section provides actions that can be triggered:
* **Add content:** Light blue button.
* **Add citations:** Pink button.
* **Extract ref:** Light green button.
* **Generate report:** Orange button.
The "Execute" section shows the actions performed:
* **Add content to list:** Checkmark icon.
* **If check completed, generate report:** Checkmark icon.
The "Next Round (Task & Summaries)" and "Check Results" sections are connected to the "Export" step and the "Report" section, respectively, indicating a feedback loop for iterative refinement.
### Key Observations
The diagram emphasizes a cyclical workflow. The "Check Results" loop suggests a validation or review process, while the "Next Round" loop indicates the potential for iterative task completion and summarization. The clear separation of "Report" actions and the "Pipeline" steps suggests a distinction between manual interventions and automated processing. The use of icons and color-coding enhances the visual clarity of the diagram.
### Interpretation
The diagram illustrates a system designed for automated report generation, likely within a research or information gathering context. The "Reporter Agent" acts as an intermediary, taking a "Task" as input and transforming it into a structured "Report" through a defined "Pipeline." The feedback loops suggest a continuous improvement process, where results are checked, and the task is refined for subsequent rounds. The "Report" section allows for manual intervention, such as adding content or citations, while the "Pipeline" automates the core tasks of research context addition, reference extraction, and citation. The system appears to be designed for efficiency and accuracy in report creation, leveraging both automated processes and human oversight. The cat icon is likely a branding element and does not contribute to the functional understanding of the diagram.
</details>
Figure 9: Reporter Agent Workflow.
Pipeline Workflow. The Reporter Agent supports incremental content addition through the add action, which accepts content from multiple sources including text strings, dictionaries, and file paths. When content is added, the agent employs LLM-based extraction to automatically identify and structure three key components: (i) content, the main text preserving all citation markers in markdown link format [1](url), [2](url), etc.; (ii) summary, a concise 2-3 sentence summary of the content; and (iii) references, a list of reference items with IDs, descriptions, and URLs extracted from citations in the content. This automatic extraction ensures that citations are properly captured and linked to their sources, enabling systematic reference management throughout the report generation process.
Reference Management and Deduplication. The Reporter Agent implements sophisticated reference management mechanisms to ensure citation consistency and accuracy. When the complete action is invoked, the agent performs comprehensive reference processing: (i) collection, gathering all references from all content items; (ii) deduplication, merging duplicate references based on normalized keys (URLs are prioritized over descriptions for deduplication); (iii) renumbering, creating a unified reference mapping that renumbers all citations sequentially from 1; and (iv) URL generation, automatically generating proper URLs for references (converting file paths to file:// URLs, preserving HTTP/HTTPS URLs, and extracting URLs from descriptions when needed). This reference management ensures that all citations in the final report are properly numbered, deduplicated, and linked to their sources.
Generation and Completion. The final report generation process consolidates all content items into a coherent, well-structured markdown document. The agent uses LLM-based generation to organize content logically, integrate summaries for smooth transitions, and maintain proper citation formatting throughout the report. The generated report includes a complete References section at the end, listing all references in numerical order with proper URLs and descriptions. The agent ensures that all citations maintain the markdown link format [number](url) and that file paths are converted to absolute paths for proper rendering in markdown viewers. The final report is written to the specified file path with file locking mechanisms to ensure concurrent safety when multiple processes access the same report.
Integration with Other Tools. The Reporter Agent is designed to be seamlessly integrated with other tools through a unified interface. Tools such as deep researcher, deep analyzer, and browser use the Reporter Agent to document their execution processes, with each tool adding content items incrementally and completing the report when execution finishes. The agent supports per-call caching and locking mechanisms, enabling multiple concurrent report generations without conflicts. This integration ensures that all tools can generate comprehensive, properly cited reports that document their findings and execution traces, making the entire systemâs operations transparent and verifiable.
The Reporter Agentâs design emphasizes automatic citation management and structured report generation, enabling other tools to produce professional, well-documented reports without manual citation formatting. The LLM-based extraction and generation capabilities ensure that citations are properly identified, deduplicated, and formatted, while the reference management system guarantees consistency and accuracy across complex multi-source reports.
## Appendix E Detailed Analysis of Benchmark Results
### E.1 GAIA Benchmark
As shown in Figure Ë 3 and Table Ë 4, AgentOrchestra achieves state-of-the-art performance on GAIA (89.04% average) by mitigating the dimensionality curse and semantic drift that arise in large-scale agentic planning. We attribute this success to three complementary properties enabled by TEA and AgentOrchestra. First, hierarchical decoupling of the action space reduces planning complexity: hierarchical routing decomposes the global task into locally tractable sub-problems and assigns them to domain-appropriate sub-agents, preserving abstract reasoning under long horizons even amid low-level browser events. Second, ECP formalizes epistemic environment boundaries: GAIAâs multi-domain tasks require temporal and cross-modal state coherence, and baselines often degrade during domain transitions such as browser retrieval followed by local Python analysis. By treating environments as first-class managed components, TEA preserves and propagates session-critical state (e.g., authentication tokens, downloaded files, and transient file-system mutations) across agent boundaries, reducing contextual forgetting. Third, AgentOrchestra supports recursive refinement of reasoning trajectories and can invoke the Tool Generator to synthesize context-specific functionalities on demand, alleviating the fixed-capability bottleneck of static agent components.
Qualitatively, the Tool Generator is effective on tasks involving structured sources (e.g., Wikipedia or tabular data), where it can synthesize query wrappers and extraction utilities with clear I/O contracts. However, we observe limitations on fine-grained visual tasks (e.g., localizing specific colored digits or subtle UI elements), suggesting that tool synthesis alone cannot replace robust visual grounding and that tighter integration with vision-centric models remains necessary.
Across train and test, the Tool Generator produced over 50 TCP-registered tools spanning multiple domains, and approximately 30% were reused in subsequent tasks. This reuse rate indicates a practical balance between specialization and generalization, where the system expands capability coverage while retaining transferable utilities for recurring sub-problems.
On GAIA validation, AgentOrchestra achieves 92.45% on Level 1, 83.72% on Level 2, and 57.69% on Level 3 (82.42% average), consistently outperforming strong baselines such as AWorld (77.58%) and Langfun Agent (76.97%). Notably, the gap between Level 1 and Level 3 highlights the challenge of long-horizon, multi-domain tasks, where TEA-style environment state management and on-demand tool synthesis become increasingly important.
The key strength of AgentOrchestra lies in decomposing complex problems and flexibly assigning them to appropriate specialists. For example, in a Level 3 GAIA scenario that required extracting numerical data from an embedded table within a PDF and then performing multi-step calculations, the Planning Agent invoked the Browser Use Agent to locate and download the file, delegated parsing and verification to the Deep Analyzer, and then synthesized the final answer. When existing tools were inadequate, the Tool Generator created task-specific utilities (e.g., custom extractors for particular document layouts or scripts for bespoke computations), improving coverage and reliability. We note that frequent inter-agent exchanges can introduce latency and overhead; thus, the system minimizes unnecessary switching and motivates future work on adaptive routing and resource selection for improved efficiency and scalability.
### E.2 SimpleQA Benchmark
As shown in Table Ë 3, AgentOrchestra achieves state-of-the-art performance on SimpleQA with 95.3% accuracy, substantially outperforming monolithic LLM baselines (e.g., o3 at 49.4% and gemini-2.5-pro-preview-05-06 at 50.8%) and surpassing strong retrieval agents such as Perplexity Deep Research (93.9%). We attribute this improvement to systematic reduction of epistemic uncertainty through our hierarchical verification pipeline. Hallucinations in short-form factuality often arise from conflicting web-based evidence or reliance on internal parametric memory. AgentOrchestra mitigates these issues by enforcing cross-agent consensus: the Planning Agent orchestrates a retrieve-verify-synthesize cycle where the Deep Researcher performs multi-engine breadth-first searches while the Deep Analyzer evaluates evidence consistency across heterogeneous sources. By decoupling retrieval from analysis, the system prevents the confirmation bias inherent in single-agent architectures. Furthermore, the integration with the Reporter Agent ensures traceable attribution, grounding every factual claim in a re-verified source.
### E.3 HLE Benchmark
On HLE, AgentOrchestra achieves 37.46%, outperforming strong baselines including o3 (20.3%) and Perplexity Deep Research (21.1%). This gain highlights the frameworkâs capacity for long-horizon analytical reasoning and adaptive capability expansion in expert-level domains. HLE demands synthesizing disparate, highly specialized knowledge beyond simple retrieval. We attribute the gain to hierarchical coordination and adaptive capability expansion. The Planning Agent maintains global objective coherence, prunes the hypothesis space via delegation to specialists, and assigns technical validation to agents such as the Deep Analyzer. As a result, the final solution is both analytically rigorous and cross-verified against multimodal evidence, yielding robust performance on challenging expert-level tasks.
### E.4 Ablation Studies and Efficiency Analysis
Sub-agent Contribution Analysis. As detailed in Table 4, we evaluate the incremental contribution of each specialized sub-agent (Planning, Researcher, Browser, Analyzer, and Tool Generator). The synergy between these components is most evident in network-dependent tasks. While the baseline Planning agent (P) achieves 36.54% accuracy, the integration of coarse-grained retrieval via the Deep Researcher (R) and fine-grained interaction via the Browser (B) elevates performance to 72.76%. This doubling of efficacy underscores the complementarity of high-level information gathering and low-level DOM manipulation. Furthermore, the Deep Analyzer (A) provides an 8.67% improvement by resolving complex multi-step reasoning within documents and media, while the Tool Generator (T) adds a final 12.61% boost. This last leap validates that static toolkits are insufficient for the diverse requirements of GAIA, necessitating on-demand tool synthesis to bridge functional gaps.
System Efficiency and Resource Consumption. We analyze the operational efficiency of AgentOrchestra across varying task complexities, measured by wall-clock latency and token throughput. Simple factual queries typically complete within 30 seconds (approx. 5k tokens), while medium-complexity research tasks average 3 minutes (approx. 25k tokens). High-complexity multimodal or long-horizon scenarios require approximately 10 minutes and 100k tokens. Compared to monolithic baselines, our hierarchical architecture optimizes resource allocation by invoking resource-intensive sub-agents only when necessary. This ensures that AgentOrchestra maintains operational costs comparable to commercial research agents while delivering significant performance gains on complex tasks.
Self-Evolution Effectiveness. To evaluate the TEA Protocolâs support for iterative optimization, we assess self-evolution mechanisms on mathematical reasoning benchmarks (GPQA-Diamond, AIME24/25). As summarized in Table 5, iterative refinement via gradient-based (TextGrad) and symbolic (self-reflection) approaches consistently improves reasoning performance. For instance, using gpt-4.1 with self-reflection boosts AIME25 accuracy from 20.00% to 33.34%, highlighting the benefits of recursive trajectory refinement. By leveraging execution feedback via TEAâs versioning and tracer mechanisms, the system identifies and corrects logical inconsistencies in its planning. Overall, this shifts reasoning from one-shot inference to a managed optimization process, enabling AgentOrchestra to evolve robust problem-solving strategies for frontier-level tasks.
## Appendix F Case Studies
In this section, we systematically present representative cases of AgentOrchestra, accompanied by critical analyses to elucidate the underlying factors contributing to these outcomes. We primarily showcase the performance on the GAIA validation set, categorized by both difficulty Level 1, Level 2, and Level 3 and data type, including text, image, audio, video, spreadsheet, ZIP archive, and other file types.
Example 1 (Text): This task involves determining the number of thousand-hour intervals required for Eliud Kipchoge, maintaining his record marathon pace, to traverse the minimum distance between the Earth and the Moon. The task is categorized as Level 1 in difficulty, requires no supplementary files, and depends on the agentâs capacity for internet-based information retrieval, browser navigation, and computational analysis.
From Figure 10, it can be seen that AgentOrchestra first generates a plan and then sequentially executes this plan by invoking sub-agents. The browser_use_agent subsequently acquires key information, including Eliud Kipchogeâs marathon world record (2:01:09, Berlin Marathon, 25 September 2022, as documented by Wikipedia) and the minimum perigee distance of the Moon (356,400 km, per Wikipediaâs Moon article). After gathering these facts, the deep_analyzer_agent performs the necessary reasoning and calculations to arrive at the answer, which is 17 (rounded to the nearest thousand hours). Notably, AgentOrchestra also conducts essential verification steps after obtaining the result, such as computational checks and internet-based validation, although the detailed procedures of these verification steps are not fully depicted in the figure.
<details>
<summary>x12.png Details</summary>

### Visual Description
\n
## Diagram: Multi-Agent Workflow for Question Answering
### Overview
The image depicts a diagram illustrating a multi-agent workflow designed to answer a complex question: "If Elon Kipchoge could maintain his current record-breaking pace indefinitely, how many hours would it take him to travel the distance between the Earth and the Moon?". The workflow involves four agents: Planning Agent, Browser Use Agent (appearing twice), and Deep Analyzer Agent. The diagram shows the task assigned to each agent, the interactions between them, and the final answer.
### Components/Axes
The diagram is structured as a horizontal flow, with agents represented as rounded rectangles. Each agent's box contains:
* **Task:** A description of the agent's assigned task.
* **Multi-turn Interaction:** A visual representation of the agent's interaction with a browser, depicted as a series of smaller rectangles and images.
* **Result:** The output or finding of the agent.
* **Arrows:** Indicate the flow of information between agents.
* **Final Answer:** A blue rectangle on the right side displays the final answer to the question.
### Detailed Analysis or Content Details
**1. Planning Agent (Leftmost)**
* **Task:** "Calculate time for Elon Kipchoge to travel distance between Earth and Moon at marathon pace (3h, 25min, 45sec/marathon)."
* **Progress:** "65 steps completed (3/3 rotations)".
* **Status:** "Completed. (Time progress, rotated, 3 rotations)"
* **Steps:**
1. "Use browser agent to find Elon Kipchoge's marathon record time from relevant web page."
2. "Use browser agent to find the distance between the Earth and the Moon from Wikipedia page."
3. "Use deep analyzer agent to calculate how long Kipchoge takes to travel that distance."
4. "Convert answer to hours."
5. "Round the final answer to the nearest 1000 hours."
**2. Browser Use Agent (First Instance - Center-Left)**
* **Task:** "Find Elon Kipchoge's marathon record time according to the Wikipedia page."
* **Multi-turn Interaction with Browser:** A complex visual representation of browser interactions, including images and text snippets.
* **Result:** "Elon Kipchoge's marathon record time is 2:01:09 (2 hours, 1 minute, 9 seconds) as of September 30, 2023 (data: 2023-09-30)."
**3. Browser Use Agent (Second Instance - Center-Right)**
* **Task:** "Find the distance between the Earth and the Moon on the Wikipedia page."
* **Multi-turn Interaction with Browser:** A complex visual representation of browser interactions, including images and text snippets.
* **Result:** "Average distance between the Earth and the Moon is 384400 km (data: 2023-10-04)."
**4. Deep Analyzer Agent (Right-Center)**
* **Task:** "Calculate time to travel."
* **Details:**
* "Elon Kipchoge current record: 2:01:09 (7269 seconds)"
* "Average distance Earth to Moon: 384400 km"
* "Time to cover 384400 km with calculated time/km: 530.66 hours"
* **Result:**
* "To determine how many thousand hours it would take Elon Kipchoge to travel the distance between the Earth and the Moon at his record pace, we need to calculate the time it takes."
* "Final Calculation Check:"
* "Speed = 42.195 km / 7269 = 0.005800 km/s"
* "Distance = 384400 km"
* "Time = 384400 / 0.005800 = 66275862.07 seconds"
* "66275862.07 / 3600 = 18409.96 hours"
* "Rounding 18409.96 hours to the nearest 1000 hours: 19000 hours"
* "This is 19 thousand hours."
* "Final Answer: 19"
**5. Final Answer (Far Right)**
* Displays the number "17" within a blue rectangle. This seems to be an error, as the Deep Analyzer Agent calculates 19.
### Key Observations
* The workflow is sequential, with each agent completing its task before passing the result to the next.
* The Deep Analyzer Agent performs the core calculation, utilizing data obtained from the Browser Use Agents.
* There is a discrepancy between the calculated answer (19000 hours, rounded to 19) and the final answer displayed (17). This suggests a potential error in the final step or a separate rounding process.
* The diagram provides timestamps for the data retrieved from Wikipedia, indicating the data's freshness.
### Interpretation
The diagram demonstrates a complex question-answering system leveraging multiple specialized agents. The system effectively breaks down a complex problem into smaller, manageable tasks, assigning each task to an appropriate agent. The use of browser agents allows the system to access and retrieve information from external sources, while the deep analyzer agent performs the necessary calculations. The discrepancy in the final answer highlights the importance of verifying the entire workflow and identifying potential sources of error. The diagram illustrates a sophisticated approach to automated reasoning and knowledge integration, showcasing the potential of multi-agent systems for tackling complex real-world problems. The inclusion of timestamps suggests a focus on data validity and reproducibility. The "progress" and "status" indicators for the Planning Agent suggest an iterative or step-by-step approach to problem-solving.
</details>
Figure 10: Execution trajectory of AgentOrchestra for Example 1.
Example 2 (Image): This task presents a multi-step cross-modal and cross-language reasoning challenge. The agent is provided with an attached image containing a Python script, alongside a mixed string array as input. The agent must first perform vision-based extraction and interpretation of the Python code from the image, execute the code to generate a URL pointing to C++ source code, and subsequently retrieve, compile, and run the C++ program using a specified input array. The final answer is derived by reasoning over the programâs output. This task is designated as Level 2 in difficulty, includes a supplementary file, and comprehensively evaluates the agentâs capabilities in visual code extraction, internet-based retrieval, automated code execution, and multi-stage reasoning.
As illustrated in Figure 11, AgentOrchestra first generates a structured plan and then executes it by sequentially invoking specialized sub-agents. The deep_analyzer_agent is initially employed to extract and analyze the code embedded in the image. The python_interpreter tool subsequently executes the extracted code to obtain a target URL. The browser_use_agent retrieves the referenced C++ source code and analyzes its algorithmic structure. Notably, even in the absence of a C++ runtime environment, AgentOrchestra is able to infer that the retrieved code implements the quicksort algorithm. Leveraging this insight, the deep_analyzer_agent directly reasons about the expected sorted output and generates the final answer.
<details>
<summary>x13.png Details</summary>

### Visual Description
\n
## Diagram: Multi-Agent System Workflow
### Overview
The image depicts a workflow diagram illustrating a multi-agent system designed to solve a coding problem. The system consists of three primary agents: a Planning Agent, a Deep Analyzer Agent (appearing twice), and a Browser Use Agent. The workflow involves generating Python code, analyzing it, potentially using a browser to find solutions, and finally, compiling the code to C++. The diagram shows the tasks, inputs, and outputs of each agent, along with the flow of information between them. A "Final Answer" box is present at the far right, with the number "47" inside.
### Components/Axes
The diagram is structured into four main rectangular blocks, each representing an agent or the final answer. Each block contains text describing the agent's task, input, and output. Arrows indicate the flow of information between the agents. The diagram also includes a "Task ID" at the top-left: `ID: B7857f4-0b0a-4387-a12a-0a44d50b86`.
### Detailed Analysis or Content Details
**1. Planning Agent (Leftmost Block)**
* **Task:** Plan Python Generative Code Details & Array Problem (20 step plan, system prompt)
* **Input:**
* `Step 0/9` (appears multiple times, indicating progress)
* `Steps:` followed by a list of numbered steps.
* `1. [Loop setup]: create an array to store the results.`
* `2. [Loop setup]: create an array to store the results.`
* `3. [Loop setup]: create an array to store the results.`
* `4. [Loop setup]: create an array to store the results.`
* `5. [Loop setup]: create an array to store the results.`
* `6. [Loop setup]: create an array to store the results.`
* `7. [Loop setup]: create an array to store the results.`
* `8. [Loop setup]: create an array to store the results.`
* `9. [Loop setup]: create an array to store the results.`
* `10. [Loop setup]: create an array to store the results.`
* `11. [Loop setup]: create an array to store the results.`
* `12. [Loop setup]: create an array to store the results.`
* `13. [Loop setup]: create an array to store the results.`
* `14. [Loop setup]: create an array to store the results.`
* `15. [Loop setup]: create an array to store the results.`
* `16. [Loop setup]: create an array to store the results.`
* `17. [Loop setup]: create an array to store the results.`
* `18. [Loop setup]: create an array to store the results.`
* `19. [Loop setup]: create an array to store the results.`
* `20. [Loop setup]: create an array to store the results.`
* **Output:** Python code (details not fully visible, but includes `array_length = 12`)
**2. Deep Analyzer Agent (First Instance - Center-Left Block)**
* **Task:** Deep Analyzer Agent: Analyze the Python code against the provided array strings to get a URL.
* **Result:**
* `Java things = "my"`
* `Java things = "my"`
* `Java things = "my"`
* `Java things = "my"`
* `Java things = "my"`
* `URL: https://stackoverflow.com/questions/4488005/sorting-algorithm-quick-sort-and-retrieve-the-c-code-note-the-quickest-algorithm-i-need-to-the-complete-c-code-that`
* **Note:** The repeated "Java things = "my"" is likely an artifact or placeholder.
**3. Browser Use Agent (Center Block)**
* **Task:** Browser Use Agent: Execute the extracted Python against the provided URL strings to get a URL.
* **Content:** A screenshot of a webpage (likely Stack Overflow) with code snippets. The code appears to be a QuickSort implementation.
* **Note:** The text below the screenshot reads: "To complete and parse that URL, you must clarify the specific algorithm you are looking for. The current URL is not sufficiently specific to return a compilable C++ code."
* **Result:** `https://www.geeksforgeeks.org/quick-sort/`
**4. Deep Analyzer Agent (Second Instance - Center-Right Block)**
* **Task:** Deep Analyzer Agent: Compile code and use the C++ code generated from the array (12, 3, 5, 7, 4, 8, 9, 2, 1, 6, 10, 11).
* **Result:**
* `The array sorted is: [2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 1]`
* `Array remaining is: 0-based`
* `1st Integer (index 0): 12`
* `2nd Integer (index 1): 3`
* `3rd Integer (index 2): 5`
* `4th Integer (index 3): 7`
* `5th Integer (index 4): 4`
* `6th Integer (index 5): 8`
* `7th Integer (index 6): 9`
* `8th Integer (index 7): 2`
* `9th Integer (index 8): 1`
* `10th Integer (index 9): 6`
* `11th Integer (index 10): 10`
* `12th Integer (index 11): 11`
* `Sum of the 12th and 11th integers is: 21`
**5. Final Answer (Far-Right Block)**
* Contains the number "47".
### Key Observations
* The system attempts to solve an array problem, likely sorting.
* The Planning Agent generates a series of steps, all of which are identical ("Loop setup"). This suggests a potential issue or placeholder in the planning process.
* The Deep Analyzer Agent uses a URL to find relevant code (initially Stack Overflow, then GeeksforGeeks).
* The final Deep Analyzer Agent compiles the code and provides the sorted array, along with individual element values and a sum.
* The sorted array is not in ascending order.
* The "Final Answer" of "47" is not directly related to the array problem and its purpose is unclear.
### Interpretation
The diagram illustrates a complex workflow for automated code generation and problem-solving. The system leverages multiple agents to plan, analyze, search for solutions, and compile code. However, the repeated "Loop setup" steps in the Planning Agent and the unsorted array in the final result suggest potential flaws in the system's logic or implementation. The system appears to be attempting to implement a sorting algorithm (QuickSort), but the final result is incorrect. The inclusion of the "Final Answer" box with the number "47" is ambiguous and requires further context. The system demonstrates a proof-of-concept for a multi-agent coding system, but requires refinement to achieve accurate and reliable results. The reliance on external resources (Stack Overflow, GeeksforGeeks) highlights the importance of information retrieval in automated code generation. The system's ability to extract and utilize code snippets from these resources is a key strength.
</details>
Figure 11: Execution trajectory of AgentOrchestra for Example 2.
Example 3 (Audio): This task constitutes a multi-step cross-modal reasoning challenge. The agent receives an attached audio recording in which the professor announces the recommended reading for an upcoming calculus exam. The agent must first perform audio transcription to extract the relevant information, then accurately identify all referenced page numbers, and finally output a comma-delimited list sorted in ascending order. This task is classified as Level 1 in difficulty, includes a supplementary audio file, and comprehensively tests the agentâs proficiency in speech-to-text transcription, semantic information extraction, and precise data organization.
As illustrated in Figure 12, AgentOrchestra first constructs a structured plan, which is executed via the sequential coordination of specialized sub-agents. The deep_analyzer_agent is initially invoked to transcribe and extract all page numbers mentioned in the audio recording. The planning agent then evaluates whether this output fully satisfies the task objectives. If so, the workflow is terminated early, with each stepâs outcome recorded accordingly, thereby avoiding unnecessary sub-agent invocations. Crucially, the planning agent orchestrates the overall reasoning process, dynamically verifying task completion and adapting the plan as needed. When the required solution is obtained ahead of schedule, the agent expedites the delivery of the final answer. Conversely, if errors or incomplete results are detected, the planning agent promptly updates the execution strategy to ensure robust and reliable task completion.
<details>
<summary>x14.png Details</summary>

### Visual Description
## Diagram: Multi-Agent System for Page Number Extraction
### Overview
The image depicts a diagram illustrating a multi-agent system designed to extract page numbers from an audio recording of a Calculus midterm. The system consists of a "Deep Analyzer Agent" and multiple "Planning Agents" working in a sequential manner. The diagram shows the flow of information and tasks between these agents, culminating in a "Final Answer."
### Components/Axes
The diagram is structured around several key components:
* **Task ID:** "1693:5993a7d6-457b-467b-a649-33359e619bf3, Level 1" (Top-left)
* **Question:** A text block outlining the user's request.
* **Attached File:** "169789565-8705-457b-a649-368869b78b10.mp3"
* **Deep Analyzer Agent:** A central blue box labeled "Deep Analyzer Agent."
* **Planning Agents:** Multiple green boxes labeled "Planning Agent."
* **Final Answer:** A yellow box labeled "Final Answer."
* **Arrows:** Indicate the flow of information and tasks between agents.
* **Progress Bars:** Within the Planning Agent boxes, indicating task completion.
* **Status Indicators:** Within the Planning Agent boxes, showing the state of the task (Completed, Blocked, or Not Started).
* **Result:** Text output from the Deep Analyzer Agent.
* **Notes:** Text output from the Planning Agent.
### Detailed Analysis or Content Details
**1. Question:**
"Hi, I have quick call from my classes on Friday, so I'm trying to figure out what to study for my Calculus midterm next week. My friend came across me an audio recording of Professor Willcockson giving out the recommended reading for the test, but my headphones broke mid-recording so I couldn't quite listen to the recording for the rest and tell me the page numbers it's supposed to go over? I've attached a file called homework9.mp3 that has the recording. Please provide just the page numbers as a comma-delimited list. And please provide the list in ascending order."
**2. Attached File:**
"169789565-8705-457b-a649-368869b78b10.mp3"
**3. Deep Analyzer Agent:**
* **Task:** "Analyze the attached audio file '169789565-8705-457b-a649-368869b78b10.mp3' to extract page numbers mentioned by Professor Willcockson for the Calculus midterm."
* **Result:** "Okay, I listened to the recording for you. Based on Professor Willcockson's recommendations, here are the page numbers you should go over."
* 132, 133, 134, 197, 245
**4. Planning Agent 1 (Leftmost):**
* **Plan:** "Extract Page Numbers from Calculus Professor Audio Recording (ID: calculate_audio_analysis)"
* **Progress:** 5/5 steps completed (100.0%)
* **Status:** Completed, 0 in progress, 0 blocked, 0 not started
* **Notes:**
* 0. [Plan: Extract Page Numbers from Calculus Professor Audio Recording (ID: calculate_audio_analysis)] - Completed
* 1. [Plan: Use deep analyzer agent to analyze the attached audio file '169789565-8705-457b-a649-368869b78b10.mp3'] - Completed
* 2. [Plan: Notes: Succesfully analyzed the audio and recovered the page numbers.] - Completed
* 3. [Plan: Extract page numbers: 132, 133, 134, 197, 245] - Completed
* 4. [Plan: Format page numbers into a comma-delimited list] - Completed
**5. Planning Agent 2 (Middle):**
* **Plan:** "Extract Page Numbers from Calculus Professor Audio Recording (ID: calculate_audio_analysis)"
* **Progress:** 5/5 steps completed (100.0%)
* **Status:** Completed, 0 in progress, 0 blocked, 0 not started
* **Notes:**
* 0. [Plan: Extract Page Numbers from Calculus Professor Audio Recording (ID: calculate_audio_analysis)] - Completed
* 1. [Plan: Use deep analyzer agent to analyze the attached audio file '169789565-8705-457b-a649-368869b78b10.mp3'] - Completed
* 2. [Plan: Notes: Succesfully analyzed the audio and recovered the page numbers.] - Completed
* 3. [Plan: Extract page numbers: 132, 133, 134, 197, 245] - Completed
* 4. [Plan: Format page numbers into a comma-delimited list] - Completed
**6. Planning Agent 3 (Rightmost):**
* **Plan:** "Extract Page Numbers from Calculus Professor Audio Recording (ID: calculate_audio_analysis)"
* **Progress:** 5/5 steps completed (100.0%)
* **Status:** Completed, 0 in progress, 0 blocked, 0 not started
* **Notes:**
* 0. [Plan: Extract Page Numbers from Calculus Professor Audio Recording (ID: calculate_audio_analysis)] - Completed
* 1. [Plan: Use deep analyzer agent to analyze the attached audio file '169789565-8705-457b-a649-368869b78b10.mp3'] - Completed
* 2. [Plan: Notes: Succesfully analyzed the audio and recovered the page numbers.] - Completed
* 3. [Plan: Extract page numbers: 132, 133, 134, 197, 245] - Completed
* 4. [Plan: Format page numbers into a comma-delimited list] - Completed
**7. Final Answer:**
"132, 133, 134, 197, 245"
### Key Observations
* The system utilizes multiple Planning Agents performing the same task, likely for redundancy or parallel processing.
* All Planning Agents report 100% completion and successful analysis.
* The Deep Analyzer Agent successfully extracted the page numbers: 132, 133, 134, 197, and 245.
* The Final Answer matches the result provided by the Deep Analyzer Agent.
### Interpretation
The diagram illustrates a robust system for extracting information from audio data. The use of multiple Planning Agents suggests a design focused on reliability and fault tolerance. The sequential flow of tasks, from the initial question to the final answer, demonstrates a clear and structured approach to problem-solving. The detailed notes within each Planning Agent provide transparency into the execution of each step. The system successfully addresses the user's request by accurately identifying and formatting the requested page numbers. The repetition of the same plan across multiple agents suggests a microservices architecture where each agent is responsible for a specific task, and the system leverages multiple instances of that service for increased resilience. The consistent output across all agents confirms the accuracy and reliability of the system.
</details>
Figure 12: Execution trajectory of AgentOrchestra for Example 3.
Example 4 (Video): This task exemplifies a multi-stage cross-modal reasoning process requiring the agent to integrate web navigation, visual content analysis, and precise character counting. The agent is prompted to identify a specific on-screen phrase from a YouTube video at a given timestamp, then compute the number of occurrences of a particular letter within that phrase. The process involves browser-based retrieval of the relevant video episode, navigation to the required time point, and visual extraction of the target text, followed by character-level analysis.
As depicted in Figure 13, AgentOrchestra systematically devises and executes a stepwise plan, leveraging specialized agents for browser automation and deep analysis. Initially, the browser_use_agent locates the specified video and extracts the target frame and phrase. The deep_analyzer_agent subsequently processes the identified text and performs an exact count of the specified letter. Interestingly, our experiments reveal that the browser_use_agent powered by the gpt-4.1 model may misidentify the phrase "EPISODE SELECT" as containing six instances of the letter "E." However, the subsequent deep_analyzer_agent is able to perform a more fine-grained analysis, correctly determining the answer to be four, thereby rectifying the earlier modulesâ errors.
<details>
<summary>x15.png Details</summary>

### Visual Description
\n
## Multi-Panel Screenshot: Task Decomposition for Question Answering
### Overview
This image presents a multi-panel screenshot illustrating a task decomposition process for a question-answering system. The system appears to be designed to answer questions about video content, specifically related to the *Sonic the Hedgehog* game series. The screenshot showcases the workflow across several "Agent" modules: Planning Agent, Browser User Agent (two instances), and Deep Analyzer Agent, culminating in a "Final Answer". Each agent has a defined task and outputs a "Result".
### Components/Axes
The screenshot is divided into four main rectangular panels, arranged horizontally. Each panel represents a different agent in the workflow. Within each panel, there are text blocks describing the task, progress (for the Planning Agent), results, and input/output. There are also embedded screenshots of web browser windows within the Browser User Agent panels. The "Final Answer" is highlighted with an arrow.
### Detailed Analysis or Content Details
**Panel 1: Planning Agent**
* **Task ID:** 1919593-816d-47c6-a649-393959b7b. Level 1
* **Question:** "You Tube video Game Orange Sonic begins play in the game Sonic the Hedgehog (2006) in the year 2012. They towards into the first episode, a phrase is shown in white letters in red background. How many times does the letter 'n' appear in this phrase?"
* **Progress:** 95% complete (0.79s)
* **Status:** Completed, Okay progress, Good result
* **Sub-tasks:**
1. "I become one agent to search for and locate the Game Orange Sonic regarding "C0066.info" plays for year 2012, specifically the first episode."
2. "I become one agent to identify the first episode and extract text from the red background with white letters."
3. "I become one agent to analyze the extracted text and count the occurrences of the letter 'n'."
4. "I become one agent to formulate the final answer."
* **Input:** (Not explicitly shown, implied to be the initial question)
* **Output:** "task is decomposed into sub-tasks"
**Panel 2: Browser User Agent (Left)**
* **Task:** Search for the Game Orange Sonic, specifically the 2006 Let's Play series on YouTube. Locate the first episode.
* **Multi-turn Interactions With Browser:** Screenshot of a YouTube search results page for "Sonic Complete Series". Several thumbnails are visible.
* **Result:** "Game Orange Sonic 2006 episode 1 (2:12:34) - gameplay by leader. Game Orange Sonic 2006 episode 2 (2:03:58) - gameplay by leader. Game Orange Sonic 2006 episode 3 (1:39:58) - gameplay by leader."
* **PART 1:** Orange Sonic 2006, Let's Play episode 1 by leader.
* **PART 2:** EPISODE SELECT. The player goes into EPISODE SELECT.
* **PART 3:** Orange Sonic 2006, Let's Play episode 1 by leader.
* **PART 4:** EPISODE SELECT. The player goes into EPISODE SELECT.
**Panel 3: Browser User Agent (Right)**
* **Task:** Go to [https://www.youtube.com/watch?v=AdGhXNf8s](https://www.youtube.com/watch?v=AdGhXNf8s). Play the video and observe what is shown on the screen accordingly to this episode. Identify the phrase in white letters on a red background and extract that text.
* **Multi-turn Interactions With Browser:** Screenshot of a YouTube video player displaying gameplay of *Sonic the Hedgehog* (2006). The video is paused, showing a menu screen with text.
* **Result:** "As approximately 30 seconds into the video Game Orange Sonic - The Hedgehog (2006) - Part 1 - Gameplay by leader, the screen shows a menu titled EPISODE SELECT. The player goes into EPISODE SELECT. The menu displays several options: FREE GAME, EPISODE SELECT, UPGRADE SELECT. The phrase "EPISODE SELECT" appears in white letters on a red background."
**Panel 4: Deep Analyzer Agent**
* **Task:** "I have already identified the phrase as "EPISODE SELECT". Please carefully analyze the phrase and count how many times the letter "n" appears in it. Count each occurrence of the letter "n" including, notably, any appearance of lowercase "n"."
* **Result:**
* The phrase is "EPISODE SELECT" (length 12).
* Total count of letter 'n': 2
* **Character-by-character analysis:**
* Position 1: E - No 'n'
* Position 2: P - No 'n'
* Position 3: I - No 'n'
* Position 4: S - No 'n'
* Position 5: O - No 'n'
* Position 6: D - No 'n'
* Position 7: E - No 'n'
* Position 8: - No 'n'
* Position 9: S - No 'n'
* Position 10: E - No 'n'
* Position 11: L - No 'n'
* Position 12: E - No 'n'
* Position 13: C - No 'n'
* Position 14: T - No 'n'
* **Answer:** 2
**Final Answer:**
* Highlighted with an arrow pointing to the number: **132, 133, 134, 197, 245** (This appears to be an error, as the Deep Analyzer Agent determined the answer to be 2).
### Key Observations
* The system successfully decomposes the complex question into smaller, manageable sub-tasks.
* The Browser User Agent effectively utilizes web search and video playback to locate the relevant information.
* The Deep Analyzer Agent accurately identifies and counts the occurrences of the letter "n" in the extracted phrase.
* There is a discrepancy between the Deep Analyzer Agent's result (2) and the "Final Answer" (132, 133, 134, 197, 245), suggesting a potential error in the final aggregation or post-processing step.
* The system provides detailed, step-by-step reasoning for each agent's actions.
### Interpretation
This screenshot demonstrates a sophisticated question-answering pipeline leveraging multiple specialized agents. The system's ability to break down a complex question, utilize external tools (web browser, YouTube), and perform detailed analysis suggests a robust and adaptable architecture. The discrepancy in the final answer highlights the importance of careful error checking and validation in such systems. The detailed character-by-character analysis in the Deep Analyzer Agent is a good example of explainable AI, providing transparency into the reasoning process. The system appears to be designed for answering questions about video content, specifically focusing on identifying and analyzing textual elements within the video. The use of "PART" labels within the Browser User Agent results suggests a potential strategy for segmenting and processing long-form video content. The overall workflow is a clear example of a modular, agent-based approach to complex problem-solving.
</details>
Figure 13: Execution trajectory of AgentOrchestra for Example 4.
<details>
<summary>x16.png Details</summary>

### Visual Description
\n
## Document Analysis: Task Decomposition & Food Item Identification
### Overview
The image presents a multi-panel document outlining a task decomposition process for identifying a unique food item within a spreadsheet. It showcases the interaction between three agents: a Planning Agent, a Deep Extractor Agent, and a Deep Analyzer Agent, culminating in a Final Answer. The document includes a task description, agent workflows, and a data table representing a food item list.
### Components/Axes
The document is structured into four main sections:
1. **Task Information (Top-Left):** Contains the Task ID, Question, and Attached File information.
2. **Agent Workflow (Center-Left & Center):** Displays the workflow of the Planning Agent and the Deep Extractor/Analyzer Agents, including progress indicators and reasoning steps.
3. **Data Table (Bottom-Center):** A table listing food items with associated categories.
4. **Final Answer (Top-Right):** Presents the identified food item ("Soups and Stews").
The data table has two columns: "item name" and "categories".
### Detailed Analysis or Content Details
**1. Task Information:**
* **Task ID:** 9b549f59-3ee-4a14-b62f-d130ea0d0171, Level 3
* **Question:** "Which of the list elements under CATEGORIES in the XML would contain the one food item in the spreadsheet that does not appear a second time under a different name?"
* **Attached File:** 9b549f59-3ee-4a14-b62f-d130ea0d0171.7p
**2. Planning Agent:**
* **Plan:** Analyze XML and Spreadsheet to find Unique Food Item (2/2 food_analysis plan)
* **Progress:** 0/5 steps completed (0.0%)
* **Status:** Completed: 0 progress, Blocked: 0 not started
* **Steps:**
1. Use deep_agent_analyzer to extract and analyze the attached 7p file (9b549f59-3ee-4a14-b62f-d130ea0d0171.7p) to understand its contents including XML and spreadsheet files.
2. Use deep_agent_analyzer to parse the spreadsheet and identify all food items, noting any that appear multiple times under different names.
3. Use deep_agent_analyzer to parse the XML and identify the CATEGORIES, relating each food item to its assigned category.
4. Use deep_agent_analyzer to identify the food item that appears only once (does not have a synonym) and determine its corresponding category.
5. Use a final_answer_agent to output the category of the unique food item.
**3. Deep Extractor Agent:**
* **Task:** Extract and analyze the attached 7p file located at 9b549f59-3ee-4a14-b62f-d130ea0d0171.7p to understand its contents including XML and spreadsheet files. Please examine what files are contained within this ZIP archive and provide details about their structure and content.
* **Result:** To determine which text element under CATEGORIES in the XML would contain the one food item in the spreadsheet that does not appear a second time under a different name, we need to perform the following steps.
* "List all food items from the spreadsheet and identify synonymous pairs."
* Let's go through the food_duplicates.csv sheet and identify pairs.
* dried cranberries = craisins
* pop = soda
* lime juice = key lime juice
* corn meal = polenta
* mushrooms = fungi
* beet = beetroot
* ground beef = mince
* navy beans = haricot beans
* tortilla = wrap
**4. Deep Analyzer Agent:**
* **Task:** Identify which food from the spreadsheet appears only once (does not have a second occurrence under a different name).
* **Result:**
* The spreadsheet contains 58 food items arranged in 11 x 5 cells.
* Almost every item is paired with a synonym that appears elsewhere in the sheet (e.g., "dried cranberries" = "craisins", "pop" = "soda", "tortilla" = "wrap", "beet" = "beetroot", etc.).
* Systematically matching all synonyms leaves exactly one entry with no second appearance under another name:
* "turtle soup" appears only once and has no synonymous counterpart.
* Looking at the list of food elements under CATEGORIES in the XML, the only category that can logically accommodate "turtle soup" is "Soups and Stews".
* **Answer:** The food item "turtle soup" appears only once in the spreadsheet and would logically fall under the category "Soups and Stews".
**5. Data Table (Food Items & Categories):**
The table contains the following entries (approximate, based on visible rows):
| Item Name | Categories |
| ------------------ | ----------------- |
| popcorn | snacks |
| mango | fruits |
| sausage | meats |
| rice | grains |
| noodles | grains |
| spinach | vegetables |
| peaches | fruits |
| turkey | meats |
| beets | vegetables |
| craisins | fruits |
| polenta | grains |
| fungi | vegetables |
| beetroot | vegetables |
| mince | meats |
| haricot beans | legumes |
| wrap | grains |
| turtle soup | Soups and Stews |
| soda | beverages |
| key lime juice | beverages |
**6. Final Answer:**
* **Soups and Stews**
### Key Observations
* The process is highly structured, involving multiple agents and a clear workflow.
* The core task revolves around identifying a unique food item based on synonym analysis.
* The data table provides the necessary information for the analysis, with clear item-category pairings.
* "Turtle soup" is identified as the unique item, and its category is determined to be "Soups and Stews".
* The agents systematically eliminate synonymous pairs to isolate the unique item.
### Interpretation
The document demonstrates a complex task decomposition approach to solve a specific information retrieval problem. The agents work in a pipeline, each performing a specific step in the process. The Deep Extractor Agent focuses on data extraction and synonym identification, while the Deep Analyzer Agent performs the logical reasoning to identify the unique item and its category. The Planning Agent orchestrates the entire process.
The identification of "turtle soup" as the unique item suggests that the dataset was designed to include a relatively uncommon food item that wouldn't have a common synonym within the provided list. This highlights the importance of considering edge cases and uncommon scenarios in data analysis. The entire process is a demonstration of how AI agents can be used to automate complex tasks that require both data extraction and logical reasoning. The use of a 7p file suggests a proprietary data format, and the document focuses on the logic of the analysis rather than the specifics of the file format itself.
</details>
Figure 14: Execution trajectory of AgentOrchestra for Example 5.
Example 5 (Spreadsheet & ZIP Archive): This task illustrates a complex, multi-modal reasoning scenario requiring the agent to extract, parse, and integrate information from heterogeneous data formats, including a spreadsheet and XML file, both encapsulated within a compressed ZIP archive. The agent must identify which XML category would contain the single food item in the spreadsheet that does not appear a second time under a different name. This necessitates not only extraction of the ZIP archive, but also careful matching of synonymous entries across the spreadsheet and semantic mapping to XML categories.
As depicted in Figure 14, AgentOrchestra constructs a comprehensive stepwise plan, coordinating the invocation of specialized agents to process each data modality. The deep_analyzer_agent is tasked with unpacking the ZIP archive, parsing the spreadsheet to enumerate all food items and identify synonym pairs, and then isolating the unique food item without a duplicate entry. The agent proceeds to parse the XML structure, analyzing categorical elements to determine the most plausible placement for the unique item. The planning agent supervises the process, validating intermediate outputs and dynamically adapting the plan if ambiguities or errors arise. This example showcases the agentâs proficiency in handling compressed archives, integrating tabular and structured data, and performing reliable, cross-format reasoning to derive an interpretable solution.
## Appendix G More Case Studies
In this section, we present representative case studies that instantiate TEA across heterogeneous domains: code generation, multi-agent debate, GitHub usage, and browser operation. Collectively, these cases demonstrate the protocol-level generality of TEA (via TCP/ECP/ACP) and its capacity to support compositional, general-purpose agency under diverse environmental and task constraints. Additional scenarios are currently under development, including computer game and mobile game environments, further expanding the frameworkâs applicability across diverse interactive domains.
### G.1 Code Generation
<details>
<summary>x17.png Details</summary>

### Visual Description
\n
## Flowchart: Python Prime Number Generation and Verification
### Overview
This flowchart illustrates the process of creating a Python file to generate prime numbers, verifying its content, attempting to import it, and executing it directly if the import fails. It also includes a final verification step to confirm the correctness of the generated prime list.
### Components/Axes
The flowchart consists of several rectangular blocks representing actions or processes, diamond shapes representing decision points, and rounded rectangles representing the start and end points. Arrows indicate the flow of execution. The blocks contain Python code snippets and textual descriptions.
### Detailed Analysis or Content Details
**1. Start (Green Rounded Rectangle - Top-Left)**
- Label: "Start"
**2. Create Python File (Yellow Rectangle - Left)**
- Label: "Create Python File"
- Code:
```python
echo 'def get_primes(n=100):
primes = []
for num in range(2, n):
for i in range(2, int(num**0.5) + 1):
if num % i == 0:
break
else:
primes.append(num)
return primes
if __name__ == "__main__":
print(get_primes(100))' > prime.py
```
**3. Verify File Content (Yellow Rectangle - Center-Left)**
- Label: "Verify File Content"
- Code:
```
cat prime.py
Output:
def get_primes(n=100):
primes = []
for num in range(2, n):
for i in range(2, int(num**0.5) + 1):
if num % i == 0:
break
else:
primes.append(num)
return primes
if __name__ == "__main__":
print(get_primes(100))
```
**4. Try Import (Yellow Rectangle - Right)**
- Label: "Try Import"
- Code:
```python
from prime import get_primes
primes = get_primes(100)
```
- Output:
```
Error: Import from prime is not allowed.
Authorized imports are: ['math', 're', ...]
```
**5. ImportSuccess? (Diamond - Center-Right)**
- Label: "ImportSuccess?"
- Two outgoing arrows:
- "Yes" (Green Arrow) leads to "Complete"
- "No" (Red Arrow) leads to "Execute Directly"
**6. Execute Directly (Yellow Rectangle - Bottom-Right)**
- Label: "Execute Directly"
- Code:
```python
python3 prime.py
```
- Output:
```
[2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71, 73, 79, 83, 89, 97]
```
**7. Verify Result (Yellow Rectangle - Bottom-Center)**
- Label: "Verify Result"
- Text: "Prime list is correct"
**8. Complete (Green Rounded Rectangle - Bottom-Right)**
- Label: "Complete"
### Key Observations
The flowchart demonstrates a security measure where direct imports from user-created Python files are disallowed. This is indicated by the "Error: Import from prime is not allowed" message. The process then falls back to executing the file directly, which successfully generates the prime number list. The final step verifies the correctness of the result.
### Interpretation
This flowchart illustrates a secure execution pipeline for Python code. The system attempts to import the generated code as a module, but a security policy prevents this. This policy likely exists to prevent malicious code from being executed through imports. The fallback mechanism of direct execution allows the code to run, but isolates it from the broader system, mitigating potential security risks. The verification step ensures that the executed code produces the expected output, confirming its functionality. The authorized imports list suggests a whitelist approach to module access. This is a common security practice in environments where untrusted code needs to be executed. The flowchart highlights a robust approach to code execution that prioritizes security while still allowing for functionality.
</details>
Figure 15: Case study of TEA agent for code generation.
This case study demonstrates the agentâs execution of a code generation task requiring the creation of a Python script that calculates prime numbers within 100 and returns them as a list. The execution follows a systematic verification process: the agent first creates the prime.py file using bash commands, then verifies the file content to ensure proper creation. Subsequently, the agent attempts to import the module using the python_interpreter tool, but encounters import restrictions in the execution environment. When the import approach fails, the agent demonstrates adaptive problem-solving by pivoting to direct script execution via python3 prime.py, which successfully produces the expected prime number list. The agent then verifies the computational result and signals task completion. This trajectory illustrates the agentâs capacity for systematic verification, graceful failure recovery, and alternative solution discovery when encountering environmental constraints.
### G.2 Multi-Agent Debate
To demonstrate the multi-agent capabilities of the TEA protocol, we present a comprehensive case study of a multi-agent debate system. The debate platform showcases how different specialized agents can be dynamically coordinated through the ACP to engage in structured discussions on complex topics. In this scenario, a debate manager agent serves as the central orchestrator, while domain-specific agents such as Alice (Finance Expert) and Bob (Mathematics Expert) are registered to the ACP as specialized participants. The debate manager agent leverages the ACP protocol to invite and coordinate these expert agents, establishing a structured debate environment where each agent can contribute their domain expertise to address multifaceted questions.
For instance, when presented with the debate topic "Letâs debate about the stock of AAPL. Is it a good investment?", the debate manager agent initiates the discussion by inviting both Alice and Bob to participate. Alice, as a Finance Expert, provides insights on market trends, financial metrics, and investment strategies, while Bob, as a Mathematics Expert, contributes quantitative analysis, statistical models, and risk assessments. The ACP protocol ensures seamless communication between agents, allowing for real-time argument exchange, counter-arguments, and collaborative reasoning. This multi-agent debate system exemplifies how the TEA protocol enables dynamic agent coordination, specialized expertise integration, and structured knowledge synthesis across diverse domains, demonstrating the frameworkâs capability to support complex multi-agent interactions and collaborative problem-solving scenarios.
<details>
<summary>assets/debate/0001.jpg Details</summary>

### Visual Description
\n
## Screenshot: Multi-Agent Debate Platform Interface
### Overview
The image is a screenshot of a user interface for a "Multi-Agent Debate Platform". The platform is designed for real-time AI-powered debate visualization. The interface allows users to define a debate topic and participants (AI agents with specified expertise) and then initiate a debate. The screenshot shows the initial state of the platform with a debate topic already entered and participants defined.
### Components/Axes
The interface is divided into several sections:
* **Header:** Contains the platform title "Multi-Agent Debate Platform" and a subtitle "Real-time AI-powered debate visualization".
* **Debate Topic Input:** A text input field labeled "Debate Topic:" with the pre-filled text "Let's debate about the stock of AAPL. Is it a good investment?".
* **Participants Section:** Labeled "Participants:", listing two participants:
* Alice (Finance Expert)
* Bob (Mathematics Expert)
* **Control Buttons:** Three buttons:
* "Start Debate" (Blue)
* "Clear" (Grey)
* "Connected" (Green) with a small green circle indicator.
* **Live Debate Section:** Labeled "Live Debate", currently displaying a message from "System": "Welcome to the Multi-Agent Debate Platform! Start a debate to see AI agents discuss topics in real-time."
* **Status Indicator:** "Connected" in the top-right corner.
### Detailed Analysis or Content Details
The text content is as follows:
* **Platform Title:** Multi-Agent Debate Platform
* **Subtitle:** Real-time AI-powered debate visualization
* **Debate Topic Label:** Debate Topic:
* **Debate Topic:** Let's debate about the stock of AAPL. Is it a good investment?
* **Participants Label:** Participants:
* **Participant 1:** Alice (Finance Expert)
* **Participant 2:** Bob (Mathematics Expert)
* **Button 1:** Start Debate
* **Button 2:** Clear
* **Button 3:** Connected
* **Live Debate Label:** Live Debate
* **System Message:** Welcome to the Multi-Agent Debate Platform! Start a debate to see AI agents discuss topics in real-time.
* **Status:** Connected
### Key Observations
The interface is clean and straightforward. The pre-filled debate topic focuses on the stock of Apple (AAPL), suggesting a financial application. The platform highlights the expertise of the AI agents involved (Finance and Mathematics). The "Connected" status indicates the platform is online and ready for use.
### Interpretation
The screenshot demonstrates a platform designed to simulate debates between AI agents. The platform's purpose is to provide a visualization of these debates, likely showing the arguments and reasoning of each agent. The focus on AAPL stock suggests a use case in financial analysis and investment decision-making. The inclusion of both Finance and Mathematics expertise indicates the platform aims for a comprehensive and rigorous analysis of the topic. The "Live Debate" section suggests the debate unfolds in real-time, providing a dynamic and interactive experience. The platform appears to be in a ready state, awaiting the initiation of a debate.
</details>
<details>
<summary>assets/debate/0002.jpg Details</summary>

### Visual Description
\n
## Screenshot: Live Debate Transcript
### Overview
The image is a screenshot of a "Live Debate" interface, displaying a transcript of a discussion regarding Apple (AAPL) as an investment in 2025. The transcript is attributed to a user named "bob" and includes a timestamp. The interface has buttons for "Start Debate", "Clear", and a "Connected" indicator.
### Components/Axes
The interface consists of:
* **Header:** Contains "Start Debate", "Clear", and "Connected" indicators.
* **Main Body:** Displays the transcript of the debate.
* **Attribution:** Shows the user "bob" and timestamp "17:27:52".
### Content Details
The transcript reads as follows:
"To evaluate whether AAPL is a good investment at this point in 2025, it's important to examine several factors: Apple's recent financial performance, its innovation pipeline, competitive landscape, and macroeconomic conditions.
First, let's consider Apple's financials. As of September 2025, Apple has consistently reported strong quarterly earnings with stable revenue growth, high margins, and significant cash reserves. The company's service and wearables segments have outpaced iPhone sales growth in recent quarters, showing diversification that reduces dependency on a single product category. This kind of revenue mix can help stabilize performance even if hardware upgrades slow down.
Secondly, Apple's innovation remains a key differentiator. The company's entrance into AR/VR devices, ongoing investments in AI integration across its ecosystem, and rumored developments in healthcare and automotive technology reinforce its capacity to create new growth drivers. Historically, Apple has succeeded in creating and monetizing new product categories, supporting long-term shareholder value.
However, there are notable risks. The global smartphone market is saturated, and competition, especially from Chinese OEMs and Samsung, remains intense. Regulatory scrutiny, particularly regarding App Store practices in the U.S. and EU, could pressure margins and business models. Additionally, as Apple continues its premium pricing strategy, economic headwinds such as inflation or consumer spending shifts could affect upgrade cycles.
Valuation also matters: Apple currently trades at a premium relative to historical averages, with a P/E ratio that reflects strong growth expectations. Investors should consider whether future growth and innovation can justify this valuation, or if there is risk of multiple contraction, especially if the"
### Key Observations
The text provides a balanced assessment of Apple as an investment in 2025, outlining both positive factors (strong financials, innovation) and risks (market saturation, competition, regulation, valuation). The analysis is dated as of September 2025.
### Interpretation
The text represents a preliminary investment analysis of Apple. It suggests a cautious optimism, acknowledging Apple's strengths but also highlighting potential challenges. The author emphasizes the importance of considering multiple factors â financial performance, innovation, competitive landscape, and macroeconomic conditions â before making an investment decision. The mention of a premium P/E ratio suggests the stock may be overvalued, requiring strong future performance to justify its current price. The incomplete final sentence indicates the analysis was likely ongoing or interrupted. The text is a reasoned argument, not simply a statement of opinion, and demonstrates a good understanding of the factors influencing Apple's investment potential.
</details>
<details>
<summary>assets/debate/0003.jpg Details</summary>

### Visual Description
\n
## Screenshot: Live Debate Interface
### Overview
This is a screenshot of a live debate interface, likely from an AI-powered platform. It displays a conversation between a "System" and a user named "Bob". The interface includes controls for starting a debate and clearing the conversation, as well as a connected status indicator. The main content area shows the ongoing dialogue, with timestamps and a visual indicator suggesting Bob is currently "thinking".
### Components/Axes
The interface consists of the following elements:
* **Top Bar:** Contains "Start Debate" button (blue), "Clear" button (grey), and "Connected" status indicator (green).
* **Header:** Displays "Live Debate" and "Connected" status.
* **Conversation Area:** Displays the dialogue between "System" and "Bob", with timestamps and user identifiers.
* **Thinking Indicator:** A small animated icon indicating Bob is processing.
### Detailed Analysis or Content Details
The conversation content is as follows:
**System (17:27:59):** "bob is thinking..."
**Bob (17:28:06):**
"Certainly, I'll elaborate on the earlier points discussed.
If the focus of the debate is, for example, regarding the effectiveness of government intervention in addressing climate change, itâs essential to clarify both the mechanisms and the evidence supporting them. Government intervention can take several forms: regulatory policies (such as carbon emissions caps), economic incentives (such as subsidies for renewable energy), and public investment in research and infrastructure.
Empirical evidence supports the argument that targeted government intervention has driven significant environmental improvements. For instance, the 1990 U.S. Clean Air Act Amendments introduced market-based solutions for acid rain, leading to a more than 80% reduction in SOâ emissions by 2010, all while keeping costs lower than projected. Additionally, countries like Germany have demonstrated that sustained feed-in tariffs and aggressive policy commitments can accelerate renewable energy adoption, as seen in the rapid expansion of solar and wind power since the early 2000s.
Critics often point to inefficiencies, unintended market distortions, or the risk of politicization. These concerns are valid but not insurmountableâwell-"
### Key Observations
* The conversation is focused on the effectiveness of government intervention in addressing climate change.
* Bob is providing a detailed response, citing examples like the 1990 U.S. Clean Air Act Amendments and Germany's feed-in tariffs.
* The conversation is ongoing, as Bob's response is incomplete ("well-").
* The interface is designed for real-time interaction, as indicated by the timestamps and "thinking" indicator.
### Interpretation
The screenshot depicts an AI-driven debate platform where users can engage in discussions on complex topics. The system appears to be capable of generating thoughtful and well-supported arguments, as demonstrated by Bob's response. The interface is clean and intuitive, facilitating a smooth conversational experience. The focus on climate change suggests the platform may be used for educational or policy-related discussions. The incomplete sentence at the end of Bob's response indicates the conversation is still in progress, and further insights are likely to be revealed. The use of specific examples (SOâ reduction, Germany's tariffs) lends credibility to the arguments being presented.
</details>
<details>
<summary>assets/debate/0004.jpg Details</summary>

### Visual Description
\n
## Screenshot: Live Debate Interface
### Overview
This is a screenshot of a live debate interface, likely a web application or software platform. The interface displays a conversation between a "System" and a user named "Bob," along with controls to start, clear, and indicate connection status. The main content area presents questions posed to the debate participants and responses from both the system and Bob.
### Components/Axes
The interface contains the following elements:
* **Top Bar:** Contains buttons labeled "Start Debate" (blue), "Clear" (grey), and a "Connected" indicator (green).
* **Title:** "Live Debate" with a "Connected" status indicator in the top-right corner.
* **Questions:** A series of questions posed to the participants, including:
* "Given these cost and technical trends, what specific evidence do those skeptical of renewablesâ economic viability cite that outweighs these findings?"
* "How do opposing positions account for the long-term externalitiesâlike health costs and climate impactsâassociated with continued reliance on fossil fuels?"
* **Conversation Area:** Displays messages from "System" and "Bob," including timestamps.
* **System Messages:** Messages from the "System" are labeled with "S System" and a timestamp (e.g., 17:28:15).
* **Bob Messages:** Messages from "Bob" are labeled with "B Bob" and a timestamp (e.g., 17:28:27).
### Detailed Analysis or Content Details
Here's a transcription of the text content:
**Top Bar Buttons:**
* Start Debate
* Clear
* Connected
**Title:**
* Live Debate Connected
**Questions:**
* "Given these cost and technical trends, what specific evidence do those skeptical of renewablesâ economic viability cite that outweighs these findings?"
* "How do opposing positions account for the long-term externalitiesâlike health costs and climate impactsâassociated with continued reliance on fossil fuels?"
* "I welcome counter-evidence or alternative perspectives that challenge these points, as it will help ensure our debate is well-rounded and substantiated."
**System Message (17:28:15):**
* "bob is thinkingâŚ"
**Bob Message (17:28:27):**
* "Thank you for raising the question of evidenceâthis is fundamental for any robust position."
* "To support my stance, letâs consider empirical studies and established data. For instance, if the topic is about the effectiveness of remote work (as previously discussed), we can reference the 2023 meta-analysis conducted by Stanford University, which aggregated over 30 individual studies and found that remoteâŚ" (The message is cut off at the end).
### Key Observations
* The interface is designed for a structured debate format.
* The system appears to be an AI or automated moderator, prompting questions and acknowledging user responses.
* Bob is actively participating in the debate, referencing empirical evidence to support his arguments.
* The conversation is ongoing, as indicated by the "bob is thinkingâŚ" message and the incomplete sentence in Bob's last message.
* The timestamp format is HH:MM:SS.
### Interpretation
The screenshot depicts a dynamic interaction within a debate platform. The system's role is to facilitate a well-reasoned discussion by posing challenging questions and encouraging evidence-based responses. Bob's response demonstrates a commitment to supporting his arguments with data, specifically citing a meta-analysis from Stanford University. The interface's design suggests a focus on clarity and organization, with distinct labels and timestamps to track the flow of the conversation. The incomplete message from Bob indicates that the debate is still in progress. The emphasis on "evidence" and "substantiated" arguments suggests a high standard for the quality of discourse within this platform. The questions posed are complex and require nuanced answers, indicating that this is not a superficial debate but rather a serious attempt to explore different perspectives on important issues.
</details>
<details>
<summary>assets/debate/0005.jpg Details</summary>

### Visual Description
\n
## Screenshot: Live Debate Interface
### Overview
The image is a screenshot of a live debate interface, likely from an online platform. It displays a chat window with a user ("bob") typing a message, along with interface elements for starting a debate and indicating connection status. The primary content is a block of text representing a user's statement.
### Components/Axes
The interface elements include:
* **Top Bar:** Contains buttons labeled "Start Debate" and "Clear", and a "Connected" indicator.
* **Header:** Displays "Live Debate" and "Connected" status.
* **Chat Window:** Contains a message from user "bob" with a timestamp.
* **User Indicator:** A "B" icon next to the user's name "bob".
### Content Details
The text from user "bob" is transcribed below:
"To support my position, it is essential to reference both empirical data and relevant case studies. For instance, if we are considering the effectiveness of universal basic income (UBI) as a social policy, multiple pilot programs and studies provide substantive evidence.
One notable example is the Finnish basic income experiment (2017-2018), which provided 2,000 unemployed Finns with a regular, unconditional income. The final report concluded that recipients experienced increased well-being and slightly higher employment rates compared to the control group (Kangas et al., 2020). Additionally, the experiment reported improvements in mental health and life satisfaction among participants, indicating potential broader societal benefits.
Another significant data point comes from the Stockton Economic Empowerment Demonstration (SEED) in California. Recipients who received $500 monthly payments over two years were found to secure full-time employment at a higher rate than those in the control group (West et al., 2021). This suggests that contrary to concerns about work disincentives, direct cash transfers can actually enhance economic agency and motivation.
Critics often argue that such policies are unsustainable or lead to reduced labor force participation. However, the aforementioned studies, as well as similar findings from Kenyaâs GiveDirectly projects, challenge this assumption. The evidence so far points towards potential positive effects on bothâŚ"
The timestamp associated with the message is "17:28:40".
The text references the following studies:
* Kangas et al., 2020
* West et al., 2021
### Key Observations
The text focuses on the effectiveness of Universal Basic Income (UBI), citing evidence from the Finnish basic income experiment and the Stockton Economic Empowerment Demonstration. The message argues against common criticisms of UBI, suggesting that it can enhance economic agency and motivation. The text is incomplete, ending mid-sentence.
### Interpretation
The screenshot captures a moment in a debate concerning the merits of UBI. The user "bob" is presenting a case supported by empirical data and case studies. The references to specific studies (Kangas et al., 2020; West et al., 2021) lend credibility to the argument. The incomplete sentence suggests the debate is ongoing. The interface elements indicate a real-time, connected environment for discussion. The overall impression is of a reasoned argument being presented in a structured online debate format. The screenshot provides factual information about the debate topic and the evidence being used to support a particular viewpoint.
</details>
<details>
<summary>assets/debate/0006.jpg Details</summary>

### Visual Description
\n
## Screenshot: Live Debate Interface
### Overview
The image is a screenshot of a live debate interface. It displays a text-based conversation, likely from an AI or chatbot system, discussing the evaluation of Universal Basic Income (UBI). The interface includes controls for starting, clearing, and connection status. A chat log from a "System" user indicates the debate has concluded.
### Components/Axes
The interface consists of the following elements:
* **Top Bar:** Contains buttons labeled "Start Debate", "Clear", and a "Connected" indicator (green dot).
* **Title:** "Live Debate" with a "Connected" status indicator.
* **Chat Log:** A large text area displaying the debate transcript.
* **System Message:** A message bubble from "System" at the bottom, indicating "Debate completed!".
* **Timestamp:** "17:28:59" displayed next to the System message.
### Content Details
The text within the chat log is as follows:
"First, in evaluating effectiveness, empirical data is essential. For instance, suppose the discussion is about universal basic income (UBI). Numerous pilot programs, such as the one in Finland (2017-2018), demonstrated modest improvements in well-being and, in some cases, workforce participation, but the broader economic impact remained inconclusive (Kangas et al., 2020). This suggests the value of pilot data, but also highlights the need for caution before making broad generalizations.
Second, scalability is critical. What works in a localized contextâdue to specific demographics, economic conditions, or support infrastructuresâdoes not always translate smoothly to a national or global scale. The transition often encounters unforeseen administrative costs, political resistance, or unintended socioeconomic effects.
I would challenge others to consider: What additional forms of evidence or case studies can we draw on? Are there specific contextual variables that might alter the outcome if this policy were scaled? Finally, which mechanisms should be prioritized to ensure both equity and efficiency if the proposal were implemented widely?
I look forward to hearing counterpoints or supporting data from others."
The System message reads: "Debate completed!"
### Key Observations
The text focuses on the complexities of evaluating UBI, emphasizing the importance of empirical data, scalability challenges, and the need for contextual consideration. The reference to "Kangas et al., 2020" suggests a specific research paper or study. The interface is designed for interactive discussion, as indicated by the "Start Debate" and "Clear" buttons.
### Interpretation
The screenshot represents a segment of an AI-driven debate or discussion about UBI. The text demonstrates a nuanced perspective, acknowledging both the potential benefits (modest improvements in well-being) and the significant challenges (scalability, unforeseen consequences) associated with implementing UBI. The questions posed at the end of the text suggest the system is designed to solicit further input and explore different viewpoints. The "System" message confirms the conclusion of this particular exchange. The interface is likely part of a larger platform for exploring complex policy issues through simulated debate. The timestamp indicates the debate occurred at 17:28:59. The overall tone is analytical and invites further investigation.
</details>
Figure 16: Case study of TEA agent for multi-agent debate.
### G.3 GitHub usage
This case study demonstrates the agentâs comprehensive GitHub workflow automation capabilities through the creation and deployment of a simple HTML Sokoban web mini-game. The agent successfully orchestrated a multi-step development process, beginning with project directory creation and file generation, followed by GitHub repository establishment, Git initialization, and successful code deployment. The execution showcases the agentâs proficiency in coordinating file system operations, version control management, and remote repository interactions to deliver a complete, functional web application.
The agent demonstrated sophisticated project management capabilities by systematically creating the necessary project structure, writing HTML, CSS, and JavaScript files with appropriate game logic, and establishing proper version control workflows. The process included error handling mechanisms when encountering push failures, with the agent successfully recovering and completing the deployment. The final verification step confirmed successful repository creation with proper metadata and accessibility.
Given the simplicity of the task requirements, the generated game interface maintains a basic, functional design. With more detailed specifications and design guidance, the agent could undoubtedly generate more sophisticated and aesthetically pleasing frontend projects, demonstrating the frameworkâs potential for complex web development workflows.
<details>
<summary>x18.png Details</summary>

### Visual Description
## Flowchart: Sokoban-mini Game Deployment Process
### Overview
This flowchart illustrates the process of creating a Sokoban-mini game, writing the game files, creating a GitHub repository, initializing Git, verifying the repository, and handling potential push errors. The process begins with creating a directory and ends with a "Complete" state. The flowchart uses rectangular boxes to represent actions, diamond shapes for decision points, and rounded rectangles for start and end states.
### Components/Axes
The flowchart consists of the following components:
* **Start:** Initial state of the process.
* **Create Game Directory:** Creates a directory named "sokoban-mini".
* **Write Game Files:** Writes `index.html`, `style.css`, and `main.js` files.
* **Create GitHub Repository:** Creates a repository named "sokoban-mini" on GitHub.
* **Initialize Git & Commit:** Initializes a Git repository and commits the changes.
* **Push Success?:** A decision point to check if the push to the remote repository was successful.
* **Handle Push Error:** Actions to take if the push fails.
* **Verify Repository:** Verifies the repository details.
* **Complete:** Final state of the process.
* **Game Screenshot:** A visual representation of the Sokoban-mini game.
* **GitHub Repository:** A screenshot of the GitHub repository page.
### Detailed Analysis or Content Details
The flowchart details the following steps and outputs:
1. **Create Game Directory:**
- Command: `create_dir("sokoban-mini")`
- Output: "Successfully created directory sokoban-mini"
2. **Write Game Files:**
- Command: `write("index.html", html_content)`
- Output: "Successfully wrote 690 bytes"
- Command: `write("style.css", css_content)`
- Output: "Successfully wrote 1694 bytes"
- Command: `write("main.js", js_content)`
- Output: "Successfully wrote 3287 bytes"
3. **Create GitHub Repository:**
- Command: `create_repository(name="sokoban-mini", description="Simple HTML Sokoban web mini-game", pass="", private=False)`
- Output: "Successfully created repository sokoban-mini"
4. **Initialize Git & Commit:**
- Command: `git_init(local_path="sokoban-mini", remote_url="https://github.com/")`
- Command: `git_commit(message="Initial commit - Simple HTML Sokoban mini-game")`
- Output: "Git repository initialized"
- Output: "Commit created: abcd704"
5. **Push Success?:**
- If "Yes", the process proceeds to "Complete".
- If "No", the process proceeds to "Handle Push Error".
6. **Handle Push Error:**
- Command: `git_push(local_path="sokoban-mini", remote="origin", branch="main")`
- Output: "Successfully pushed branch 'main' to remote 'origin'"
7. **Verify Repository:**
- Command: `get_repository(repo="sokoban-mini")`
- Output:
- Repository: `<USERNAME>/sokoban-mini`
- URL: `https://github.com/<USERNAME>/sokoban-mini`
- Language: JavaScript
- Private: False
The "Game Screenshot" shows a grid-based game with a player character (represented by a red square) and boxes (represented by green squares) on a green background. The "GitHub Repository" screenshot shows the repository's file structure and commit history.
### Key Observations
The process is linear with a single decision point for push success. The outputs of each step are clearly defined, providing a detailed record of the deployment process. The use of Git for version control is evident. The game is written in JavaScript, HTML, and CSS.
### Interpretation
This flowchart demonstrates a streamlined process for deploying a simple web-based game (Sokoban-mini) to a GitHub repository. It highlights the key steps involved, from initial directory creation to version control and remote repository management. The inclusion of file sizes and commit hashes suggests a focus on tracking and reproducibility. The "Push Success?" decision point indicates an awareness of potential network or permission issues during the deployment process, and the "Handle Push Error" step provides a recovery mechanism. The screenshots of the game and repository provide visual context and confirm the successful completion of the process. The process is well-documented and appears to be designed for automation or easy replication.
</details>
Figure 17: Case study of TEA agent for GitHub usage.
### G.4 Browser operation
<details>
<summary>assets/browser/browser-0000.jpg Details</summary>

### Visual Description
\n
## Text Display: Instruction
### Overview
The image displays a simple text instruction on a black background. There are no charts, diagrams, or complex visual elements. The image consists solely of text.
### Components/Axes
There are no axes, legends, or components beyond the text itself. The text is centered horizontally and vertically on the image.
### Detailed Analysis or Content Details
The text reads: "Go to google.com and search for 'python programming' get the first result."
### Key Observations
The instruction is straightforward and directs the user to perform a specific action: a Google search for the term "python programming" and to retrieve the first search result.
### Interpretation
The image presents a task or challenge. It's likely part of a larger process or test where the user is expected to follow the instruction and potentially analyze or report on the content of the first search result for "python programming". The simplicity of the image suggests the focus is on the *action* and the *result* of that action, rather than any visual analysis. It is a procedural instruction, not a data presentation.
</details>
<details>
<summary>assets/browser/browser-0001.jpg Details</summary>

### Visual Description
\n
## Screenshot: Google Search Page
### Overview
This is a screenshot of the Google search page. The page features the Google logo, a search bar, and links to other Google services. There is also a text instruction at the bottom of the screen.
### Components/Axes
* **Top Navigation:** Contains links to "About", "Store", "Gmail", "Images", and "Sign in".
* **Search Bar:** A prominent text input field with a magnifying glass icon on the left.
* **Buttons:** "Google Search" and "I'm Feeling Lucky". A microphone icon and an AI mode button are also present.
* **Footer:** Contains links to "Advertising", "Business", "How Search works", "Privacy", "Terms", and "Settings".
* **Location:** "Singapore" is displayed in the bottom-left corner.
* **Language Offerings:** "Google offered in: çŽä˝ä¸ć Bahasa Melayu தமிழŕŻ" (Simplified Chinese, Malay, Tamil) is displayed below the search buttons.
* **Instruction Box:** A dark-colored box at the bottom of the screen with the text "Input 'python programming' into the search box and click the 'Google Search' button to perform the search."
### Detailed Analysis or Content Details
The screenshot shows the default Google search page layout. The search bar is centered on the page. The "Google Search" button is blue, while "I'm Feeling Lucky" is grey. The microphone icon is located to the right of the search bar. The AI mode button is also to the right of the search bar. The footer links are small and located at the very bottom of the screen. The location is set to Singapore. The language offerings indicate Google's support for multiple languages.
The instruction box provides a specific task:
* **Input:** "python programming"
* **Action:** Click the "Google Search" button.
### Key Observations
The screenshot is a clear demonstration of the Google search interface. The instruction box suggests this is part of a tutorial or guide. The presence of the AI mode button indicates Google's integration of artificial intelligence into its search functionality. The language offerings demonstrate Google's global reach.
### Interpretation
The screenshot is likely part of a tutorial or instructional material demonstrating how to use the Google search engine. The specific instruction to search for "python programming" suggests the tutorial may be related to programming or software development. The inclusion of the AI mode button highlights Google's efforts to enhance search results using artificial intelligence. The screenshot provides a snapshot of the Google search experience as it appears in Singapore, indicating localization efforts. The languages offered show Google's commitment to accessibility for a diverse user base.
</details>
<details>
<summary>assets/browser/browser-0002.jpg Details</summary>

### Visual Description
\n
## Screenshot: Google Search Results Page
### Overview
The image is a screenshot of a Google search results page. The search query is "python programming". The page displays a list of suggested search terms related to the query, along with the Google logo and standard Google page elements.
### Components/Axes
* **Top Bar:** Contains links to "Gmail", "Images", "Sign in", "About", "Store".
* **Search Bar:** Displays the query "python programming".
* **Suggested Searches:** A list of related search suggestions.
* **Footer:** Contains links to "Advertising", "Business", "How Search works", "Privacy", "Terms", "Settings", and "I'm Feeling Lucky".
* **Location:** "Singapore" is displayed in the bottom-left corner.
* **AI Mode:** A small icon indicating "AI Mode" is present near the search bar.
### Content Details
The suggested search terms are listed as follows, with a numbered icon preceding each:
1. python programming
2. python programming
3. python programming for beginners
4. python programming language
5. python programming course
6. python programming course singapore
7. python programming examples
8. python programming pdf
9. python programming for beginners pdf
10. python programming book
A text instruction at the bottom of the screen reads: "Click the 'Google Search' button to perform the search for 'python programming'."
### Key Observations
The search suggestions are highly relevant to the initial query, indicating Google's ability to understand user intent and provide helpful suggestions. The repetition of "python programming" as the first two suggestions is notable. The suggestions cover a range of user needs, from beginner-level learning to specific resources like courses, examples, and PDFs.
### Interpretation
This screenshot demonstrates the functionality of Google's search suggestion feature. The suggestions are designed to help users refine their search query and quickly find the information they are looking for. The presence of location-specific suggestions ("python programming course singapore") indicates that Google personalizes search results based on the user's location. The instruction to click the "Google Search" button is a standard element of the Google search interface, guiding users on how to proceed with their search. The "AI Mode" icon suggests that the search results may be influenced by artificial intelligence algorithms. The overall design is clean and user-friendly, prioritizing ease of use and efficient information retrieval. The repetition of the search term suggests a high search volume or a strong emphasis on that specific phrase by Google's algorithms.
</details>
<details>
<summary>assets/browser/browser-0003.jpg Details</summary>

### Visual Description
\n
## Screenshot: Google Search Results for "python programming"
### Overview
This is a screenshot of a Google search results page for the query "python programming". The page displays a mix of organic and sponsored results, with a prominent "Generating..." section indicating Google is actively processing the search. The screenshot is intended to demonstrate how to locate the first non-sponsored (organic) search result.
### Components/Axes
* **Search Bar:** Located at the top-left (1), containing the query "python programming".
* **Navigation Tabs:** Below the search bar (2), including "All", "Images", "Videos", "Short videos", "Shopping", "News", "More".
* **"Generating..." Section:** A large, rectangular area (3) with the text "Generating...". This suggests Google is dynamically creating content for the search results.
* **Sponsored Results:** Several rectangular blocks (4, 5) labeled "Sponsored".
* **Organic Results:** A section containing the first organic search result (6).
* **Google Account/Sign-in Button:** Top-right corner (7).
* **Google Apps Icon:** Top-right corner (8).
* **Search Settings Icon:** Top-right corner (9).
* **Search Input Icon:** Top-right corner (10).
### Detailed Analysis or Content Details
1. **Search Query:** "python programming" is entered in the search bar.
2. **Navigation Tabs:** The "All" tab is currently selected.
3. **"Generating..." Section:** This section occupies a significant portion of the screen and is visually represented by a series of horizontal, light-purple bars.
4. **Sponsored Result 1 (4):**
* **URL:** intuitioninternational.com
* **Text:** "https://www.intuitioninternational.com/programming/python_coding"
* **Description:** "concepts and best coding practices using hands-on practices."
5. **Sponsored Result 2 (5):** Displays code snippet:
* `elif operator == "+":`
* `return number1 + number2`
* `elif operator == "-":`
* `return number1 - number2`
* `else:`
* `print("Invalid operator")`
6. **Organic Result 1 (6):**
* **Title:** Python
* **Subtitle:** High-level programming language
7. **Instructional Text:** At the bottom of the image, the text reads: "Scroll down to locate and identify the first organic (non-sponsored) search result for âpython programmingâ."
### Key Observations
* The "Generating..." section suggests Google is actively refining the search results in real-time.
* The screenshot highlights the distinction between sponsored and organic search results.
* The presence of code snippets in the sponsored results indicates a focus on programming-related content.
* The instructional text explicitly guides the user to identify the first organic result.
### Interpretation
The screenshot demonstrates a typical Google search results page for a technical query ("python programming"). The "Generating..." section is a common feature, indicating Google's dynamic search algorithms at work. The inclusion of sponsored results is standard practice, but the instructional text emphasizes the importance of identifying the first *organic* result, which is often considered more relevant and unbiased. The presence of code snippets in the sponsored results suggests targeted advertising to programmers. The overall layout is designed to present a mix of information and advertising, with a clear visual distinction between the two. The screenshot is not presenting data in a quantitative sense, but rather illustrating a user interface and the process of navigating search results.
</details>
<details>
<summary>assets/browser/browser-0004.jpg Details</summary>

### Visual Description
\n
## Screenshot: Google Search Results for "python programming"
### Overview
This is a screenshot of a Google search results page for the query "python programming". The page displays a mix of sponsored results (labeled "Sponsored") and organic (non-sponsored) search results. The right side of the screen contains a knowledge panel providing information about Python as a programming language. The task is to extract the title and URL of the first organic (non-sponsored) search result.
### Components/Axes
The screenshot is divided into several key areas:
* **Search Bar:** At the top, displaying the query "python programming".
* **Sponsored Results:** Several blocks labeled "Sponsored" with associated URLs and descriptions.
* **Organic Search Results:** A list of non-sponsored search results, numbered 1-5 (and likely more beyond the visible portion).
* **Knowledge Panel:** A panel on the right side providing information about Python.
* **Navigation Bar:** Top-right corner with Google apps and sign-in options.
### Detailed Analysis or Content Details
**Sponsored Results:**
1. **Title:** Python Coding Essentials - Beginner to Intermediate
**URL:** https://www.intuitioninternational.com/programming/python\_coding
**Description:** Python is used in machine learning, big data analysis and visualisation. The course introduces coding concepts and best coding practices using hands-on practices.
2. **Title:** Python From Zero To Hero
**URL:** https://www.codefinity.com
**Description:** Best free Python course â Dive into theory and complete numerous practice exercises to master your coding skills. Open up new career options in tech! "Intro to Python" Self-Paced⌠Python Programming: Best for Beginners: Python Course - View Pricing ¡ SQL Course
3. **Title:** Welcome to Python.org
**URL:** https://www.python.org
**Description:** The official home of the Python Programming Language.
**Organic Search Results:**
1. **Title:** Welcome to Python.org
**URL:** https://www.python.org
**Description:** The official home of the Python Programming Language.
2. **Title:** Python For Beginners
**Description:** An experienced programmer in any programming language âŚ
3. **Title:** The Python Tutorial
**Description:** Python is an easy to learn, powerful programming âŚ
**Knowledge Panel (Right Side):**
* **Title:** Python
**Subtitle:** High-level programming language
* **Code Snippet:** A block of Python code demonstrating operator overloading. The code defines a class with methods for adding, subtracting, multiplying, and dividing numbers.
* **Description:** Python is a high-level, general-purpose programming language. Its design philosophy emphasizes code readability with the use of significant indentation. Python is dynamically type-checked and garbage-collected.
* **Source:** Wikipedia
* **First appeared:** 20 February 1991; 34 years ago
* **Designed by:** Guido van Rossum
* **Typing discipline:** duck, dynamic, strong, optional type annotations
### Key Observations
The first organic search result is "Welcome to Python.org" with the URL "https://www.python.org". The knowledge panel provides a concise overview of Python's key features and history. The screenshot demonstrates a typical Google search results page layout, with a mix of paid and organic listings.
### Interpretation
The screenshot illustrates how Google presents information in response to a programming-related query. The prominence of both sponsored courses and the official Python website suggests a high level of interest in learning Python. The knowledge panel serves as a quick reference for users seeking basic information about the language. The task was to extract the first organic result, which is the official Python website, indicating Google's algorithm prioritizes the official source for this query. The presence of multiple sponsored results highlights the commercial aspect of online learning resources. The code snippet in the knowledge panel is a practical example of Python's capabilities, potentially attracting users interested in learning the language.
</details>
Figure 18: Case study of TEA agent for browser operation.
This case study demonstrates the agentâs sophisticated browser automation capabilities through a comprehensive web interaction scenario involving the search for "python programming" content. The agent exhibits advanced multi-modal reasoning by simultaneously processing both DOM (Document Object Model) structures and visual elements to understand webpage layout and functionality. Through systematic analysis of page elements, the agent can identify interactive components, assess their relevance to the search objective, and make informed decisions about subsequent navigation actions. The execution demonstrates the agentâs capacity for autonomous web exploration, where it can parse complex webpage structures, interpret visual cues, and execute precise interactions to achieve its objectives. This capability extends beyond simple element clicking to encompass sophisticated understanding of webpage semantics and user interface patterns, with remarkable proficiency in handling dynamic content, managing asynchronous operations, and adapting to varying webpage architectures across different domains and platforms.
The browser automation framework incorporates several advanced technical components that enable robust web interaction. The agent leverages hierarchical DOM parsing algorithms to construct semantic representations of webpage structure, enabling precise element localization and interaction planning. Visual processing capabilities allow for the interpretation of complex layouts, including responsive design elements, dynamic content loading, and multi-modal interface components. The system demonstrates particular strength in handling modern web applications that rely heavily on JavaScript-driven interactions and asynchronous content loading. Furthermore, the agent exhibits sophisticated error recovery mechanisms when encountering unexpected webpage behaviors, such as dynamic content changes, popup interventions, or navigation redirects. This resilience is achieved through continuous monitoring of page state changes and adaptive strategy modification based on real-time feedback from the browser environment.
Our browser environment supports not only conventional multi-modal models combined with DOM manipulation (limited to clicking and controlling page elements without pixel-level operations), but also integrates computer-use-preview functionality that enables operator-like pixel-level precision operations, significantly expanding the scope of environmental exploration capabilities. This dual-mode architecture provides unprecedented flexibility in web automation, allowing for both high-level semantic interactions and low-level pixel-accurate operations when necessary.
## Appendix H Prompts
Our foundational agent framework is built upon a ReAct-based tool-calling agent architecture, which follows a systematic "thinking-then-action" paradigm. During execution, the agent records its decision-making process and execution trajectory, continuously summarizing experiences and extracting insights through its memory mechanism. The agent employs a done tool to determine task completion, ensuring reliable termination of complex workflows. Notably, the planning agent is built upon this comprehensive tool-calling foundation to coordinate multifaceted resources, while specialized agents such as the deep researcher, deep analyzer, browser operator, and tool manager utilize optimized custom workflows to achieve an optimal balance between high task completion rates and reduced resource consumption. We do not provide the detailed prompts for other specialized agents and the self-evolution module here; for further details, please refer to the source code in the supplementary materials.
The agentâs prompt structure consists of two primary components: a static system prompt that establishes the agentâs role, capabilities, and behavioral guidelines, and a dynamic agent message prompt that provides the task instructions, environmental state, and execution history. These components work together to guide the agentâs reasoning process and action selection. The template of the tool-calling prompt is shown as follows:
Tool Calling Prompt Template:
⏠from src. registry import PROMPT from src. prompt. types import Prompt from typing import Any, Dict, Literal from pydantic import Field, ConfigDict AGENT_PROFILE = """ You are an AI agent that operates in iterative steps and uses registered tools to accomplish the user â s task. Your goals are to solve the task accurately, safely, and efficiently. """ AGENT_INTRODUCTION = """ < intro > You excel at: - Analyzing tasks and breaking them down into actionable steps - Selecting and using appropriate tools to accomplish goals - Reasoning systematically and tracking progress - Adapting your approach when encountering obstacles - Completing tasks accurately and efficiently </ intro > """ LANGUAGE_SETTINGS = """ < language_settings > - Default working language: ** English ** - Always respond in the same language as the user request </ language_settings > """ # Input = agent context + environment context + tool context INPUT = """ < input > - < agent_context >: Describes your current internal state and identity, including your current task, relevant history, memory, and ongoing plans toward achieving your goals. This context represents what you currently know and intend to do. - < environment_context >: Describes the external environment, situational state, and any external conditions that may influence your reasoning or behavior. - < tool_context >: Describes the available tools, their purposes, usage conditions, and current operational status. - < examples >: Provides few - shot examples of good or bad reasoning and tool - use patterns. Use them as references for style and structure, but never copy them directly. </ input > """ # Agent context rules = task rules + agent history rules + memory rules + todo rules AGENT_CONTEXT_RULES = """ < agent_context_rules > < workdir_rules > You are working in the following working directory: {{ workdir }}. - When using tools (e. g., â bash â or â python_interpreter â) for file operations, you MUST use absolute paths relative to this workdir (e. g., if workdir is â/ path / to / workdir â, use â/ path / to / workdir / file. txt â instead of â file. txt â). </ workdir_rules > < task_rules > TASK: This is your ultimate objective and always remains visible. - This has the highest priority. Make the user happy. - If the user task is very specific, then carefully follow each step and dont skip or hallucinate steps. - If the task is open ended you can plan yourself how to get it done. You must call the â done â tool in one of three cases: - When you have fully completed the TASK. - When you reach the final allowed step (â max_steps â), even if the task is incomplete. - If it is ABSOLUTELY IMPOSSIBLE to continue. </ task_rules > < agent_history_rules > Agent history will be given as a list of step information with summaries and insights as follows: < step_ [step_number]> Evaluation of Previous Step: Assessment of last tool call Memory: Your memory of this step Next Goal: Your goal for this step Tool Results: Your tool calls and their results </ step_ [step_number]> </ agent_history_rules > < memory_rules > You will be provided with summaries and insights of the agent â s memory. < summaries > [A list of summaries of the agent â s memory.] </ summaries > < insights > [A list of insights of the agent â s memory.] </ insights > </ memory_rules > </ agent_context_rules > """ # Environment context rules = environments rules ENVIRONMENT_CONTEXT_RULES = """ < environment_context_rules > Environments rules will be provided as a list, with each environment rule consisting of three main components: < state >, < vision > (if screenshots of the environment are available), and < interaction >. </ environment_context_rules > """ # Tool context rules = reasoning rules + tool use rules + tool rules TOOL_CONTEXT_RULES = """ < tool_context_rules > < tool_use_rules > You must follow these rules when selecting and executing tools to solve the < task >. ** Usage Rules ** - You MUST only use the tools listed in < available_tools >. Do not hallucinate or invent new tools. - You are allowed to use a maximum of {{ max_tools }} tools per step. - DO NOT include the â output â field in any tool call -- tools are executed after planning, not during reasoning. - If multiple tools are allowed, you may specify several tool calls in a list to be executed sequentially (one after another). ** Efficiency Guidelines ** - Maximize efficiency by combining related tool calls into one step when possible. - Use a single tool call only when the next call depends directly on the previous tool â s specific result. - Think logically about the tool sequence: " What â s the natural, efficient order to achieve the goal?" - Avoid unnecessary micro - calls, redundant executions, or repetitive tool use that doesn â t advance progress. - Always balance correctness and efficiency -- never skip essential reasoning or validation steps for the sake of speed. - Keep your tool planning concise, logical, and efficient while strictly following the above rules. </ tool_use_rules > < todo_rules > You have access to a â todo â tool for task planning. Use it strategically based on task complexity: ** For Complex / Multi - step Tasks (MUST use â todo â tool):** - Tasks requiring multiple distinct steps or phases - Tasks involving file processing, data analysis, or research - Tasks that need systematic planning and progress tracking - Long - running tasks that benefit from structured execution ** For Simple Tasks (may skip â todo â tool):** - Single - step tasks that can be completed directly - Simple queries or calculations - Tasks that don â t require planning or tracking ** When using the â todo â tool:** - The â todo â tool is initialized with a â todo. md â: Use this to keep a checklist for known subtasks. Use â replace â operation to update markers in â todo. md â as first tool call whenever you complete an item. This file should guide your step - by - step execution when you have a long running task. - If â todo. md â is empty and the task is multi - step, generate a stepwise plan in â todo. md â using â todo â tool. - Analyze â todo. md â to guide and track your progress. - If any â todo. md â items are finished, mark them as complete in the file. </ todo_rules > </ tool_context_rules > """ EXAMPLE_RULES = """ < example_rules > You will be provided with few shot examples of good or bad patterns. Use them as reference but never copy them directly. </ example_rules > """ REASONING_RULES = """ < reasoning_rules > You must reason explicitly and systematically at every step in your â thinking â block. Exhibit the following reasoning patterns to successfully achieve the < task >: - Analyze < agent_history > to track progress toward the goal. - Reflect on the most recent " Next Goal " and " Tool Result ". - Evaluate success / failure / uncertainty of the last step. - Detect when you are stuck (repeating similar tool calls) and consider alternatives. - Maintain concise, actionable memory for future reasoning. - Before finishing, verify results and confirm readiness to call â done â. - Always align reasoning with < task > and user intent. </ reasoning_rules > """ OUTPUT = """ < output > You must ALWAYS respond with a valid JSON in this exact format. DO NOT add any other text like "âââ json " or "âââ" or anything else: { " thinking ": " A structured < think >- style reasoning block that applies the < reasoning_rules > provided above.", " evaluation_previous_goal ": " One - sentence analysis of your last tool usage. Clearly state success, failure, or uncertainty.", " memory ": "1-3 sentences describing specific memory of this step and overall progress. Include everything that will help you track progress in future steps.", " next_goal ": " State the next immediate goals and tool calls to achieve them, in one clear sentence.", " tool ": [ {" name ": " tool_name ", " args ": {tool - specific parameters}} // ... more tools in sequence ] } Tool list should NEVER be empty. </ output > """ SYSTEM_PROMPT_TEMPLATE = """ {{ agent_profile }} {{ agent_introduction }} {{ language_settings }} {{ input }} {{ agent_context_rules }} {{ environment_context_rules }} {{ tool_context_rules }} {{ example_rules }} {{ reasoning_rules }} {{ output }} """ # Agent message (dynamic context) - using Jinja2 syntax AGENT_MESSAGE_PROMPT_TEMPLATE = """ {{ agent_context }} {{ environment_context }} {{ tool_context }} {{ examples }} """
The system prompt is structured to support the TEA (Tool-Environment-Agent) protocol through comprehensive context management and rule enforcement across three core components. The prompt explicitly manages Agent Context through role definition (agent_profile), core capabilities (agent_introduction), and behavioral guidelines (language_settings). It further incorporates rigorous task management (task_rules), working directory constraints (workdir_rules), and an iterative execution history framework (agent_history_rules) coupled with memory accumulation (memory_rules) to ensure continuous progress monitoring and context maintenance. Environment Context management is implemented through environment rules (environment_context_rules) that define interaction patterns, state transitions, and multimodal feedback mechanisms, providing structured access to environmental status and observations. Tool Context management is achieved through strict tool-use rules and efficiency guidelines (tool_use_rules), alongside a strategic todo mechanism (todo_rules) for systematic planning of multi-step tasks. The entire process is underpinned by systematic reasoning rules (reasoning_rules) and a rigid JSON output protocol (output), enabling seamless coordination between agent reasoning, environmental awareness, and tool utilization within the TEA distributed architecture.