2510.27051v1
Model: gemini-2.0-flash
# Adaptive Data Flywheel: Applying MAPE Control Loops to AI Agent Improvement
Abstract
Enterprise AI agents must continuously adapt to maintain accuracy, reduce latency, and remain aligned with user needs. We present a practical implementation of a data flywheel in NVInfo AI, NVIDIA’s Mixture-of-Experts (MoE) Knowledge Assistant serving over 30,000 employees. By operationalizing a MAPE-driven data flywheel, we built a closed-loop system that systematically addresses failures in retrieval-augmented generation (RAG) pipelines and enables continuous learning. Over a 3-month post-deployment period, we monitored feedback and collected 495 negative samples. Analysis revealed two major failure modes: routing errors (5.25%) and query rephrasal errors (3.2%). Using NVIDIA NeMo microservices, we implemented targeted improvements through fine-tuning. For routing, we replaced a Llama 3.1 70B model with a fine-tuned 8B variant, achieving 96% accuracy, a 10× reduction in model size, and 70% latency improvement. For query rephrasal, fine-tuning yielded a 3.7% gain in accuracy and a 40% latency reduction. Our approach demonstrates how human-in-the-loop (HITL) feedback, when structured within a data flywheel, transforms enterprise AI agents into self-improving systems. Key learnings include approaches to ensure agent robustness despite limited user feedback, navigating privacy constraints, and executing staged rollouts in production. This work offers a repeatable blueprint for building robust, adaptive enterprise AI agents capable of learning from real-world usage at scale.
I Introduction
Enterprise adoption of generative AI (GenAI) agents has accelerated rapidly, with applications ranging from knowledge retrieval to workflow automation. However, the performance of these systems often deteriorates post-deployment due to evolving user intent, domain drift, and the absence of systematic feedback integration. A central challenge in operationalizing such agents lies in enabling them to continuously adapt based on real-world usage patterns and user feedback, without requiring full-scale retraining or infrastructure overhauls.
While retrieval-augmented generation (RAG) pipelines and Mixture-of-Experts (MoE) architectures have improved the relevance and efficiency of enterprise AI agents, most production deployments remain static and reactive. Feedback mechanisms, if present, are frequently decoupled from the model improvement process. This disconnect results in stagnant accuracy, increasing latency, and declining user trust. There is a pressing need for closed-loop systems that can monitor agent performance, analyze failure modes, and execute targeted optimizations in a cost-efficient and privacy-aware manner.
In this work, we introduce a MAPE-based data flywheel framework that enables continuous learning in enterprise AI agents through a modular, feedback-driven pipeline. Adapted from self-adaptive control loops, this framework supports the deployment of agents that evolve incrementally over time. We apply this approach to NVIDIA’s deployment of NVInfo AI, an internal Knowledge Assistant Agent that serves over 30,000 employees across diverse domains including engineering, operations, HR, and sales. NVInfo AI integrates user feedback with performance telemetry to identify actionable failure signals and execute targeted updates using parameter-efficient fine-tuning (PEFT) and model specialization.
Over a three-month observation window, we collected and analyzed 495 negative feedback samples, revealing two dominant sources of failure: routing errors (5.25 %) and query rephrasal errors (3.2 %). Utilizing NVIDIA NeMo microservices, we applied lightweight, component-specific fine-tuning strategies to improve performance.
- For routing, we reduced model size by a factor of ten (from 70 billion to 8 billion parameters) while maintaining 96% accuracy and reducing latency by 70%.
- For query rephrasal, we achieved a 3.7% improvement in accuracy (measured on a synthetic dataset generated from manually analyzed incorrect queries, expanded to 5,000 examples and split 80/10/10), along with a 40% reduction in response latency.
This work makes three key contributions:
- We demonstrate a novel application of the MAPE control loop to the domain of GenAI agent improvement, bridging observability and action in a continuous feedback pipeline.
- We present an empirical analysis of post-deployment failure modes in a production-grade enterprise AI agent, informed by real user feedback.
- We provide a modular implementation blueprint using NVIDIA NeMo microservices, offering a practical architecture for organizations seeking to build adaptive and self-correcting AI agents.
II Background and Related Work
II-A From MAPE-K to Agentic AI: Foundations of Self-Adaptive Systems
The MAPE-K (Monitor, Analyze, Plan, Execute – Knowledge) reference model, introduced by IBM [1], remains foundational for designing self-adaptive software systems by structuring behavior into a control loop that continuously responds to environmental changes, with its modular architecture enabling broad adoption across multiple domains [2, 3, 4, 5, 6]. Central to its evolution is the Knowledge component, which supports long-term reasoning and intelligent adaptation, especially when integrated with machine learning to enable predictive and causal decision-making [7, 8, 9]. Within agentic AI frameworks, MAPE-K cycles are increasingly embedded in autonomous agents to drive real-time, decentralized adaptation that enables intelligent decision-making and responsive behavior in dynamic environments [10, 11]. As reinforcement learning and GenAI capabilities are incorporated into these loops, agents gain the ability to synthesize adaptive strategies and reason across modalities [12]. These advancements illustrate MAPE-K’s synergy with the data flywheel paradigm: each monitoring cycle enriches the knowledge base, fueling increasingly effective planning and adaptation through a self-reinforcing loop [13].
II-B Modular Pipelines for Scalable RAG: Retrieval, Routing, and Rephrasal
As enterprises adopt Retrieval-Augmented Generation (RAG) through staged pipelines that involve retrieval, grounding, reasoning, and feedback, they increasingly realize its value for scalability, compliance, and trustworthy AI, positioning RAG as a core enabler of intelligent adaptive systems [14]. By grounding large language models in enterprise knowledge and facilitating continuous real-world feedback, RAG complements the MAPE-K trajectory and reinforces the data flywheel paradigm [15, 16]. To support enterprise RAG deployments at scale, expert routing has emerged as a key architectural strategy for enabling adaptive, modular reasoning. Modular approaches such as Mixture of Experts (MoE) and multi-agent systems [17, 18] dynamically direct inputs to specialized components using techniques like embedding selectors, symbolic routing, and LLM-as-a-Router, enhancing efficiency and task-specific alignment [19, 20]. A complementary strand of research focuses on query understanding and rephrasal, which are critical for strengthening RAG pipelines. By mitigating ambiguity and poorly structured queries, these methods enhance retrieval accuracy and reduce hallucinations [21, 22]. Recent advances leverage LLM-based rephrasal, semantic parsing, and uncertainty-aware frameworks such as RaFe and Omni-RAG to clarify intent, improve retrievability, and boost response reliability in enterprise contexts [23, 24, 25].
As these pipelines scale, the choice of model architecture becomes critical for balancing performance and cost. While 8B models offer lower latency and cost, they typically underperform compared to 70B models [26, 27, 28]. Parameter-efficient fine-tuning methods such as Low-Rank Adaptation (LoRA) and Quantized Low-Rank Adaptation (QLoRA) narrow this gap by adapting smaller models to specific tasks with minimal overhead, achieving near-parity in performance while reducing memory and compute demands by up to 100 $×$ . This enables up to 60–80% savings in GPU costs without compromising accuracy [29, 30, 31, 32].
II-C Feedback and Evaluation: Closing the Adaptation Loop
To close the loop in these adaptive systems, human-in-the-loop (HITL) pipelines serve as a critical counterpart to automated feedback mechanisms. By embedding human expertise into monitoring, annotation, and evaluation stages, HITL workflows enhance the reliability and contextual accuracy of enterprise RAG deployments, particularly in high-stakes domains where model errors can lead to significant consequences [33, 34]. Modern approaches integrate subject matter experts, active learning, weak supervision, and toolkits such as Snorkel, Label Studio, and Prodigy to reduce annotation effort while enabling scalable and domain-aligned feedback cycles that continuously refine model behavior [35, 36, 37].
Robust evaluation plays a vital role in transforming HITL and system-generated feedback into actionable signals for model refinement. As a key driver of the data flywheel, it determines which behaviors to reinforce, retrain, or discard. Beyond traditional metrics like accuracy and latency, emerging methods such as LLM-as-a-Judge, reward modeling, and preference-based scoring more effectively capture alignment and robustness [38, 39, 40, 41, 42].
II-D MAPE-K-Aligned Data Flywheel for Self-Improving Enterprise GenAI Systems
Despite significant advances in retrieval, expert routing, rephrasal, fine-tuning, HITL feedback, and evaluation, enterprise GenAI systems often lack a cohesive architecture to support continuous adaptation. These components are typically implemented in isolation, limiting coordination between observability, retraining, and evaluation workflows. This fragmentation hampers responsiveness and diminishes reliability in production environments.
This paper presents the first comprehensive application of MAPE-K principles to the improvement of AI agents in enterprise settings. We introduce a MAPE-K-aligned data flywheel architecture that consolidates monitoring, analysis, planning, and execution into a modular, closed-loop pipeline. Leveraging NVIDIA’s NeMo Microservices [43, 44, 45, 46], our framework integrates observability, feedback ingestion, fine-tuning, and evaluation, and supports secure, low-latency deployment across cloud, on-premises, and hybrid environments with built-in policy enforcement and real-time feedback handling [47, 48].
By applying control-theoretic foundations to retrieval-augmented, multi-agent GenAI systems, our approach enables dynamic, self-improving behavior in production. The data flywheel continuously refines the knowledge base through each cycle of monitoring and evaluation, guiding targeted adaptations over time. This architecture provides a scalable and reliable foundation for building enterprise GenAI systems that evolve with real-world usage.
III System Architecture
III-A NVInfo AI: Mixture of Experts Architecture
Before describing the Adaptive Data Flywheel, we first present the underlying AI system it enhances. The NVInfo AI system operates as NVIDIA’s internal enterprise chatbot which provides services to more than 30,000 staff members spread across different locations worldwide. The system operates with an advanced Mixture of Experts (MoE) framework which optimizes its performance when processing various enterprise information requests.
III-A1 Architecture Components
The NVInfo AI system consists of multiple essential components (Figure 2, Appendix A) which work together to generate precise answers that understand user context.
- User Interface: The intranet portal functions as the main access point which allows staff members to ask questions and handles complex business information requirements across various domains. The system offers
- User questions through natural language while maintaining context understanding
- Response Generation in table, lists and formatted data structure
- Source references which link directly to SharePoint documentation
- Follow-up question suggestions generated from conversational context
- Feedback system which uses thumbs up/down buttons to help agents improve their performance.
- Router Module: The system uses Llama 3.1 70B as its initial large language model to classify user queries which then get sent to one of seven specialized experts.
- Financial Info Expert (earnings reports, transcripts)
- IT Help & HR Benefits Expert (ServiceNow knowledge and catalog)
- SharePoint Expert (intranet content)
- Holidays Expert (region-specific holiday calendars)
- Cafe Menu Expert (cafeteria information)
- People Expert (organization charts, reporting chains)
- NVIDIA Policies Expert (corporate policies and procedures)
- Query Processing Pipeline: The system processes queries through multiple stages after they pass through the router module.
1. Conversation Rephrasing: Incorporates prior turns for multi-turn dialogue.
1. Query Variations: Generates multiple rephrasings to improve retrieval coverage.
1. Retriever: Conducts semantic document searches across all available document collections.
1. Re-ranking & De-duplication: Ranks documents based on their relevance while removing duplicate results.
1. Answer Generation: Creates a unified response by processing the retrieved information.
1. Citation Generation: Produces trustworthy source links which enable users to verify information sources.
1. Suggested Follow-ups: Generates additional questions which help users discover new content while enhancing their interaction with the system.
III-A2 NVInfo AI Conversation and Feedback Collection
NVInfo AI system responds to user inquiries and simultaneously records extensive conversational data and user feedback information. The system records detailed response metrics and structured feedback metrics which get processed through a single data pipeline for system monitoring, evaluation and performance enhancement (Figure 3, Appendix B).
- User Interaction: The NVInfo User Interface enables Users to initiate queries. The NVInfo Agent receives input from the interface to execute query interpretation, information retrieval and response generation.
- Response Metrics Collection: The system generates NVInfo Response Metrics for every response it produces to enable observability and future analysis. The system tracks the following information points
- Query – the original user input
- Response – the agent’s output
- Category – the knowledge source from which information was retrieved
- Expert Selected – subject-matter expert or expert route chosen
- Time Taken – latency observed across different components in the agentic AI workflow
- Agent Thought – reasoning trace behind the response
- Rephrased Query – any reformulation of the user’s input
- IR Results – intermediate retrieval results
- Prompts – the prompt(s) used in response generation
- Guardrail Metrics – policy or safety checks applied to the response
These metrics are stored in DynamoDB, allowing fast, scalable storage of large volumes of response data.
- Feedback Metrics Collection: Users can give direct feedback through the thumbs up / thumbs down system after reviewing their responses. Users can access a feedback modal through these icons which allows them to add more details about their feedback. The system records the following NVInfo Feedback Metrics:
- Positive or negative signal (thumbs up/down)
- Contextual reasons for feedback, such as:
- Usefulness of cited sources
- Relevance of the generated response
- Clarity and completeness of the output
- Suggestions for improvement
These structured feedback metrics are stored in a SQL Database, making them easy to query for analytics and fine-grained error analysis.
- Unified Data Pipeline: The data ingestion pipeline receives both response metrics from DynamoDB and feedback metrics from the SQL Database. The data transformation process within this pipeline creates a uniform schema structure while adding supplementary information and establishing signal type connections to achieve system interaction understanding. The processed data moves to a central Data Lake where it becomes available for comprehensive analysis, continuous monitoring, and iterative improvements to the NVInfo AI system.
III-A3 Performance Characteristics
The baseline NVInfo AI operated with the following system metrics before Data Flywheel implementation:
- Average response time: $\,\sim\,$ 12 seconds per query
- LLM as judge ratings: 4.2 correctness score out of 5 measured on our regression dataset (see Appendix F)
- Weekly query volume: $\,\sim\,$ 2000 unique queries across 800 unique users
This Mixture of Experts framework serves as the base structure which our Adaptive Data Flywheel system uses to enhance particular experts through user feedback analysis.
<details>
<summary>dataflywheel_architecture.png Details</summary>

### Visual Description
## System Diagram: Nemo Microservices Platform
### Overview
The image is a system diagram illustrating the data flow and components of the Nemo Microservices Platform. It shows how user input is processed through various modules, including LLM-as-a-Judge, Retriever, and Answer Generation, with feedback loops for continuous model optimization. The diagram also highlights potential error points and the integration of user feedback and logs.
### Components/Axes
* **Nodes:**
* User (represented by a blue person icon)
* NeMo Guardrails (green hexagon with a network icon)
* NVINFO Front-end (gray speech bubble icon)
* Router (green hexagon with a network icon)
* NVINFO Expert (yellow gear icon)
* Financial Expert (blue person icon with a speech bubble)
* Holiday Expert (blue person icon with a speech bubble)
* NVHelp Expert (blue person icon with a speech bubble)
* mixtral-8x22b NIM (green hexagon with a network icon)
* Rephrased Queries (yellow gear icon)
* Vector DB (brown cylinder icon with a magnifying glass)
* NeMo Retriever Reranking NIM (green hexagon with a network icon)
* Answer Generation NIM (green hexagon with a network icon)
* NeMo Retriever Embedding NIM (green hexagon with a network icon)
* Answer Generation with Citations (yellow gear icon)
* NIM (green hexagon with a network icon)
* NeMo Customizer (green hexagon with a network icon)
* NeMo Datastore (green hexagon with a network icon)
* NeMo Evaluator (green hexagon with a network icon)
* NeMo Deployment Manager (green hexagon with a network icon)
* SME (represented by a blue person icon)
* **Error Nodes:**
* Router Error (pink box)
* Query Rephrasal Error (pink box)
* Retriever Error (pink box)
* Answer Generation Error (pink box)
* **Flow Direction:** The diagram uses arrows to indicate the flow of data and processes.
* **Labels:**
* LLM-as-a-Judge
* Query Rephrasal Error
* Retriever Error
* Answer Generation Error
* Vector DB
* mixtral-8x22b NIM
* Rephrased Queries
* NeMo Retriever Reranking NIM
* Answer Generation NIM
* NeMo Retriever Embedding NIM
* Router Error
* NVINFO Expert
* User
* NeMo Guardrails
* NVINFO Front-end
* Router
* Financial Expert
* Holiday Expert
* NVHelp Expert
* User Feedback
* SME
* NIM Logs
* NeMo Customizer
* NeMo Datastore
* NeMo Evaluator
* NeMo Deployment Manager
* NIM
* Continuous Model Optimizations
* Answer Generation with Citations
* Data Flywheel: Nemo Microservices Platform + User Feedback + Logs
* Response
* Fine-Tuned Smaller, Faster Model
### Detailed Analysis or Content Details
1. **User Interaction:**
* The process starts with a "User" interacting with "NeMo Guardrails."
* "NeMo Guardrails" connects to "NVINFO Front-end."
2. **Routing and Expert Selection:**
* "NVINFO Front-end" feeds into a "Router."
* The "Router" directs the flow to "NVINFO Expert."
* "NVINFO Expert" branches out to "Financial Expert," "Holiday Expert," and "NVHelp Expert."
3. **LLM as a Judge:**
* "NVINFO Front-end" also connects to "mixtral-8x22b NIM" under "LLM-as-a-Judge."
4. **Query Processing:**
* From "NVINFO Expert," the flow goes to "Rephrased Queries."
* "Rephrased Queries" interacts with "Vector DB" (two instances).
5. **Answer Generation:**
* "Vector DB" feeds into "NeMo Retriever Reranking NIM."
* "NeMo Retriever Reranking NIM" connects to "Answer Generation NIM."
* "Answer Generation NIM" connects to "NeMo Retriever Embedding NIM."
* "Answer Generation NIM" also connects to "Answer Generation with Citations."
6. **Model Optimization:**
* "NVINFO Expert" sends a "Response" to "NeMo Retriever Reranking NIM."
* "NeMo Retriever Reranking NIM" connects to "NIM" via "Fine-Tuned Smaller, Faster Model."
* "NIM" generates "NIM Logs."
* "User Feedback" and "NIM Logs" are fed into the "Data Flywheel."
7. **Data Flywheel:**
* The "Data Flywheel" consists of "NeMo Customizer," "NeMo Datastore," "NeMo Evaluator," and "NeMo Deployment Manager."
* The "Data Flywheel" is connected to "SME" and "Continuous Model Optimizations."
8. **Error Handling:**
* "Router Error" is associated with the "Router."
* "Query Rephrasal Error" is associated with "Rephrased Queries."
* "Retriever Error" is associated with "Vector DB."
* "Answer Generation Error" is associated with "Answer Generation NIM."
### Key Observations
* The diagram illustrates a complex system with multiple interconnected components.
* The system incorporates feedback loops for continuous improvement.
* Error handling is explicitly addressed with designated error nodes.
* The "Data Flywheel" represents a closed-loop system for model optimization.
### Interpretation
The diagram depicts a sophisticated microservices architecture designed for question answering or information retrieval. The system leverages multiple models and data sources to provide accurate and relevant responses to user queries. The inclusion of feedback loops and error handling mechanisms suggests a focus on continuous improvement and robustness. The "Data Flywheel" concept emphasizes the importance of data-driven optimization in this architecture. The system is designed to handle user input, route it to the appropriate expert, retrieve relevant information, generate an answer, and continuously improve its performance based on user feedback and system logs.
</details>
Figure 1: Adaptive Data Flywheel Architecture showing the MAPE control loop implementation for AI agent improvement
III-A4 RAG System Challenges
Before introducing our solution, it is important to understand the failure points inherent in RAG-based systems. Figure 4 illustrates the failure modes observed in production, ordered by their occurrence in the processing pipeline (see Appendix C):
The RAG pipeline encounters multiple processing challenges throughout its entire operation:
1. Router - Query Understanding: Misclassification of user intent leading to wrong expert selection. Example: ”vacation days” queries routed to Holiday Expert instead of Policies Expert (5.25% of our failures).
1. Query Rephrasing Error: Incorrect expansion or interpretation of queries for the selected agent. Example: ”RESS planning team” incorrectly rephrased as ”Resource Planning team” instead of ”Real Estate & Site Services” (3.2% of failures).
1. Retriever Error: Failure to find relevant documents which exist in the knowledge base because of semantic search limitations or embedding mismatches.
1. Reranking Error: Retrieved documents incorrectly prioritized which results in important information being hidden beyond the context window threshold.
1. LLM Hallucination: The model produces believable yet false information when it lacks sufficient context which leads to confident but incorrect responses.
1. Citation Generation Error: Incorrect or missing source references which decreases answer reliability and blocks users from verifying the information.
1. Answer Generation Error: A poor final response by combining retrieved context which results in incomplete or unclear answers even though it has access to correct information.
The MAPE control loop of our Adaptive Data Flywheel system solves these problems by implementing automatic detection and correction of these system failures.
III-B Adaptive Data Flywheel Overview
Building upon the NVInfo AI architecture and addressing RAG system challenges, Figure 1 illustrates how our Adaptive Data Flywheel wraps around the core system to enable continuous improvement. The flywheel contains the four MAPE phases with dedicated components for AI agent management which operate through a unified knowledge base.
III-B1 Monitor Component (M in MAPE)
1. Problem: The numerous failure points in RAG pipelines make it challenging to identify between situational and systemic problems. Systemic problems remain undetected until users file complaints which results in negative user experiences and delayed solution implementation.
1. Solution: We implemented a comprehensive monitoring system that tracks both direct user feedback (thumbs up/down) and implicit signals (re-queries, session abandonment). Table I shows two examples out of 495 queries that users gave a ”thumbs-down” in the first three months after release. These queries demonstrate how two specific data points could be generalized into larger patterns for developers to address.
1. Challenges: The collection of user feedback in AI systems encounters multiple obstacles which reduce the quality of evaluation signals. The main problem stems from insufficient user engagement because users only provide feedback to a limited extent which results in unrepresentative data collection. Users tend to report negative feedback more frequently because they focus on sharing their dissatisfaction with unsatisfactory results instead of verifying positive outcomes. The process of data collection becomes harder because organizations need to remove all personally identifiable information (PII) from query–response pairs while following strict privacy and security regulations. The available feedback data remains incomplete because users mainly provide basic binary feedback such as thumbs up or down without explaining their reasons for dissatisfaction. Users sometimes provide incorrect feedback through comments that fail to pinpoint the actual cause of their negative ratings. The combination of these factors makes it difficult to obtain dependable feedback that can help systems improve their performance.
1. Learnings: User feedback collection needs systems that combine user-friendly interfaces with privacy protection features to achieve better participation rates and useful data insights. The system should allow users to rate content directly and complete brief surveys and follow step-by-step prompts for feedback collection while maintaining full compliance with GDPR and CCPA regulations and enterprise data protection policies that ban personal information storage. The system should record both direct ratings and indirect feedback indicators which include user interaction statistics, search pattern changes and repeated query attempts. The system needs to request positive feedback through confirmation questions (e.g. “Was this answer helpful?”) together with negative feedback collection methods. Real-time feedback processing systems need to be established for immediate input evaluation and classification to shorten model improvement cycles.
TABLE I: Representative Error Examples Captured by Monitor Component During 3-Month Deployment
| ”What is the role of the RESS planning team at NVIDIA?” | Unable to find answer - RESS incorrectly expanded to ”Resource Planning team” instead of ”Real Estate & Site Services” | Query Rephrasing | Failed to retrieve correct department information |
| --- | --- | --- | --- |
| ”How many vacation days does NVIDIA Canada have?” | ”I don’t have enough information to answer this question” | Router Error | Sent to Holiday Expert instead of Policies Expert |
III-B2 Analyze Component (A in MAPE)
1. Problem: Raw feedback data tends to lack actionable insights. The RAG pipeline contains multiple failure points (see Figure 4) which makes it difficult to identify original causes and determine which components caused the errors. Without accurate error attribution, developers may introduce fixes that fail to significantly improve answer quality.
1. Solution: We developed systematic error attribution techniques combining manual analysis with automated classification. From 495 thumbs-down samples:
- Routing Errors: 26/495 (5.25%) - Queries sent to wrong expert
- Rephrasal Errors: $\sim$ 3.2% (extrapolated from analyzing 250/495 samples)
Although the NVInfo expert routing classifier demonstrated high overall accuracy, our analysis revealed that certain low-frequency query classes exhibited poor data representation. This distributional imbalance led to occasional misclassifications within those specific subsets. Recognizing this gap, we designed targeted experiments to enrich the data and improve performance in those underrepresented domains. Specific examples identified:
- Routing Error: ”How many vacation days does NVIDIA Canada have?” was sent to NVINFO’s Holiday Expert instead of the Policies Expert
- Rephrasal Error: ”RESS planning team” incorrectly rephrased as ”NVIDIA Resource Planning team” instead of ”Real Estate & Site Services”
1. Challenges: The RAG pipeline contains multiple failure points throughout its different stages as shown in Section III. The propagation of initial routing mistakes through subsequent components leads to cascading errors which grow more severe with each stage. The process of manual analysis creates a bottleneck because expert review is needed to perform accurate attribution. The identification of root causes becomes difficult when issues present as ambiguous failures because multiple dependent factors create the overall error.
1. Learnings: The RAG pipeline needs tracing functionality to track queries, retrieval operations and model choices because this will help developers debug the system efficiently and identify where failures occur. The attribution models which use heuristics or machine learning classifiers help identify which stages of the pipeline produce errors. The system needs to distinguish between model-related breakdowns and non-model problems because this separation enables developers to identify LLM-related errors from retrieval and ranking system errors. The evaluation of different system configurations (chunking methods and embedding models) through A/B testing will show their individual performance effects. The process of error classification and root-cause identification becomes faster through automated issue labeling which uses weak supervision or heuristic tagging methods.
III-B3 Plan Component (P in MAPE)
1. Problem: The developers need to make extensive modifications across multiple system components to fix the fundamental problems they have discovered. The combination of restricted labeled data, privacy restrictions and specialized domain requirements makes standard model retraining methods ineffective.
1. Solution: We developed targeted data curation and fine-tuning strategies for different problems leveraging NVIDIA NeMo microservices: Routing Error Remediation:
- Collected user feedback + SME corrected completions
- Used LLM-as-a-Judge: 140 incorrect routing identified, 32 truly incorrect
- Created ground truth: 761 data points (729 original + 32 corrections)
- Final dataset: 685 samples after removing duplicates (60/40 train/test split)
Rephrasal Error Remediation:
- Collected user feedback and manually analyzed 250/495 thumbs-down samples
- 10 incorrect rephrasal identified, shortlisted 4
- Synthetic data generation: Generated 5,000 synthetic samples using 4 shortlisted examples as few-shot prompts to Llama 3.1 405B (see Appendix G)
- Final dataset: 5,000 synthetic data samples (80/10/10 train/validation/test split)
Implementation Tools:
- NeMo Curator: Data processing and cleaning
- NeMo Customizer: Model fine-tuning and adaptation
- NeMo Evaluator: Model evaluation and testing
- NeMo Guardrails: Safety and quality assurance
1. Challenges: Developing targeted remediation strategies presents several challenges. The available training data consists of restricted labeled information because 495 production cases includes only 32 incorrect routing examples and 10 incorrect rephrasing instances. The learning process becomes more difficult because enterprise terminology and acronyms need specialized knowledge to understand their context. The model size requirements force developers to find an optimal point between performance and response time for maintaining system performance. The quality of synthetic data remains a problem because artificial examples need to exactly replicate actual user input and error behavior to achieve success.
1. Learnings: The LLM-as-a-Judge approach delivered excellent results by accurately detecting routing errors at a rate of 77%. The few-shot synthetic data generation method demonstrated excellent results because it needed only four to five examples to create high-quality training data. The domain-specific fine-tuning of smaller models produced results that were comparable to those of larger 70B models. The NVIDIA NeMo microservices stack modular design allowed developers to quickly test and optimize individual components which sped up the entire development cycle.
III-B4 Execute Component (E in MAPE)
1. Problem: The deployment of enhanced models to production requires various sequential operations which help reduce system downtime. The deployment of 70B parameter models leads to negative impacts on user experience and operational efficiency because they tend to have higher latency and cost.
1. Solution: Using NVIDIA NeMo Customizer, we executed model fine-tuning and progressive deployment: Router Optimization Results:
- Baseline: Llama 3.1 70B - 96% accuracy, 0.26s latency
- Fine-tuned: Llama 3.1 8B - 96% accuracy, 0.08s latency
- Achievement: 10x model size reduction, 70% latency reduction
Rephrasal Enhancement Results:
- Baseline: Llama 3.1 70B - 73.8% accuracy, 1.9s latency
- Fine-tuned: Llama 3.1 8B - 77.5% accuracy, 1.1s latency
- Achievement: 3.7% accuracy improvement, 40% latency reduction
1. Challenges: The system faces major production risks because any unwanted changes will affect more than 30,000 users by degrading system performance. The system requires effective rollback mechanisms to perform fast updates and reduce system downtime during problematic changes. The system requires ongoing performance tracking to monitor change effects on different query domains while maintaining uniform quality standards. The deployment process requires teams to work together effectively because data scientists need to coordinate with engineers and operations staff to handle dependencies and preserve system stability.
1. Learnings: The deployment process should include Canary and staged deployments to introduce changes to limited user groups before complete system deployment helps protect against unexpected system problems. The implementation of defined rollback procedures enables teams to safely return to previous updates when performance deterioration occurs. The monitoring of essential performance indicators including accuracy, latency and user feedback after deployment helps detect system deterioration at its beginning stages. The release process benefits from clear handoffs between data scientist, engineer and product manager which enables effective team collaboration. Users will develop more trust in new model versions when organizations maintain open communication about system updates.
IV Experimental Evaluation
IV-A Experimental Setup
We evaluated our Data Flywheel implementation on NVIDIA’s NVInfo bot through systematic feedback analysis and targeted improvements:
- User Base: 800 active users per week
- Feedback Dataset: 1,224 human feedback samples from production (729 thumbs-up, 495 thumbs-down)
- Baseline Models: Llama 3.1 70B for routing and query rephrasal
- Fine-tuning Models: Llama 3.1 8B, Llama 3.2 3B/1B
- Infrastructure: NVIDIA NeMo Customizer microservices for customization
IV-B Error Analysis from User Feedback
Through analysis of 495 negative feedback samples, we identified two primary failure modes:
TABLE II: Error Classification from User Feedback
| Routing Errors Rephrasal Errors Other Errors | 26/495 $\sim$ 16/495 453/495 | 5.25% 3.2% (extrapolated) 91.5% |
| --- | --- | --- |
Example failures identified:
- Routing Error: ”How many vacation days does NVIDIA Canada have?” was incorrectly routed to Holiday Expert instead of Policies Expert
- Rephrasal Error: ”What is the role of the RESS planning team?” failed due to incorrect expansion (RESS = Real Estate & Site Services)
IV-C Fine-Tuning Results
IV-C1 Infrastructure and Fine tuning Method
To address key failure modes in the NVInfo RAG pipeline, we adopted LoRA via PEFT to optimize critical components such as routing and query rephrasal. LoRA enables targeted updates to transformer weights using lightweight, low-rank matrices, making it well suited for rapid iteration on curated failure samples without requiring full model retraining.
All fine-tuning was performed on an NVIDIA DGX Station equipped with 4Ă— A100 GPUs (80 GB each), which provided the compute capacity needed for efficient parallel training on long-sequence, high-volume datasets.
IV-C2 Expert Routing Optimization
After gathering user feedback and incorporating subject-matter expert (SME) corrections, we compiled a curated dataset. In total, we collected 761 data points, consisting of 729 original samples and 32 additional corrections generated by the LLM-as-Judge. After removing duplicates, the dataset was reduced to 685 unique samples. For experimentation, we adopted a 60/40 split between training and testing sets.
TABLE III: Router Fine-Tuning Results: 10x Model Size Reduction
| Llama 3.1 70B (baseline) | 96% | 0.26 |
| --- | --- | --- |
| Llama 3.1 8B (no tuning) | 14% | 0.08 |
| Llama 3.1 8B + prompt-tuning | 86% | 0.08 |
| Llama 3.1 8B + fine-tuning | 96% | 0.08 |
| Llama 3.2 3B + fine-tuning | 94% | – |
| Llama 3.2 1B + fine-tuning | 94% | – |
Key achievement: Maintained 96% accuracy while reducing model size by 10x and latency by 70%.
IV-C3 Query Rephrasal Enhancement
We manually analyzed 250 samples, from which we identified 10 candidates for rephrasing. To further expand the dataset, we generated 5,000 synthetic samples using the Llama 3.1 405B model with few-shot examples. For downstream experiments, the data was partitioned into an 80/10/10 split across training, validation, and test sets.
TABLE IV: Query Rephrasal Fine-Tuning Results
| Llama 3.1 70B (baseline) Llama 3.1 8B Fine-Tuned | 73.8% 77.5% | 1.9 1.1 |
| --- | --- | --- |
Key achievement: 3.7% accuracy improvement with 40% latency reduction and 10x model size reduction.
IV-D Improvements Achieved Through the Data Flywheel
Table V shows examples of issues resolved through the data flywheel.
TABLE V: Examples of Corrected Issues through Data Flywheel
| ”What is the role of the RESS planning team at NVIDIA?” | Rephrasal Error: Incorrectly expanded to ”Resource Planning team” | Correct Rephrase: ”NVIDIA RESS planning team role”, ”RESS planning team responsibilities” | The role of RESS (Real Estate and Site Services) Planning team is to manage site operations, support lease delivery… |
| --- | --- | --- | --- |
| ”How many vacation days does NVIDIA Canada have?” | Router Error: Sent to Holiday Expert instead of Policies Expert | Correctly routed to Policies Expert | According to the Canada Vacation Policy, employees receive… |
V Discussion
V-A Key Achievements
V-A1 Model Size and Efficiency
The model achieved a 10Ă— reduction in size from 70B to 8B parameters while maintaining 96% routing accuracy. The results show that domain-specific models with smaller sizes can achieve comparable performance to their larger general-purpose counterparts when fine-tuned properly. The model size reduction enabled a 70% decrease in latency which made real-time applications more practical.
V-A2 Accuracy Improvements
The 3.7% improvement in query rephrasal accuracy shows significant advancement for production systems although it seems small at first glance. The combination of faster performance through reduced latency by 40% and improved accuracy leads to enhanced user experience through quicker and more accurate system interactions.
V-A3 Systematic Error Identification
The analysis of 495 feedback samples revealed that routing errors combined with rephrasal errors made up less than 10% of all system failures at 5.25% and 3.2% respectively. The results indicate that the retrieval and ranking and generation stages of the RAG pipeline need further optimization efforts.
V-B Practical Challenges Encountered
V-B1 Low Feedback Participation
The system received feedback from 495 employees out of 30,000 users which shows difficulties in obtaining large-scale feedback data. The insufficient number of participants in the study creates sampling bias which reduces the generalizability of the obtained results. The system uses query reformulation as an additional data source but it does not replace the need for direct user feedback.
V-B2 Manual Analysis Bottleneck
The process of reviewing 250 samples for rephrasal errors took too much time and proved to be unworkable at scale. The feedback pipeline faces a major bottleneck because LLM-as-Judge helps identify routing errors but there is no automated solution for this process.
V-B3 Privacy and Compliance
The enterprise policies restricted storing complete query-response pairs which restricted thorough analysis of the data. The process of handling feedback data became more complicated because of PII removal requirements and GDPR and CCPA compliance regulations.
V-B4 Synthetic Data Generation
The creation of 5,000 synthetic examples for rephrasal training proved successful. The process of maintaining high-quality and contextually accurate data required advanced prompt engineering techniques and validation procedures which raised operational costs for data augmentation.
V-C Implications for Enterprise AI
V-C1 Proprietary Data as Differentiator
The research proves that properly managed enterprise data through a data flywheel system creates a sustainable business advantage for organizations. Real-world usage data serves as the foundation for developing AI agents which become both adaptive and resilient.
V-C2 Importance of Modular Architecture
NVIDIA NeMo tools consisting of Curator and Customizer and Evaluator and Guardrails proved essential for establishing the data flywheel system. The platform’s modular design allowed separate optimization of individual components and fast development cycles which are essential for enterprise flexibility.
V-C3 TCO Reduction Through Model Optimization
The model size reduction by 10 times delivered better performance while simultaneously reducing infrastructure expenses. Organizations that operate multiple AI agents can achieve substantial TCO reductions through optimization efforts which preserve service quality standards.
V-D Future Work
V-D1 Automated Error Attribution
The development of machine learning classifiers which can identify all RAG pipeline errors stands as the main objective. The system will achieve better scalability in root cause analysis through automated error classification which decreases human involvement in review processes.
V-D2 Continuous Learning Without Forgetting
The ability to update models continuously while preserving current performance levels and preventing catastrophic forgetting remains vital for supporting incremental learning progress in changing enterprise environments.
V-D3 Multi-Agent Coordination
The data flywheel concept should advance to support coordinated development between different specialized agents which represents a logical progression. System-wide intelligence and stable performance in complex enterprise systems depend on coordinated efforts between different system components.
VI Conclusion
The research presented real-world data flywheel implementation results for NVInfo Knowledge Assistant at NVIDIA which showed how enterprise AI agents can learn from their errors to improve their performance. The MAPE framework enabled us to convert user feedback into quantifiable performance enhancements which improved both model accuracy and user interaction quality.
Our main accomplishments included reducing model size by 10 times (from 70B to 8B parameters) while achieving 96% routing accuracy and improving query rephrasal accuracy by 3.7% and decreasing latency by 40%. The experimental results confirm that optimized smaller models perform at least as well as larger models while reducing operational expenses.
The evaluation of 495 feedback samples showed that routing errors and rephrasal errors combined to make up only 8.45% of total failure cases thus becoming the most suitable targets for optimization efforts. The targeted refinement approach using restricted training data produced substantial performance enhancements which proved the effectiveness of focused model improvement without needing extensive retraining.
The deployment process revealed two essential lessons about the challenges of obtaining large-scale feedback data and the restrictions that enterprise privacy regulations create. The solution we developed using implicit signal detection and synthetic data creation and privacy-friendly processing methods provides organizations with a functional method to handle their data collection and privacy restrictions.
The success of enterprise workflows in the future will depend on continuous AI agent improvement capabilities. Organizations that establish data flywheel systems will experience faster AI system development and operational efficiency and enhanced user satisfaction. Enterprise AI success depends on creating adaptive systems which learn from all user interactions rather than seeking flawless models at deployment. AI agents that receive feedback-based learning capabilities will develop into self-enhancing assets which improve their capabilities and value through continuous evolution.
References
- [1] IBM Corporation, “An architectural blueprint for autonomic computing (4th ed.),” IBM White Paper, 2006.
- [2] D. L. Iglesia and D. Weyns, “Mape-k formal templates to rigorously design behaviors for self-adaptive systems,” ACM Transactions on Autonomous and Adaptive Systems, vol. 10, no. 3, p. Article 15, October 2015, 31 pages.
- [3] P. Arcaini, E. Riccobene, and P. Scandurra, “Modeling and analyzing mape-k feedback loops for self-adaptation,” in 2015 IEEE/ACM 10th International Symposium on Software Engineering for Adaptive and Self-Managing Systems, Florence, Italy, 2015, pp. 13–23.
- [4] E. Rutten, N. Marchand, and D. Simon, “Feedback control as mape-k loop in autonomic computing,” in Software Engineering for Self-Adaptive Systems III. Assurances, ser. Lecture Notes in Computer Science, R. de Lemos, D. Garlan, C. Ghezzi, and H. Giese, Eds. Springer, Cham, 2017, vol. 9640.
- [5] A. Romero-GarcĂ©s, A. Hidalgo-Paniagua, M. González-GarcĂa, and A. Bandera, “On managing knowledge for mape-k loops in self-adaptive robotics using a graph-based runtime model,” Applied Sciences, vol. 12, no. 17, p. 8583, 2022.
- [6] J. Andersson, M. Caporuscio, M. D’Angelo et al., “Architecting decentralized control in large-scale self-adaptive systems,” Computing, vol. 105, pp. 1849–1882, 2023.
- [7] O. Gheibi, D. Weyns, and F. Quin, “Applying machine learning in self-adaptive systems: A systematic literature review,” ACM Transactions on Autonomous and Adaptive Systems, vol. 15, no. 3, p. Article 9, September 2020.
- [8] I. Abdennadher, “Daacs: A decision approach for autonomic computing systems,” The Journal of Supercomputing, vol. 78, pp. 3883–3904, 2022.
- [9] N. Belhaj, “Generic autonomic service management for component-based applications,” Ph.D. dissertation, Université Paris Saclay (COmUE), 2018, artificial Intelligence [cs.AI].
- [10] K. Patel, “Agentic ai for self-healing production lines: Autonomous root cause analysis & correction,” Journal of Information Systems Engineering and Management, 2025.
- [11] C.-E. Hrabia, M. Lützenberger, and S. Albayrak, “Towards adaptive multi-robot systems: self-organization and self-adaptation,” The Knowledge Engineering Review, vol. 33, p. E16, 2018.
- [12] J. Li, M. Zhang, N. Li, D. Weyns, Z. Jin, and K. Tei, “Generative ai for self-adaptive systems: State of the art and research roadmap,” ACM Transactions on Autonomous and Adaptive Systems, vol. 19, no. 3, p. Article 13, September 2024.
- [13] B. P. Sanwouo, C. Quinton, and P. Temple, “Breaking the loop: Aware is the new mape-k,” in Proceedings of the 33rd ACM International Conference on the Foundations of Software Engineering (FSE Companion ’25). New York, NY, USA: Association for Computing Machinery, 2025, pp. 626–630.
- [14] R. Akkiraju, A. Xu, D. Bora, T. Yu, L. An, V. Seth, A. Shukla, P. Gundecha, H. Mehta, A. Jha, and P. Raj, “Facts about building retrieval augmented generation-based chatbots,” arXiv preprint arXiv:2407.07858, 2024.
- [15] Microsoft Research, “Arena learning: Build data flywheel for llms post-training via simulated chatbot arena,” 2024.
- [16] NVIDIA, “Maximize ai agent performance with data flywheels using nvidia nemo microservices,” NVIDIA Developer Blog, 2025.
- [17] W. Cai, J. Jiang, F. Wang, J. Tang, S. Kim, and J. Huang, “A survey on mixture of experts in large language models,” IEEE Transactions on Knowledge and Data Engineering, 2025.
- [18] Y. Zhou, T. Lei, H. Liu, N. Du, Y. Huang, V. Zhao, A. M. Dai, Q. V. Le, and J. Laudon, “Mixture-of-experts with expert choice routing,” in Advances in Neural Information Processing Systems, vol. 35, 2022, pp. 7103–7114.
- [19] J.-C. Y. Chen, S. Yun, E. Stengel-Eskin, T. Chen, and M. Bansal, “Symbolic mixture-of-experts: Adaptive skill-based routing for heterogeneous reasoning,” arXiv preprint arXiv:2503.05641, 2025.
- [20] S. Chen, W. Jiang, B. Lin, J. Kwok, and Y. Zhang, “Routerdc: Query-based router by dual contrastive learning for assembling large language models,” in Advances in Neural Information Processing Systems, vol. 37, 2024, pp. 66 305–66 328.
- [21] R. Li, L. He, Q. Liu, Z. Zhang, H. Yu, Y. Ye, L. Zhu, and Y. Su, “Unirag: Unified query understanding method for retrieval augmented generation,” in Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (ACL 2025), 2025, pp. 14 163–14 178.
- [22] G. Dong, X. Li, Y. Zhang, and M. Deng, “Leveraging llm-assisted query understanding for live retrieval-augmented generation,” arXiv preprint arXiv:2506.21384, 2025.
- [23] S. Mao, Y. Jiang, B. Chen, X. Li, P. Wang, X. Wang, P. Xie, F. Huang, H. Chen, and N. Zhang, “Rafe: Ranking feedback improves query rewriting for rag,” arXiv preprint arXiv:2405.14431, 2024.
- [24] A. Yang, C. Chen, and K. Pitas, “Just rephrase it! uncertainty estimation in closed-source language models via multiple rephrased queries,” arXiv preprint arXiv:2405.13907, 2024.
- [25] A. Shrivastava, S. Desai, A. Gupta, A. Elkahky, A. Livshits, A. Zotov, and A. Aly, “Retrieve-and-fill for scenario-based task-oriented semantic parsing,” arXiv preprint arXiv:2202.00901, 2022.
- [26] MyScale, “Llama 3.1 models: 405b vs 70b vs 8b - which one to choose?” 2024.
- [27] Amazon AWS, “Announcing llama 3.1 405b, 70b, and 8b models from meta in amazon bedrock,” 2024.
- [28] BentoML, “Benchmarking llm inference backends,” 2023.
- [29] E. J. Hu, Y. Shen, P. Wallis, Z. Allen-Zhu, Y. Li, S. Wang, L. Wang, and W. Chen, “Lora: Low-rank adaptation of large language models,” in International Conference on Learning Representations (ICLR), vol. 1, no. 2, 2022, p. 3.
- [30] T. Dettmers, A. Pagnoni, A. Holtzman, and L. Zettlemoyer, “Qlora: Efficient finetuning of quantized llms,” in Advances in Neural Information Processing Systems, vol. 36, 2023, pp. 10 088–10 115.
- [31] E. N. Coleman, L. Quarantiello, Z. Liu, Q. Yang, S. Mukherjee, J. Hurtado, and V. Lomonaco, “Parameter-efficient continual fine-tuning: A survey,” arXiv preprint arXiv:2504.13822, 2025.
- [32] J. Kim, J. Lee, G. Park, B. Kim, S. J. Kwon, D. Lee, and Y. Lee, “An inquiry into datacenter tco for llm inference with fp8,” arXiv preprint arXiv:2502.01070, 2025.
- [33] V. Vats, M. B. Nizam, M. Liu, Z. Wang, R. Ho, M. S. Prasad, V. Titterton, S. V. Malreddy, R. Aggarwal, Y. Xu, and L. Ding, “A survey on human-ai teaming with large pre-trained models,” arXiv preprint arXiv:2403.04931, 2024.
- [34] J. Gama, I. Žliobaitė, A. Bifet, M. Pechenizkiy, and A. Bouchachia, “A survey on concept drift adaptation,” ACM Computing Surveys, vol. 46, no. 4, p. 44, 2014.
- [35] A. Ratner, S. Bach, H. Ehrenberg, J. Fries, S. Wu, and C. Ré, “Snorkel: rapid training data creation with weak supervision,” Proceedings of the VLDB Endowment, vol. 11, no. 3, pp. 269–282, 2017.
- [36] Quotient Blog, “Subject-matter expert language liaison (smell): A framework for aligning llm evaluators to human feedback,” 2024.
- [37] D. Gong, P. Lu, Z. Wang, M. Zhou, and X. He, “Training agents with weakly supervised feedback from large language models,” arXiv preprint arXiv:2411.19547, 2024.
- [38] M. T. R. Laskar, S. Alqahtani, M. S. A. Bari, M. Rahman, M. A. M. Khan, H. Khan, I. Jahan, A. Bhuiyan, C. W. Tan, M. R. Parvez, and E. Hoque, “A systematic survey and critical review on evaluating large language models: Challenges, limitations, and recommendations,” arXiv preprint arXiv:2407.04069, 2024.
- [39] M. Gao, X. Hu, X. Yin, J. Ruan, X. Pu, and X. Wan, “Llm-based nlg evaluation: Current status and challenges,” Computational Linguistics, pp. 1–27, 2025.
- [40] S. Tan, S. Zhuang, K. Montgomery, W. Y. Tang, A. Cuadron, C. Wang, R. A. Popa, and I. Stoica, “Judgebench: A benchmark for evaluating llm-based judges,” arXiv preprint arXiv:2410.12784, 2024.
- [41] E. Frick, T. Li, C. Chen, W.-L. Chiang, A. N. Angelopoulos, J. Jiao, B. Zhu, J. E. Gonzalez, and I. Stoica, “How to evaluate reward models for rlhf,” arXiv preprint arXiv:2410.14872, 2024.
- [42] L. Zheng, W.-L. Chiang, Y. Sheng, S. Zhuang, Z. Wu, Y. Zhuang, Z. Lin, Z. Li, D. Li, E. Xing, and H. Zhang, “Judging llm-as-a-judge with mt-bench and chatbot arena,” in Advances in Neural Information Processing Systems, vol. 36, 2023, pp. 46 595–46 623.
- [43] NVIDIA, “Nemo — build, monitor, and optimize ai agents,” 2025, accessed: 2025-09-09.
- [44] NVIDIA Docs, “About evaluating — nvidia nemo microservices,” 2025, accessed: 2025-09-09.
- [45] NVIDIA, “Overview of nemo microservices,” 2025, accessed: 2025-09-09.
- [46] Constellation Research, “Nvidia nemo microservices generally available, aims for ai agent data flywheel,” 2025, accessed: 2025-09-09.
- [47] Unit8, “Road to on-premise llm adoption – part 3,” 2024, accessed: 2025-09-09.
- [48] Bitrock, “A comparative analysis of open-source large language models on hugging face,” 2024, accessed: 2025-09-09.
Appendix A NVInfo AI Architecture
NVInfo AI Architecture
<details>
<summary>nvinfo_architecture.png Details</summary>

### Visual Description
## Flow Diagram: Expert System Response Generation
### Overview
The image is a flow diagram illustrating the process of generating a response to a user query within an expert system. The diagram outlines the steps from initial user input to the final response, including various expert modules, data processing stages, and response generation components.
### Components/Axes
* **User:** Represents the initial user input.
* **Chat History and Query Contextualization:** A green hexagon that processes the user's input and chat history.
* **Task Planner Agent:** A green hexagon that plans the tasks required to answer the user's query.
* **Pre-facto Expert Selection:** A green hexagon that selects the appropriate expert modules based on the query.
* **Guardrailing (Direct Responses):** A green hexagon that provides direct responses or rejections for small talk.
* **Financial Info Expert:** A blue circle containing a robot icon, representing an expert module for financial information.
* **ITHelp & HR Benefits Expert:** A blue circle containing a robot icon, representing an expert module for IT help and HR benefits. This module has a light blue outline.
* **Sharepoint Expert:** A blue circle containing a robot icon, representing an expert module for Sharepoint information. This module has a light blue outline.
* **Holiday Expert:** A blue circle containing a robot icon, representing an expert module for holiday information.
* **Cafe Menu Expert:** A blue circle containing a robot icon, representing an expert module for cafe menu information.
* **People Expert:** A blue circle containing a robot icon, representing an expert module for people information.
* **Merged IR Content:** A yellow magnifying glass, representing the merging of information retrieval content.
* **NV Embedding Reranker:** A green hexagon that reranks the embedded content.
* **LLM Answer Summarization:** A green hexagon that summarizes the answer using a Large Language Model.
* **Citation Generation:** A yellow gear that generates citations for the response.
* **Financial Response with Citations:** A yellow gear that provides a financial response with citations.
* **Holiday Response with Citations:** A yellow gear that provides a holiday response with citations.
* **Cafe Menu Response with Citations:** A yellow gear that provides a cafe menu response with citations.
* **People Response with Citations:** A yellow gear that provides a people response with citations.
* **Suggested Follow up Question Generation:** A green hexagon that generates suggested follow-up questions.
* **Final Response:** An orange rectangle representing the final output to the user.
* **Direct Response (Small Talk or Rejection):** A text label indicating a direct response path.
### Detailed Analysis
1. **User Input:** The process begins with the user's query, represented by a blue person icon on the left.
2. **Contextualization:** The query is then processed by the "Chat History and Query Contextualization" module (green hexagon).
3. **Task Planning:** The contextualized query is passed to the "Task Planner Agent" (green hexagon).
4. **Expert Selection:** The "Pre-facto Expert Selection" module (green hexagon) determines the relevant expert modules.
5. **Expert Modules:**
* **Financial Info Expert:** If the query relates to financial information, this module (blue circle) is activated. It leads to "Financial Response with Citations" (yellow gear).
* **ITHelp & HR Benefits Expert:** For IT help or HR benefits queries, this module (blue circle with light blue outline) is used. It leads to "Merged IR Content" (yellow magnifying glass).
* **Sharepoint Expert:** For Sharepoint queries, this module (blue circle with light blue outline) is used. It leads to "Merged IR Content" (yellow magnifying glass).
* **Holiday Expert:** For holiday-related queries, this module (blue circle) is activated. It leads to "Holiday Response with Citations" (yellow gear).
* **Cafe Menu Expert:** For cafe menu queries, this module (blue circle) is activated. It leads to "Cafe Menu Response with Citations" (yellow gear).
* **People Expert:** For people-related queries, this module (blue circle) is activated. It leads to "People Response with Citations" (yellow gear).
6. **Data Processing:**
* **Merged IR Content:** The output from the "ITHelp & HR Benefits Expert" and "Sharepoint Expert" modules is processed by the "Merged IR Content" module (yellow magnifying glass).
* **NV Embedding Reranker:** The merged content is then reranked by the "NV Embedding Reranker" (green hexagon).
* **LLM Answer Summarization:** The reranked content is summarized by the "LLM Answer Summarization" module (green hexagon).
* **Citation Generation:** Citations are generated by the "Citation Generation" module (yellow gear).
7. **Response Generation:**
* The outputs from the "Financial Response with Citations," "Holiday Response with Citations," "Cafe Menu Response with Citations," and "People Response with Citations" modules, as well as the "Citation Generation" module, are combined to form the "Final Response" (orange rectangle).
8. **Follow-up Questions:** The "Suggested Follow up Question Generation" module (green hexagon) generates potential follow-up questions based on the response.
9. **Direct Response:** The "Guardrailing (Direct Responses)" module (green hexagon) can provide direct responses or rejections for small talk, bypassing the expert modules.
### Key Observations
* The diagram illustrates a modular approach to response generation, with different expert modules handling specific types of queries.
* The system incorporates several data processing steps, including content merging, reranking, and summarization, to improve the quality of the response.
* The inclusion of citation generation ensures that the response is well-supported and credible.
* The system can handle both complex queries requiring expert knowledge and simple queries that can be answered directly.
### Interpretation
The flow diagram depicts a sophisticated expert system designed to provide accurate and relevant responses to user queries. The system leverages a combination of expert modules, data processing techniques, and response generation components to deliver high-quality results. The modular design allows for easy expansion and adaptation to new domains of knowledge. The inclusion of follow-up question generation enhances the user experience by encouraging further interaction and exploration. The system is designed to handle a variety of queries, from simple small talk to complex requests requiring expert knowledge.
</details>
Figure 2: NVInfo AI Mixture of Experts Architecture showing the complete RAG pipeline with Router, seven specialized domain experts, query rephrasing, retrieval, reranking, answer generation, and citation generation components
The architecture shown in Figure 2 illustrates the complete NVInfo AI system, which processes employee queries through a sophisticated pipeline:
- Router: Classifies incoming queries and routes them to the appropriate domain expert
- Seven Domain Experts: Specialized models for Financial Info, IT Help & HR Benefits, SharePoint, Holidays, Cafe Menu, People, and NVIDIA Policies
- Query Processing Pipeline: Includes rephrasing, retrieval, reranking, answer generation, and citation generation
- Feedback Loop: Captures user satisfaction signals for continuous improvement
Appendix B NVInfo Response and Feedback Capture Architecture
NVInfo Response and Feedback Capture Architecture
<details>
<summary>nvinfo_data_capture.png Details</summary>

### Visual Description
## Diagram: NVInfo Data Collection
### Overview
The image is a diagram illustrating the NVInfo data collection process. It shows how user interactions, both through a user interface and direct feedback, are processed and stored in different databases before being ingested, transformed, and finally stored in a data lake.
### Components/Axes
* **Title:** NVInfo Data Collection
* **Left Side:**
* **User:** A blue icon representing a user.
* **NVInfo User Interface:** A green square icon with a grid and a plus sign.
* **NVInfo Agent:** A green hexagon icon with a cartoon of a person wearing glasses.
* **Middle Section (within a dashed rounded rectangle):**
* **On Response:** A blue circle icon with a chat bubble and code snippets.
* **NVInfo Response Metrics:** An orange square icon with lines representing a document.
* **Event Trigger:** A yellow square icon with a gear.
* **DynamoDB:** A purple square icon with stacked cylinders and a lightning bolt.
* **On user feedback:** A green circle icon with a thumbs-up and a red circle icon with a thumbs-down.
* **NVInfo Feedback Metrics:** An orange square icon with a grid pattern.
* **SQL Database:** An orange square icon with stacked cylinders and the letters "SQL".
* **Right Side:**
* **Data Ingestion:** A yellow square icon with three lines converging into a single line.
* **Data Transformation:** A yellow square icon with a gear and arrows rotating around it.
* **Data Lake:** An orange square icon with wavy lines.
* **Arrows:** Arrows indicate the flow of data between components.
### Detailed Analysis
1. **User Interaction:**
* A user interacts with the system through the NVInfo User Interface.
* The user also interacts with the NVInfo Agent.
2. **Data Collection:**
* User interactions trigger "On Response" events, which are then processed into NVInfo Response Metrics.
* User feedback (thumbs up/down) is collected as "On user feedback" and processed into NVInfo Feedback Metrics.
3. **Event Trigger and Storage:**
* Both Response Metrics and Feedback Metrics trigger events.
* Response Metrics are stored in DynamoDB.
* Feedback Metrics are stored in an SQL Database.
4. **Data Processing:**
* Data from DynamoDB and the SQL Database is ingested.
* The ingested data undergoes transformation.
5. **Data Lake:**
* The transformed data is stored in a Data Lake.
### Key Observations
* The diagram illustrates a data pipeline that collects user interactions and feedback, stores them in different databases, and then processes them for storage in a data lake.
* The use of different databases (DynamoDB and SQL Database) suggests that different types of data are being stored and managed.
* The data transformation step is crucial for preparing the data for analysis and use in the data lake.
### Interpretation
The diagram depicts a comprehensive data collection and processing pipeline for NVInfo. It highlights the importance of capturing both direct user input (through the interface and agent) and indirect feedback. The use of separate databases for response metrics and feedback metrics suggests a deliberate strategy for organizing and managing different data types. The final step of storing the transformed data in a data lake indicates a focus on enabling data-driven decision-making and analysis. The diagram suggests a system designed to learn from user interactions and feedback to improve its performance and user experience.
</details>
Figure 3: NVInfo AI Response and Feedback Capture Architecture showing the complete data collection, ingestion and transformation components
The figure 3 illustrates the end-to-end data flow from user interaction with the NVInfo AI system to structured data storage for future system improvement. It highlights two main types of data captured(response metrics and user feedback metrics) and their subsequent processing.
- User Interaction and Metrics Collection: The data flow begins when a user interacts with the NVInfo User Interface, which connects to the NVInfo Agent, a domain-aware generative AI assistant that delivers structured, context-rich responses with citations. Each response is logged as part of NVInfo Response Metrics, capturing details such as query intent, routing, latency, and completeness. If the user provides feedback (e.g., thumbs up or down), NVInfo Feedback Metrics are recorded, including sentiment, error types, and optional comments. These metrics trigger events that stream response data to DynamoDB and feedback data to a SQL database, enabling structured downstream processing.
- Data Ingestion and Transformation: A centralized data ingestion pipeline runs every 4 hours via a scheduled cron job to extract the latest response and feedback records from DynamoDB and SQL databases. This ensures timely synchronization while minimizing system load during peak usage periods.
- PySpark-based Data Transformation: The ingested data is processed through a PySpark-based pipeline that performs cleaning, normalization, and enrichment. It maps feedback to specific conversation sessions, standardizes sentiment scores, and parses routing and rephrasal trace logs to identify failure modes. The resulting structured views capture model-side performance metrics such as routing accuracy and response latency, as well as user-side indicators like feedback sentiment and interaction quality, together providing a holistic picture of system effectiveness.
- Data Lake Storage: The structured outputs are stored in a scalable data lake for long-term access and analysis. These views support downstream tasks such as dashboarding, fine-tuning, error analysis, and offline evaluation, contributing to continuous improvement of the NVInfo Agent.
Appendix C RAG System Failure Points
RAG System Failure Points
<details>
<summary>failure_points_in_rag.png Details</summary>

### Visual Description
## Diagram: NVIDIA Expert System Flow
### Overview
The image is a flowchart illustrating the flow of a user query through an NVIDIA expert system. It shows the different stages of processing, from the initial user prompt to the final answer generation and citation. The diagram also highlights potential error points within the system.
### Components/Axes
* **Nodes:** The diagram consists of several nodes representing different components of the system. These nodes are represented by different shapes and colors.
* **Edges:** Arrows indicate the flow of information between the nodes.
* **Labels:** Each node is labeled with its function within the system.
* **Error Indicators:** Pink boxes indicate potential error points in the process.
### Detailed Analysis
1. **User:** A blue circle on the left represents the user.
2. **Prompt:** A gray speech bubble represents the user's prompt.
3. **Router:** A green hexagon represents the router, which directs the prompt to the appropriate expert.
* **Router Error:** A pink box below the Router indicates a potential "Router Error".
4. **Experts (Blue Circles):**
* NVIDIA Policies Expert
* IT Help & HR Benefits Expert
* NVIDIA Holiday Expert
* Financial Earnings Expert
5. **NVIDIA Company Info Expert (Yellow Square):** This expert is represented by a yellow square.
6. **Processing Pipeline (Enclosed in a dashed red box):**
* **Rephrased Queries (Yellow Square):** The query is rephrased.
* **Query Rephrasal Error:** A pink box above indicates a potential "Query Rephrasal Error".
* **Retriever (Orange Cylinder):** The rephrased query is used to retrieve relevant information.
* **Retriever Error:** A pink box above indicates a potential "Retriever Error".
* **Retriever Reranking (Green Hexagon):** The retrieved information is reranked.
* **Reranking Error:** A pink box above indicates a potential "Reranking Error".
* **Answer Generation (Green Hexagon):** An answer is generated based on the reranked information.
* **LLM Hallucination Error:** A pink box above indicates a potential "LLM Hallucination Error".
* **Citation Generation (Green Hexagon):** Citations are generated for the answer.
* **Citation Generation Error:** A pink box above indicates a potential "Citation Generation Error".
7. **Answer Generation with Citations (Yellow Square):** An alternative path leads to answer generation with citations.
* **Answer Generation Error:** A pink box below indicates a potential "Answer Generation Error".
### Key Observations
* The diagram illustrates a complex system with multiple stages of processing.
* The router directs the user's prompt to the appropriate expert.
* The processing pipeline involves query rephrasing, information retrieval, reranking, answer generation, and citation generation.
* Several potential error points are identified within the system.
* There are two paths to answer generation: one with citations and one without.
### Interpretation
The diagram provides a high-level overview of the NVIDIA expert system. It highlights the key components and the flow of information between them. The identification of potential error points is crucial for system monitoring and improvement. The presence of two answer generation paths suggests that the system can provide answers with or without citations, depending on the specific query or expert. The system uses a combination of different shapes and colors to represent the different components, making it easy to understand the overall flow. The diagram suggests a modular design, where each component can be independently developed and maintained. The system is designed to handle a variety of queries, as indicated by the different experts. The use of a router ensures that the query is directed to the appropriate expert.
</details>
Figure 4: Sequential failure points in the RAG pipeline from query routing to answer generation
The RAG pipeline, as shown in Figure 4, faces challenges at each processing stage. These failure points were identified through analysis of 495 negative feedback samples collected over 3 months:
- Router - Query Understanding: Misclassification of user intent (5.25% of failures)
- Query Rephrasing Error: Incorrect query expansion (3.2% of failures)
- Retriever Error: Failure to find relevant documents despite their existence
- Reranking Error: Incorrect prioritization of retrieved documents
- LLM Hallucination: Generation of plausible but incorrect information
- Citation Generation Error: Incorrect or missing source attribution
- Answer Generation Error: Poor synthesis of retrieved context
Appendix D NVInfo Interface Examples
NVInfo Interface Examples
<details>
<summary>IT_Help_QA.png Details</summary>

### Visual Description
## Screenshot: Remote Employee Monitor Ordering Instructions
### Overview
The image is a screenshot of a help document or FAQ addressing the question of how a remote employee can order a monitor. It provides step-by-step instructions and links to relevant resources.
### Components/Axes
* **Header:** "I am a remote employee. How do i order a monitor?"
* **Main Content:**
* Introductory statement: "As a remote employee, you can request a monitor by following these steps: [1]"
* Numbered list of steps:
* "Go to the Monitors catalog item on the ServiceNow portal. [2]"
* "Click on the 'Request a New or Replacement Monitor' button. [1]"
* "Fill out the request form with the required information, including your shipping address."
* "Submit the request."
* Additional information: "Please note that monitors are provided to remote employees, and they will be shipped directly to your home address via the vendor. [1] [3]"
* Alternative method: "Alternatively, you can also submit a request through the Coupa platform. [2]"
* Contact information: "If you have any issues or questions, you can reach out to the IT team for assistance."
* Closing statement: "Please let me know if you have any other questions or need further assistance."
* Links section: "For more details, please refer to the following links:"
* "[1] Monitors"
* "[2] Laptop / Desktop"
* "[3] Workstation Monitor Policy"
* **Footer:** "Here are some follow-up questions you can ask me:"
* "How long does it take to ship a monitor?"
* "What types of monitors are available for remote employees?"
* "Can I request a non-standard monitor configuration?"
### Detailed Analysis or ### Content Details
The document provides two primary methods for requesting a monitor:
1. **ServiceNow Portal:**
* Navigate to the "Monitors" catalog item.
* Click the "Request a New or Replacement Monitor" button.
* Complete the request form, including the shipping address.
* Submit the request.
2. **Coupa Platform:**
* Submit a request through the Coupa platform.
The document also notes that monitors are shipped directly to the employee's home address by the vendor.
The links section provides access to more detailed information about monitors, laptops/desktops, and the workstation monitor policy.
The follow-up questions section suggests potential inquiries related to shipping time, available monitor types, and non-standard configurations.
### Key Observations
* The document is designed for remote employees.
* Two methods are provided for requesting a monitor.
* The document emphasizes the direct shipping of monitors to the employee's home.
* Links are provided for additional information.
* The follow-up questions section anticipates common employee inquiries.
### Interpretation
The document aims to streamline the process of ordering a monitor for remote employees. By providing clear instructions and links to relevant resources, it reduces the need for employees to contact IT support for basic inquiries. The inclusion of follow-up questions suggests an effort to anticipate and address common concerns. The document highlights the company's policy of providing monitors to remote employees and shipping them directly to their homes, indicating a commitment to supporting remote work arrangements.
</details>
(a) IT Help Expert providing technical support for WFH monitor ordering
<details>
<summary>Financial_Earning_QA.png Details</summary>

### Visual Description
## Data Table and Explanatory Text: NVIDIA's Data Center Revenue
### Overview
The image presents NVIDIA's data center revenue for the past three fiscal quarters (Q4 2025, Q1 2026, and Q2 2026) in a table format, followed by a list of factors contributing to the revenue growth and the sources used to generate the information.
### Components/Axes
* **Table:**
* Columns: "Period" and "Data Center Revenue"
* Rows: Representing fiscal quarters (Q4 2025, Q1 2026, Q2 2026)
* **Explanatory Text:**
* A list of factors contributing to NVIDIA's data center revenue growth.
* A list of sources used to generate the response.
### Detailed Analysis
**Table Data:**
| Period | Data Center Revenue |
| ------------- | ------------------- |
| Q2 Fiscal 2026 | $41.1 billion |
| Q1 Fiscal 2026 | $39.1 billion |
| Q4 Fiscal 2025 | $35.6 billion |
**Explanatory Text:**
NVIDIA's data center revenue growth can be attributed to several factors, including:
* Strong demand for NVIDIA's AI infrastructure and Blackwell AI supercomputers
* Increased adoption of NVIDIA's data center products and technologies by cloud service providers, enterprises, and research institutions
* Expansion of NVIDIA's data center offerings, including the introduction of new products and services such as NVIDIA Blackwell Ultra, NVIDIA Dynamo, and NVIDIA Spectrum-X
* Partnerships and collaborations with leading companies in the industry, such as AWS, Google Cloud, Microsoft Azure, and Oracle Cloud Infrastructure
* Growing demand for AI computing and accelerated computing in various industries, including healthcare, finance, and autonomous vehicles.
**Sources:**
* NVIDIA Announces Financial Results for First Quarter Fiscal 2026 | NVIDIA Newsroom
* NVIDIA Announces Financial Results for First Quarter Fiscal 2026 | NVIDIA Newsroom
* NVIDIA Announces Financial Results for Fourth Quarter and Fiscal 2025 | NVIDIA Newsroom
### Key Observations
* NVIDIA's data center revenue has been increasing over the past three quarters.
* The revenue increased from $35.6 billion in Q4 Fiscal 2025 to $39.1 billion in Q1 Fiscal 2026, and further to $41.1 billion in Q2 Fiscal 2026.
* The growth is attributed to factors such as strong demand for AI infrastructure, increased adoption of NVIDIA's products, expansion of data center offerings, partnerships with leading companies, and growing demand for AI computing.
### Interpretation
The data indicates a positive trend in NVIDIA's data center revenue, suggesting strong performance in this segment. The identified contributing factors highlight the key drivers behind this growth, including the increasing demand for AI-related technologies and NVIDIA's strategic partnerships. The consistent citation of NVIDIA's official newsroom as the source adds credibility to the reported figures. The growth from Q4 2025 to Q2 2026 is approximately 15%, indicating a substantial increase in revenue over a short period.
</details>
(b) Financial Earnings Expert analyzing quarterly revenue growth
<details>
<summary>HR_Benefits_QA.png Details</summary>

### Visual Description
## Chat Log: NVIDIA HSA Contribution
### Overview
The image is a screenshot of a chat log or FAQ entry answering the question "How much does nvidia contribute to HSA?". The response indicates that NVIDIA contributes up to $3,000 to HSA plans, with the exact amount varying based on the plan and coverage level. It also provides links for more details and suggests follow-up questions.
### Components/Axes
* **Question:** "How much does nvidia contribute to HSA?"
* **Answer:** "Up to $3,000. NVIDIA contributes to the HSA plans, with the amount varying by plan and coverage level."
* **References:**
* "[1] Health Savings Accounts (HSA) | NVIDIA Benefits"
* "[2] What is a Health Savings Account (HSA)?"
* **Follow-up Questions:**
* "How to set up custom HSA contributions?"
* "What is the deadline to use my Day Care FSA balance?"
* "Can I roll over HSA funds to another account?"
### Detailed Analysis or ### Content Details
The main content is the answer to the question about NVIDIA's HSA contributions. The answer states a maximum contribution of $3,000, but clarifies that the actual amount depends on the specific plan and coverage level. The references provide links to further information about Health Savings Accounts and NVIDIA benefits. The follow-up questions suggest related topics that users might be interested in.
### Key Observations
* The maximum contribution amount is explicitly stated as $3,000.
* The contribution amount is not fixed and varies based on plan and coverage.
* Links are provided for more detailed information.
* Suggested follow-up questions are related to HSA contributions and related accounts.
### Interpretation
The chat log entry provides a concise answer to a common question about NVIDIA's HSA contributions. It sets expectations by stating the maximum possible contribution while also clarifying that the actual amount may vary. The links and follow-up questions enhance the user experience by providing additional resources and addressing related concerns. The information suggests that NVIDIA offers a Health Savings Account benefit to its employees, with varying contribution levels depending on the chosen plan and coverage.
</details>
(c) HR Benefits Expert explaining HSA contribution details
<details>
<summary>NVIDIA_product_doc_QA.png Details</summary>

### Visual Description
## Screenshot: NVIDIA Cloud Functions Explanation
### Overview
The image is a screenshot of a text-based explanation of NVIDIA Cloud Functions (NVCF). It describes NVCF as a serverless API for deploying and managing AI workloads on GPUs, highlighting its security, scale, and reliability. It also mentions suitable workloads and provides links for further details. Finally, it suggests follow-up questions.
### Components/Axes
* **Question:** "what is nvidia cloud functions?" at the top.
* **NVIDIA Logo:** Located to the left of the explanation text.
* **Explanation Text:** A detailed description of NVIDIA Cloud Functions.
* **References:** Links to "Overview - NVIDIA Cloud Functions" and "Function Lifecycle - NVIDIA Cloud Functions".
* **Follow-up Questions:** Three suggested questions: "How does NVCF secure AI workloads?", "What are the benefits of using NVCF?", and "Is NVCF suitable for long-running tasks?".
* **Edit/Delete Icons:** Located at the top-right corner.
* **Copy/Thumbs Up Icons:** Located to the right of the explanation text.
### Detailed Analysis or ### Content Details
* **NVIDIA Cloud Functions (NVCF) Description:** "NVIDIA Cloud Functions (NVCF) is a serverless API to deploy & manage AI workloads on GPUs, which provides security, scale and reliability to your workloads." \[1] \[2]
* **Workload Suitability:** "It is primarily suited for shorter running, preemptable workloads such as inferencing and fine-tuning."
* **Call to Action:** "Please let me know if you have any other questions or need further assistance."
* **Links:**
* "[1] Overview - NVIDIA Cloud Functions"
* "[2] Function Lifecycle - NVIDIA Cloud Functions"
* **Follow-up Questions:**
* "How does NVCF secure AI workloads?"
* "What are the benefits of using NVCF?"
* "Is NVCF suitable for long-running tasks?"
### Key Observations
* The explanation focuses on the benefits and use cases of NVIDIA Cloud Functions.
* It emphasizes the suitability of NVCF for shorter, preemptable workloads.
* The provided links offer further information on the overview and lifecycle of NVCF.
* The follow-up questions suggest common concerns or areas of interest regarding NVCF.
### Interpretation
The text provides a concise overview of NVIDIA Cloud Functions, targeting users who want to understand its purpose and capabilities. The emphasis on serverless architecture, AI workload management, and GPU utilization highlights the key value propositions of NVCF. The inclusion of follow-up questions indicates an attempt to anticipate user needs and encourage further engagement. The links to more detailed documentation allow users to delve deeper into specific aspects of NVCF.
```
</details>
(d) Company Info Expert providing NVC function information
Figure 5: Representative NVInfo AI interface examples showing mixture-of-experts responses across different enterprise domains
The interface examples above demonstrate the system’s capabilities:
- IT Support (Fig. 5(a)): Shows step-by-step guidance for ordering WFH equipment through the ServiceNow portal
- Financial Analysis (Fig. 5(b)): Provides quarterly revenue data with year-over-year growth metrics and detailed breakdowns
- HR Benefits (Fig. 5(c)): Displays HSA contribution tables with employer matching details for different fiscal quarters
- Product Documentation (Fig. 5(d)): Explains technical concepts like NVIDIA Cloud Functions with architecture overview
Appendix E Prompt for router error LLM-as-a-judge classification
Prompt for router error LLM-as-a-judge classification
⬇
1 Question: How do I submit a referral?
2 Tools: [’ it_benefits_help ’, ’ nvinfo_policies_expert ’]
3 Reasoning: This question is related to NVIDIA policy which means it should be sent to either ’ it_benefits_help ’ or ’ nvinfo_policies_expert ’.
4 Answer: YES
5
6 Question: When can I sign up for a new health plan?
7 Tools: [’ finance_expert ’]
8 Reasoning: This question is related to employee benefits which means it should be sent to ’ it_benefits_help ’ instead of ’ finance_expert ’.
9 Answer: NO
10
11 Question: what was NVIDIA ’ s Q3 revenue in fiscal 2024?
12 Tools: [’ finance_expert ’]
13 Reasoning: This question is related to NVIDIA ’ s earnings which means it should go to ’ finance_expert ’.
14 Answer: YES
15
16 Question: Is Mercedes Benz using NVIDIA ’ s digital twin technology?
17 Tools: [’ it_benefits_help ’, ’ nvinfo_policies_expert ’]
18 Reasoning: This question is related to NVIDIA ’ s products and therefore should have gone to ’ it_benefits_help ’.
19 Answer: YES
20
21 Question: What is the vacation policy at NVIDIA?
22 Tools: [’ nvinfo_holiday_expert ’]
23 Reasoning: This question is related to NVIDIA policy which means it should be sent to either ’ it_benefits_help ’ or ’ nvinfo_policies_expert ’.
24 Answer: NO
25
26 Question: When is the next free day at NVIDIA?
27 Tools: [’ nvinfo_holiday_expert ’]
28 Reasoning: The user is trying to find the date of a holiday which means that the question should be sent to [’ nvinfo_holiday_expert ’].
29 Answer: YES
30
31 Question: When is the first open stock sale period in 2025?
32 Tools: [’ finance_expert ’]
33 Reasoning: This question is related to NVIDIA ’ s company finances and should therefore be sent to ’ finance_expert ’.
34 Answer: YES
35
36 Question: How many unused vacation days can I carry over?
37 Tools: [’ it_benefits_help ’, ’ nvinfo_policies_expert ’]
38 Reasoning: This question is related to NVIDIA policy and employee benefits which means it should be sent to either ’ it_benefits_help ’ or ’ nvinfo_policies_expert ’.
39 Answer: YES
40
41 Question: Who heads up wwfo?
42 Tools: [’ finance_expert ’]
43 Reasoning: This question is related to NVIDIA ’ s leadership which means that it should be sent to ’ finance_expert ’.
44 Answer: YES
45
46 Question: Who is John Smith?
47 Tools: [’ finance_expert ’]
48 Reasoning: The user is trying to find information about a specific person which means that this question should go to ’ it_benefits_help ’ or ’ nvinfo_policies_expert ’.
49 Answer: NO
50
51 Question: What are the latest hardware offerings by nvidia?
52 Tools: [’ it_benefits_help ’, ’ nvinfo_policies_expert ’]
53 Reasoning: This question is related to NVIDIA ’ s products and therefore should have gone to ’ it_benefits_help ’.
54 Answer: YES
55
56 Question: What is gb200 nvl72?
57 Tools: [’ finance_expert ’]
58 Reasoning: This question is related to NVIDIA ’ s products and therefore should have gone to ’ it_benefits_help ’.
59 Answer: NO
60
61 Question: When will the 2025 free days be officially announced?
62 Tools: [’ it_benefits_help ’, ’ nvinfo_policies_expert ’]
63 Reasoning: This question is related to NVIDIA ’ s policies or benefits, so it should be sent to ’ it_benefits_help ’ or ’ nvinfo_policies_expert ’.
64 Answer: YES
65
66 Question: Does nvidia offer financial advice services?
67 Tools: [’ finance_expert ’]
68 Reasoning: This question is related to NVIDIA ’ s policies or benefits, so it should be sent to ’ it_benefits_help ’ or ’ nvinfo_policies_expert ’.
69 Answer: NO
70
71 Question: What was the year - over - year (YoY) and quarter - over - quarter (QoQ) growth for Q2 Fiscal 2025?
72 Tools: [’ finance_expert ’]
73 Reasoning: This question is related to NVIDIA ’ s earnings and should therefore be routed to ’ finance_expert ’.
74 Answer: YES
75
76 Question: How do I order a mouse?
77 Tools: [’ it_benefits_help ’, ’ nvinfo_policies_expert ’]
78 Reasoning: This question is related to procuring a work accessory, which means that it should go to either ’ it_benefits_help ’ or ’ nvinfo_policies_expert ’.
79 Answer: YES
80
81 Question: I ’ m getting a VPN error
82 Tools: [’ finance_expert ’]
83 Reasoning: This question is related to an IT issue, which means that it should go to either ’ it_benefits_help ’ or ’ nvinfo_policies_expert ’.
84 Answer: NO
85
86 QUERY: {query}
87 TOOLS: {experts}
Listing 1: Prompt for router error LLM-as-a-judge classification (complete example)
Appendix F Regression dataset
Regression Dataset
The NVInfo regression dataset is actively curated and regularly updated, currently comprising around 300 queries that cover a range of domains including NVIDIA benefits, holidays, company policies, and IT Help. Each query in the dataset contains the corresponding ground truth and expected citation values. LLM-as-judge framework is leveraged to evaluate the quality of NvInfo generated answers against the regression dataset. The criteria for judgment is based on metrics such as correctness, helpfulness, and conscientiousness.
Appendix G Synthetic data generation
Synthetic data generation
As part of our continuous improvement efforts, we identified that refining the way queries are phrased could significantly enhance retrieval accuracy. Given that our retrieval system is highly sensitive to keyword usage, an in-depth analysis of existing feedback was conducted. During this analysis, we discovered that decomposing certain queries into sub-queries improved the recall of relevant information from our internal retrieval systems. This, in turn, led to more accurate and contextually appropriate responses. To tackle this, we conducted a detailed review of about 250 examples from our ”thumbs down” feedback dataset, focusing specifically on queries related to our SharePoint expert system. We noticed that some queries weren’t retrieving the most relevant information due to a lack of understanding of NVIDIA-specific acronyms or context. By manually rephrasing these queries, we found that we could significantly improve the retrieval of the intended information. From an initial set of 250 examples, we identified 10 key candidates showcasing common patterns of misinterpretation or context loss. For instance, the query ”I am based in the netherlands, when is pay day?” was initially rephrased as ”payday schedule united states, employer pay dates usa.” We manually corrected this to ”payday schedule netherlands” and ”netherlands pay days.” To extend this improvement beyond the feedback dataset, we integrated these rephrased examples into our synthetic data generation pipeline. We leveraged these examples as few-shot prompts for our large language model. By providing SharePoint-related website content as context, we instructed the LLM to generate both original and rephrased queries for all documents. This method allowed us to produce approximately 5,000 rephrased queries, thereby enriching our dataset and facilitating more effective fine-tuning of the agent. This focused enhancement significantly improved the SharePoint expert’s ability to retrieve and deliver the most relevant information with increased accuracy.
Prompt for Synthetic Data Generation
You are a data annotator generating questions, answers, and rephrased questions from an input document and its URL. Guidelines •
Identify key phrases and entities in the document and generate questions around them. •
Generate questions answerable using information contained in the input document. •
Do not write questions that require viewing the document to understand the question. •
Avoid phrases like “according to the document/author”, “in this document”, etc. •
Questions may also be key phrases found in the document. •
Ensure the document contains the complete answer to your question. •
Provide enough context in the question to lead to the specific answer in the document. •
Vary phrasing, vocabulary, complexity, and type of questions. •
Do not copy exact phrasing; use your own words. •
Prefix questions with Question: and answers with Answer:. •
Rephrase each question at least twice (query decomposition/expansion) to aid search. •
Final output must be a Python list. •
Rephrased queries are short, concise keyword/entity mixes; you may replace nvidia with employer or company. •
Provide two or more rephrased queries preserving intent and timeframe. •
If the question asks for “the next X date” without time context, append YYYY (current or next year) in rephrased queries. Example: Question: “when is the next NTech conference” $→$ “upcoming ntech 2024”, “ntech dates 2024”, “ntech schedule 2025”. Use the EnterpriseKnowledge tool when The user asks for non-sensitive information such as organization info, direct reports, phone numbers, benefits alternate ID, email addresses, working addresses, tax explanations, updating SSN instructions, or stock trading policies. Your action format MUST be ⬇ Thought: Provide a short analysis of your understanding from the Question. Process: I need to use the Enterprise Knowledge tool Action: EnterpriseKnowledge Action Input: A single line Python list of rephrased queries MUST be generated. Strict JSON schema (return nothing else) ⬇ { "type": "object", "properties": { "Question": { "type": "string", "description": "Generated Question from the input document." }, "Answer": { "type": "string", "description": "Corresponding Answer from the input document that answers the Question." }, "Thought": { "type": "string", "description": "Short analysis of your understanding from the Question." }, "Process": { "type": "string", "description": "I need to use the Enterprise Knowledge tool." }, "Action": { "type": "string", "description": "EnterpriseKnowledge" }, "Action Input": { "type": "list", "description": "A single line Python list of rephrased queries." } } } Examples Input Document: <Content of input document> Input Document url: <url of input document> Output ⬇ { "Question": "I am based in the Netherlands, when is pay day?", "Answer": "25th of every month", "Thought": "Payroll timing question; include location keywords in rephrased queries.", "Process": "I need to use the Enterprise Knowledge tool", "Action": "EnterpriseKnowledge", "Action Input": [ "payday schedule netherlands", "netherlands pay days" ] } Input Document: <Content of input document> Input Document url: <url of input document> Output ⬇ { "Question": "point me to gpu fcv page?", "Answer": "https://nvidia.sharepoint.com/sites/TechnicalTraining/ASIC%20teams.aspx", "Thought": "Needs GPU FCV (Full Chip Verification) page.", "Process": "I need to use the Enterprise Knowledge tool", "Action": "EnterpriseKnowledge", "Action Input": [ "gpu fcv page company", "fcv gpu url" ] } Input Document: <Content of input document> Input Document url: <url of input document> Output ⬇ { "Question": "ok, i’m looking for an nvidia icon for biotech / pharmaceuticals to use in a presentation. can you help me find that?", "Answer": "https://nvidia.sharepoint.com/sites/nvinfo/brand/Pages/default.aspx", "Thought": "Needs a company icon for biotech/pharma use.", "Process": "I need to use the Enterprise Knowledge tool", "Action": "EnterpriseKnowledge", "Action Input": [ "company icons", "company logos biotech" ] } Task output format Generate 3 pairs by following the instructions based on the Input Document. Strictly return only a Python list of pairs and nothing else. Input Document: <Content of input document> Input Document url: <url of input document> Output: ###