# WorldGym: World Model as An Environment for Policy Evaluation
**Authors**:
- Percy Liang Sherry Yang (Stanford UniversityââNYUâââGoogle DeepMind)
Abstract
Evaluating robot control policies is difficult: real-world testing is costly, and handcrafted simulators require manual effort to improve in realism and generality. We propose a world-model-based policy evaluation environment (WorldGym), an autoregressive, action-conditioned video generation model which serves as a proxy to real world environments. Policies are evaluated via Monte Carlo rollouts in the world model, with a vision-language model providing rewards. We evaluate a set of VLA-based real-robot policies in the world model using only initial frames from real robots, and show that policy success rates within the world model highly correlate with real-world success rates. Moreoever, we show that WorldGym is able to preserve relative policy rankings across different policy versions, sizes, and training checkpoints. Due to requiring only a single start frame as input, the world model further enables efficient evaluation of robot policiesâ generalization ability on novel tasks and environments. We find that modern VLA-based robot policies still struggle to distinguish object shapes and can become distracted by adversarial facades of objects. While generating highly realistic object interaction remains challenging, WorldGym faithfully emulates robot motions and offers a practical starting point for safe and reproducible policy evaluation before deployment. See videos and code at https://world-model-eval.github.io
1 Introduction
Robots can help humans in ways that range from home robots performing chores (Shafiullah et al., 2023; Liu et al., 2024) to hospital robots taking care of patients (Soljacic et al., 2024). One of the major road blocks in the development robots lies in evaluation â how should we ensure that these robots will work reliably without causing any physical damage when deployed in the real world? Traditionally, people have used handcrafted software simulators to develop and evaluate robot control policies (Tedrake et al., 2019; Todorov et al., 2012; Erez et al., 2015). However, handcrafted simulation based on our understanding of the physical world can be limited, especially when it comes to hardcoding complex dynamics with high degrees of freedom or complex interactions such as manipulating soft objects (SĂŒnderhauf et al., 2018; Afzal et al., 2020; Choi et al., 2021). As a result, the sim-to-real gap has hindered progress in robotics (Zhao et al., 2020; Salvato et al., 2021; Dulac-Arnold et al., 2019).
With the development of generative models trained on large-scale video data (Ho et al., 2022; Villegas et al., 2022; Singer et al., 2022), recent work has shown that video world models can visually emulate interactions with the physical real world, by conditioning on control inputs in the form of text (Yang et al., 2023; Brooks et al., 2024) or keyboard strokes (Bruce et al., 2024). This brings up an interesting question â could video world models be used to emulate robot interactions with the real world, hence being used as an environment to evaluate robot policies in the world model before real-world testing or deployment?
Learning a dynamics model from past experience and performing rollouts in the learned dynamics model has been extensively studied in model-based reinforcement learning (RL) (Hafner et al., 2019; Fonteneau et al., 2013; Zhang et al., 2021; Kaiser et al., 2019; Yu et al., 2020). However, most of the existing work in model-based RL considers single-task settings, which puts itself at a disadvantage compared to model-free RL, since learning a dynamics model can be much harder than learning a policy in the single-task setting. Nevertheless, we make the important observation that
- While there can be many tasks and policies, there is only one physical world in which we live that is governed by the same set of physical laws.
This makes it possible to learn a single world model that, in principle, can be used as an interactive environment to evaluate any policies on any tasks.
<details>
<summary>x1.png Details</summary>

### Visual Description
## Diagram: Robot Task Execution System with World Model and VLM Reward
### Overview
This diagram illustrates a robotic task execution system that integrates a world model, policy execution, and vision-language model (VLM) reward evaluation. The system processes initial frames, language instructions, and out-of-distribution (OOD) inputs to generate and evaluate robotic actions.
### Components/Axes
1. **Left Panel: Initial Inputs**
- **Initial Frame and Language Instruction**: Contains two scenarios:
- *Evaluation Dataset Example*: "Put the eggplant in the pot" (correct instruction)
- *OOD Image Input*: Modified image with additional objects (red border)
- *OOD Language Instruction*: Modified instruction "Put the eggplant in the drying rack" (red border)
- **Key Elements**: Robot arm, sink environment, objects (eggplant, pot, drying rack)
2. **Central Panel: World Model and Policies**
- **World Model**: Central processing unit receiving sequential observations (oâ, oâ, oâ)
- **Policy Blocks**: Three identical policy modules processing observations (oââoâ) and outputting actions (oΞ)
- **Flow**: Observations feed into world model â policies â world model (recurrent loop)
3. **Right Panel: VLM Reward**
- **VLM as Reward**: Hexagonal symbol representing vision-language model
- **Output**: Reward value (RÌ) derived from policy evaluation
### Detailed Analysis
- **Initial Inputs**:
- Correct instruction: "Put the eggplant in the pot" (yellow box)
- OOD variations:
- Image: Additional objects (red border)
- Language: "Put the eggplant in the drying rack" (red border)
- **World Model**:
- Processes sequential observations (oââoâ) showing robot arm movement
- Maintains internal state (g) representing environment dynamics
- **Policy Execution**:
- Three identical policy modules process different observation states
- Outputs action sequences (oΞ) for robotic execution
- **VLM Reward System**:
- Evaluates policy outputs using vision-language model
- Generates scalar reward (RÌ) for action quality assessment
### Key Observations
1. **OOD Handling**: Red borders highlight system's ability to process instruction/image mismatches
2. **Recurrent Architecture**: World model maintains state between policy executions
3. **Modular Design**: Separate policy blocks suggest parallel processing capability
4. **Reward Integration**: VLM directly influences policy evaluation without explicit training signals
### Interpretation
This system demonstrates a closed-loop robotic control architecture where:
1. **World Model** serves as both environment simulator and memory
2. **Policies** generate actions based on current observations and historical context
3. **VLM Reward** provides real-time evaluation of action quality through vision-language understanding
4. **OOD Robustness**: The system explicitly handles instruction-image mismatches through separate OOD input channels
The architecture suggests a hierarchical approach where:
- Low-level policies execute basic actions
- World model maintains high-level context
- VLM provides semantic evaluation of action-instruction alignment
- Recurrent connections enable continuous learning from execution outcomes
The use of identical policy blocks implies transfer learning capabilities across different observation states, while the VLM reward system enables value-based policy selection without explicit reward shaping.
</details>
Figure 1: Overview of WorldGym. Given an initial frame and an action sequence predicted by a policy, WorldGym uses a world model to interactively predict future frames, serving as a generative simulator. WorldGym then passes the generated rollout to a VLM which provides rewards. WorldGym can easily be used to test policies on OOD tasks and environments by changing the input language instruction or directly modifying the initial image.
Inspired by this observation, we propose a world-model-based policy evaluation environment (WorldGym), as shown in Figure 1. WorldGym first combines knowledge of the world across diverse environments by learning a single world model that generates videos conditioned on actions. To enable efficient rollouts of policies which predict different-length action chunks, WorldGym aligns its diffusion horizon length with policiesâ chunk sizes at inference time. With video rollouts from the world model, WorldGym then uses a vision-language model (VLM) to determine tasksâ success from generated videos.
Our experiments show that WorldGym can emulate end-effector controls across different control axes highly effectively for robots with different morphologies. We then use the world model to evaluate VLA-based robot policies by rolling out the policies in the world model starting from real initial frames, and compare their success rates (policy values) in WorldGym to those achieved in real-world experiments. Our result suggests that policy values in WorldGym are highly correlated with policy performance in the real world, and the relative rankings of different policies are preserved.
Furthermore, as WorldGym requires only a single initial frame as input, we show how we can easily design out-of-distribution (OOD) tasks and environments and then use WorldGym to evaluate robot policies within these newly âcreatedâ environments. We find that modern robot policies still struggle to distinguish some classes of objects by their shape, and can even be distracted by adversarial facades of objects.
Although simulating realistic object interactions remains challenging, we believe WorldGym can serve as a highly useful tool for sanity check and testing robot policies safely and reproducibly before deploying them on real robots. Key contributions of this paper include:
- We propose to use video world model to evaluate robot policies across different robot morphologies, and perform a comprehensive set of studies to understand its feasibility.
- We propose flexibly aligning diffusion horizon length with policiesâ action chunk sizes for efficient rollouts of a variety of policies over hundreds of interactive steps.
- We show a single world model learned on data from diverse tasks and environments can enable policy value estimates that highly correlate with real-world policy success rates.
- We demonstrate the ease of testing robot policies on OOD tasks and environments within an autoregressive video generation-based world model.
2 Problem Formulation
In this section, we define relevant notations and review the formulation of offline policy evaluation (OPE). We also situate OPE in practical settings with partially observable environments and image-based observations.
Multi-Task POMDP.
We consider a multi-task, finite-horizon, partially observable Markov Decision Process (POMDP) (Puterman, 2014; Kaelbling et al., 1995), specified by $\mathcal{M}=(S,A,O,G,R,T,\mathcal{E},H)$ , which consists of a state space, action space, observation space, goal space, reward function, transition function, emission function, and horizon length. A policy $\pi$ interacts with the environment for a goal starting from an initial state $g,s_{0}\sim G$ , producing a distribution $\pi(·|s_{t},g)$ over $A$ from which an action $a_{t}$ is sampled and applied to the environment at each step $tâ[0,H]$ . The environment produces a scalar reward $r_{t}=R(s_{t},g)$ , and transitions to a new state $s_{t+1}\sim T(s_{t},a_{t})$ and emits a new observation $o_{t+1}\sim\mathcal{E}(s_{t+1})$ . We consider the sparse reward setting with $R(s_{H},g)â\{0,1\}$ and $R(s_{t},g)=0,â t<H$ , where $g$ is a language goal that defines the task. Data is logged from previous interactions into an offline dataset $D=\{g,s_{0},o_{0},a_{0},...,s_{H},o_{H},r_{H}\}$ . The value of a policy $\pi$ can be defined as the total expected future reward:
$$
\displaystyle\rho(\pi)= \displaystyle\mathbb{E}[R(s_{H},g)|s_{0},g\sim G,a_{t}\sim\pi(s_{t},g), \displaystyle s_{t+1}\sim T(s_{t},a_{t}),\forall t\in[0,H]]. \tag{1}
$$
Estimating the value of $\rho(\pi)$ from previously collected data $D$ , known as offline policy evaluation (OPE) (Levine et al., 2020), has been extensively studied (Thomas & Brunskill, 2016; Jiang & Li, 2016; Fu et al., 2021; Yang et al., 2020; Thomas et al., 2015b). However, existing work in OPE mostly focuses on simulated settings that are less practical (e.g., assumptions about full observability, access to ground truth states).
Model-Based Evaluation.
Motivated by characteristics of a real-robot system such as image based observations, high control frequencies, diverse offline data from different tasks/environments, and the lack of access to the ground truth state of the world, we consider the use of offline data to learn a single world model $\hat{T}(·|\mathbf{o},\mathbf{a})$ , where $\mathbf{o}$ represents a sequence of previous image observations and $\mathbf{a}$ represents a sequence of next actions. A sequence of next observations can be sampled from the world model $\mathbf{o^{\prime}}\sim\hat{T}(\mathbf{o},\mathbf{a})$ . With this world model, one can estimate the policy value $\rho(\pi)$ with Monte-Carlo sampling using stochastic rollouts from the policy and the world model:
$$
\displaystyle\hat{\rho}(\pi)= \displaystyle\mathbb{E}[\hat{R}([o_{0},...,o_{H}],g)|s_{0},g\sim G,\mathbf{a}\sim\pi(\mathbf{o},g), \displaystyle\mathbf{o^{\prime}}\sim\hat{T}(\mathbf{o},\mathbf{a}),\mathbf{o}=\mathbf{o^{\prime}}], \tag{2}
$$
where $\hat{R}$ is a learned reward function. Previously, model-free policy evaluation may be more preferable since in a single task setting, dynamics models are potentially harder to learn than policy values themselves, and doing rollouts in a dynamics model may lead to compounding errors (Xiao et al., 2019). However, we make the key observations that while there can be many tasks and many policies, there is only one physical world that is governed by the same set of physical laws. As a result, learning a world model can benefit from diverse data from different tasks and environments with different state spaces, goals, and reward functions. More importantly, a world model can be directly trained on image-based observations, which is often the perception modality of real-world robots.
3 Building and Evaluating the World Model
In this section, we first describe our implementation of world model training and inference. Then, we discuss how we validate our world modelâs performance prior to rolling out real robot policies within it in the next section.
3.1 Building the World Model
First, we describe the architecture and key implementation details, followed by our proposed inference scheme for policy rollouts.
3.1.1 World Model Training
We train a latent Diffusion Transformer (Peebles & Xie, 2023) on sequences of frames paired with actions, using Diffusion Forcing (Chen et al., 2024) to enable autoregressive frame generation. Per-frame robot action vectors are linearly projected to the model dimension and added elementwise to diffusion timestep embeddings, the result of which is used to condition the model through AdaLN-Zero modulation, similar to class conditioning in Peebles & Xie (2023). To ensure the world model is fully controllable by robot actions, we propose to randomly drop out actions for entire video clips, and use classifier-free guidance to improve the world modelâs adherence to action inputs. Conditioning on previous framesâ latents is achieved via causal temporal attention blocks interleaved between spatial attention blocks, as in Bruce et al. (2024); Ma et al. (2025). See Appendix A for additional implementation details.
<details>
<summary>x2.png Details</summary>

### Visual Description
## Image Comparison: Ground-Truth vs. Generated Video Frames
### Overview
The image presents a side-by-side comparison of **ground-truth video frames** (actual robotic actions) and **generated video frames** (simulated or reconstructed actions) across four distinct robotic manipulation tasks. Each task is labeled with a descriptive title, and the frames are arranged in two rows: the top row shows ground-truth, while the bottom row shows generated results.
---
### Components/Axes
1. **Task Labels**:
- **Top-Left**: "Object Manipulation (Blue Towel)"
- **Top-Right**: "Object Arrangement (Stacked Items)"
- **Bottom-Left**: "Object Interaction (Drawer and Bottle)"
- **Bottom-Right**: "Object Relocation (Blue Bowl)"
2. **Frame Structure**:
- Each task contains **4 frames** (temporal sequence).
- Frames are labeled implicitly by their position in the sequence (left to right).
3. **Visual Elements**:
- **Ground-truth**: High-resolution, realistic robotic actions.
- **Generated**: Simulated actions with noticeable discrepancies in object placement, motion, and interaction fidelity.
---
### Detailed Analysis
#### 1. Object Manipulation (Blue Towel)
- **Ground-truth**:
- A robotic arm grasps a blue towel on a wooden surface, lifts it, and folds it neatly.
- Towel remains upright during manipulation.
- **Generated**:
- Towel appears slightly misaligned in the final frame (tilted or partially unfolded).
- Arm trajectory deviates slightly from ground-truth, suggesting motion planning inaccuracies.
#### 2. Object Arrangement (Stacked Items)
- **Ground-truth**:
- Items (e.g., cans, plates) are stacked symmetrically on a dark surface.
- Robotic arm adjusts positions with precision.
- **Generated**:
- Items are misaligned (e.g., cans tilted, plates uneven).
- Arm fails to maintain consistent spacing between objects.
#### 3. Object Interaction (Drawer and Bottle)
- **Ground-truth**:
- Arm opens a wooden drawer, places a black bottle inside, and closes it.
- Drawer opens fully, bottle is centered.
- **Generated**:
- Drawer opens only partially; bottle is misplaced (off-center or tilted).
- Arm motion is jerky compared to smooth ground-truth movement.
#### 4. Object Relocation (Blue Bowl)
- **Ground-truth**:
- Arm lifts a blue bowl from a table and places it on a secondary surface.
- Bowl remains upright throughout.
- **Generated**:
- Bowl is dropped or misplaced in the final frame (e.g., tilted or on the wrong surface).
- Arm trajectory diverges significantly from ground-truth.
---
### Key Observations
1. **Task Complexity Correlation**:
- Simpler tasks (e.g., towel folding) show smaller discrepancies than complex ones (e.g., drawer interaction).
2. **Motion Fidelity**:
- Generated frames exhibit unnatural arm movements (e.g., jerky motions, incorrect trajectories).
3. **Object Placement Errors**:
- Final frames often show objects in incorrect orientations or positions (e.g., tilted bowls, misaligned stacks).
---
### Interpretation
The image highlights limitations in generated video reconstruction for robotic tasks. Discrepancies suggest challenges in:
- **Physics Simulation**: Generated frames fail to replicate realistic object dynamics (e.g., bowl dropping).
- **Motion Planning**: Arm trajectories deviate from ground-truth, indicating potential issues in inverse kinematics or sensor feedback.
- **Task-Specific Accuracy**: Complex interactions (e.g., drawer opening) are more error-prone, possibly due to higher degrees of freedom or occlusions.
These findings underscore the need for improved simulation models to bridge the gap between ground-truth and generated data in robotics training and testing.
</details>
Figure 2: Qualitative evaluation of the world model on Bridge, RT-1, VIOLA, and Berkeley UR5. In each group, top row shows the ground truth video from the real robot. Bottom row shows the generated video from the world model conditioned on the same actions as the original video. The world model closely follows the true dynamics across different robot morphologies.
3.1.2 Rolling Out a Policy in the World Model
Our policy evaluation pipeline operates through an iterative loop between the robot policy and the world model. First, the world model is initialized with an initial observation $o_{0}$ , which is then passed as input to a policy $\pi$ which produces a chunk of actions $\mathbf{a}_{\text{pred}}$ . The actions are passed back to the world model, which predicts a new frame for each action in $\mathbf{a}_{\text{pred}}$ . The latest frame produced by the world model is then returned to the policy as its next input observation.
Since different robot policies output a different number of actions at once (Kim et al., ; Brohan et al., 2022; Chi et al., 2023), WorldGym needs to support efficient prediction of a chunk of videos conditioned on a chunk of (variable length) actions. By virtue of being trained with Diffusion Forcing, as well as our usage of a causal temporal attention mask, we can flexibly control how many frames our world model denoises in parallel at inference time, i.e. its prediction horizon length. We propose setting the horizon equal to the policyâs action chunk size, $|\mathbf{a}_{\text{pred}}|$ . This has the benefit of efficient frame generation for policies with differing action chunk sizes, all from a single world model checkpoint. This contrasts with prior diffusion world models for robotics, such as Cosmos (NVIDIA et al., 2025), which, due to being trained with bidirectional attention and a fixed context length, must always denoise 16 latent frames in parallel. This constraint results in wasted compute for action chunk sizes less than the context length and unrealized parallelism for chunk sizes which are larger. On the other hand, our design allows parallelism to flexibly match the number of actions, thus utilizing hardware more effectively (see Appendix F.2).
3.1.3 VLM as Reward
We opt for GPT-4o (OpenAI et al., 2024) as a reward model, passing in the sequence of frames from the generated rollout and the language instruction (see the prompt for the VLM in Appendix B). In certain cases where both policies being evaluated fail to perform a task end-to-end, it is still helpful to get signals on which policy is closer to completing a task. We can specify these partial credit criteria to the VLM to further distinguish performance between different policies, which has been done manually using heuristics in prior work (Kim et al., ). We validate the accuracy of VLM-predicted rewards in Appendix B.2.
3.2 Evaluating the World Model
Next, we describe how we validate the performance of our world model prior to policy evaluation, ensuring that it exhibits realistic robot movement and adheres to arbitrary action controls.
<details>
<summary>x3.png Details</summary>

### Visual Description
## Diagram: Robotic Arm Movement and Gripper Control
### Overview
The image depicts a robotic arm performing controlled movements across four distinct operational modes: Gripper, X Sweep, Y Sweep, and Z Sweep. Each mode is represented by a 2x2 grid of frames showing sequential actions, with directional indicators and labels embedded in the visuals.
### Components/Axes
1. **Gripper Section**
- Labels: "close" (left frame), "open" (right frame)
- Positioning: Top-left quadrant, centered above the gripper mechanism.
- Spatial grounding: Text is horizontally aligned with the gripper's jaw position.
2. **X Sweep Section**
- Arrows: Right arrow (â) in the first frame, left arrow (â) in the second frame.
- Positioning: Bottom-left quadrant, aligned with the robotic arm's horizontal axis.
3. **Y Sweep Section**
- Arrows: Upward arrow (â) in the first frame, downward arrow (â) in the second frame.
- Positioning: Top-right quadrant, aligned with the vertical axis of the workspace.
4. **Z Sweep Section**
- Arrows: Downward arrow (â) in the first frame, upward arrow (â) in the second frame.
- Positioning: Bottom-right quadrant, aligned with the vertical axis of the robotic arm's base.
### Detailed Analysis
- **Gripper**: Explicitly labeled "close" and "open" in the top-left quadrant, indicating the robot's ability to manipulate objects via jaw movement.
- **X Sweep**: Horizontal movement (left/right) is denoted by bidirectional arrows in the bottom-left quadrant.
- **Y Sweep**: Vertical movement (up/down) is shown in the top-right quadrant with opposing arrows.
- **Z Sweep**: Vertical movement (down/up) is depicted in the bottom-right quadrant, suggesting multi-axis coordination.
### Key Observations
- The diagram emphasizes **bidirectional control** for all axes (X, Y, Z) and gripper functionality.
- Arrows are consistently placed in the first frame of each section, with opposing directions in the second frame.
- No numerical data or quantitative metrics are present; the focus is on qualitative movement directions.
### Interpretation
This diagram illustrates the **operational range and control logic** of a robotic arm in a 3D workspace. The labeled movements (X, Y, Z sweeps) and gripper actions suggest a system designed for precise object manipulation, likely for tasks like assembly, sorting, or material handling. The absence of numerical data implies the diagram prioritizes **functional demonstration** over quantitative performance metrics. The bidirectional arrows indicate the robot's ability to reverse motion, critical for dynamic task execution.
</details>
-5mm
Figure 3: Results on end-effector control across action dimensions. Generated videos closely follow the gripper controls such as open and close the gripper as well as moving in different directions starting from any initial observation frame. Results for control sweeps on the Bridge robot can be found in Figure 16 in Appendix E.1.
3.2.1 Agreement with Validation Split
First, we test the world modelâs ability to generate similar videos as running a robot in the real world. Specifically, we take the validation split of initial images from the Open-X Embodiment dataset, and predict videos conditioned on the same action sequences as in the original data. Figure 2 shows that the generated rollouts generally follow the real-robot rollouts across different initial observations and different robot morphologies.
3.2.2 End-Effector Control Sweeps
Next, we need a way to evaluate whether our world model can emulate arbitrary action sequences, beyond the kinds of action sequences present in the training data. We propose hard-coding a robot control policy by only moving one action dimension at once (and keeping the other action dimensions as zeros). The robot is then expected to move along that one action dimension with non-zero input, corresponding to moving in different horizontal and vertical directions as well as open and close its gripper. Figure 3 shows that the generated videos faithfully follow the intended end-effector movement, Results are best viewed as videos in the supplementary material. despite the fact that these particular sequences of controls are not present in the training data.
4 Evaluating Policies in WorldGym
Having established confidence in the world modelâs performance, we now use the world model to evaluate policies. We begin by rolling out three recent VLA policies in WorldGym and check whether WorldGym reflects real-world success. (Section 4.1). We then assess whether relative policy performance is preserved, comparing different versions, sizes, and training stages of the same models (Section 4.2). Finally, we explore WorldGymâs potential to test policies on out-of-distribution (OOD) tasks and environments (Section 4.3), including novel instructions and altered visual contexts.
4.1 Correlation between Real-World and Simulated Policy Performance
<details>
<summary>x4.png Details</summary>

### Visual Description
## Scatter Plot: Per-Task Success Rates: Real World vs World Model
### Overview
The image is a scatter plot comparing real-world success rates (x-axis) to world model success rates (y-axis) across three AI systems: RT-1-X, Octo, and OpenVLA. A dashed trend line (labeled "Fit") shows a strong positive correlation (r = 0.78, p < 0.001), indicating that higher real-world success rates generally correspond to higher model success rates.
### Components/Axes
- **X-axis**: Real World Success Rate (%)
- Scale: 0% (left) to 100% (right)
- Labels: Discrete ticks at 0, 20, 40, 60, 80, 100
- **Y-axis**: World Model Success Rate (%)
- Scale: 0% (bottom) to 100% (top)
- Labels: Discrete ticks at 0, 20, 40, 60, 80, 100
- **Legend**: Located in the bottom-right corner
- RT-1-X: Blue circles
- Octo: Orange squares
- OpenVLA: Red triangles
- **Trend Line**: Dashed black line labeled "Fit"
- Equation: Not explicitly provided
- Correlation: r = 0.78 (strong positive relationship)
- Significance: p < 0.001 (statistically significant)
### Detailed Analysis
1. **RT-1-X (Blue Circles)**
- **Positioning**: Clustered in the lower-left quadrant (real-world success: 0â40%, model success: 0â30%).
- **Trend**: Data points generally align below the trend line, suggesting underperformance relative to real-world success.
- **Outliers**: One point at (60%, 50%) deviates slightly above the trend line.
2. **Octo (Orange Squares)**
- **Positioning**: Spread across the plot, with concentrations near (20â60% real-world, 10â50% model success).
- **Trend**: Mixed alignment with the trend line; some points above (e.g., (40%, 40%)) and below (e.g., (20%, 10%)).
3. **OpenVLA (Red Triangles)**
- **Positioning**: Dominates the upper-right quadrant (real-world success: 60â100%, model success: 60â100%).
- **Trend**: Most points align closely with or above the trend line, indicating strong performance relative to real-world success.
- **Outliers**: One point at (80%, 70%) falls slightly below the trend line.
### Key Observations
- **Strong Correlation**: The trend line (r = 0.78) confirms a robust relationship between real-world and model success rates.
- **Model Performance**:
- OpenVLA consistently outperforms expectations (above the trend line).
- RT-1-X underperforms relative to real-world success (below the trend line).
- Octo shows moderate alignment with the trend line but higher variability.
- **Statistical Significance**: The p-value (< 0.001) rules out random chance as the cause of the correlation.
### Interpretation
The data demonstrates that world models trained on real-world data (e.g., OpenVLA) achieve higher success rates in tasks where real-world performance is high. RT-1-X, by contrast, struggles to match real-world outcomes, suggesting limitations in its training or architecture. The trend lineâs slope implies that improving real-world success rates could directly enhance model performance, but the spread in data points (especially for Octo and OpenVLA) highlights task-specific variability. OpenVLAâs outlier at (80%, 70%) may indicate a task where real-world success is high but model performance lags, warranting further investigation. Overall, the plot underscores the importance of real-world data in training effective world models.
</details>
(a) Per-Task Task Success Rates. Each point represents a task from Table 5, with different policies being represented by different shaped markers. There is a strong correlation ( $r=0.78$ ) between policy performance in our world model (y-axis) and within the real world (x-axis).
<details>
<summary>x5.png Details</summary>

### Visual Description
## Bar Chart: Mean Success Rates: Real World vs World Model
### Overview
The chart compares mean success rates (%) between "Real World" and "World Model" across three categories: RT-1-X, Octo, and OpenVLA. Error bars indicate variability, with OpenVLA showing the largest uncertainty.
### Components/Axes
- **X-axis**: Categories (RT-1-X, Octo, OpenVLA)
- **Y-axis**: Success Rate (%) from 0 to 80 (increments of 10)
- **Legend**:
- Light blue = Real World
- Dark blue = World Model
- **Error Bars**: Vertical lines extending from each barâs midpoint
### Detailed Analysis
1. **RT-1-X**:
- Real World: 18.5% (±3.2%)
- World Model: 15.6% (±2.8%)
2. **Octo**:
- Real World: 20.0% (±4.1%)
- World Model: 23.8% (±3.5%)
3. **OpenVLA**:
- Real World: 70.6% (±6.3%)
- World Model: 67.4% (±5.9%)
### Key Observations
- **Real World** consistently outperforms **World Model** in all categories.
- **OpenVLA** has the highest success rates but also the largest error margins.
- **RT-1-X** shows the lowest success rates and smallest error bars.
### Interpretation
The data suggests that "Real World" models achieve higher success rates than "World Model" counterparts across all tasks. OpenVLA demonstrates the greatest performance gap (3.2% difference) but also the highest variability, potentially indicating complex or edge-case scenarios. RT-1-Xâs lower performance and smaller error bars may reflect simpler tasks or more stable model behavior. The error bars for OpenVLA suggest significant uncertainty, possibly due to smaller sample sizes or diverse real-world conditions.
</details>
(b) Mean Success Rates. Robot policiesâ mean success rates in the world model differ by an average of only 3.3% between from the real world, near the standard error range for each policy. Relative performance rankings between RT-1-X, Octo, and OpenVLA are also preserved.
Figure 4: Success rates of modern VLAs, as evaluated within WorldGym and the real world.
Qualitative Evaluation.
To ensure WorldGym is useful for policy evaluation, we test whether policy performance within the world model is similar to that of the real world. To do so, we perform a direct comparison with the Bridge evaluation trials from OpenVLA (Kim et al., ). Specifically, the OpenVLA Bridge evaluation consists of 17 challenging tasks which are not present in the Bridge V2 (Walke et al., 2023) dataset. We use WorldGym to evaluate the three open-source policies evaluated in Kim et al. : RT-1-X (OâNeill et al., 2023), Octo (Octo Model Team et al., 2024), and OpenVLA (Kim et al., ). For each task and each policy, Kim et al. perform 10 trials, each with randomized initial object locations. We obtain the first frame of the recorded rollouts for all trials of all tasks. We then simulate each of the 10 real-world trials by using the original initial frame to roll out the policy within the world model as described in Section 3.1.2. We show qualitative rollouts in WorldGym from different policies in Figure 5, which shows that rollouts from OpenVLA generally perform better than rollouts from RT-1-X and Octo on the Bridge robot (top two rows). We further show that WorldGym can be easily used to perform rollouts in other environments with other robots, such as the Google Robot (bottom row in Figure 5).
<details>
<summary>x6.png Details</summary>

### Visual Description
## Photo Comparison: Robotic Manipulation Task Performance
### Overview
The image presents a side-by-side comparison of robotic manipulation tasks across three methods: **OpenVLA**, **Octo**, and **RT-1-X**. Three distinct tasks are evaluated:
1. **Put eggplant into pot**
2. **Put corn on plate**
3. **Pick blue slip bag**
Each method is visualized through a sequence of frames showing the robot's actions, with the **Initial Frame** provided for context.
---
### Components/Axes
- **Tasks (Rows)**:
- Row 1: "Put eggplant into pot"
- Row 2: "Put corn on plate"
- Row 3: "Pick blue slip bag"
- **Methods (Columns)**:
- Column 1: **Initial Frame** (baseline setup)
- Column 2: **OpenVLA**
- Column 3: **Octo**
- Column 4: **RT-1-X**
- **Visual Elements**:
- Robot arms (black/gray) interacting with colored objects (e.g., purple eggplant, red/yellow corn, blue slip bag).
- Backgrounds vary slightly (e.g., blue tray, wooden surface).
---
### Detailed Analysis
#### Task 1: "Put eggplant into pot"
- **OpenVLA**: Smooth, controlled motion; eggplant is placed precisely into the pot.
- **Octo**: Slightly erratic movement; eggplant is misaligned but still in the pot.
- **RT-1-X**: Overly aggressive motion; eggplant is dropped outside the pot.
#### Task 2: "Put corn on plate"
- **OpenVLA**: Accurate placement; corn is centered on the plate.
- **Octo**: Corn is placed off-center but still on the plate.
- **RT-1-X**: Corn is knocked off the plate entirely.
#### Task 3: "Pick blue slip bag"
- **OpenVLA**: Precise grip; bag is lifted cleanly.
- **Octo**: Bag is partially grasped, causing instability.
- **RT-1-X**: Bag is crushed or dropped during the attempt.
---
### Key Observations
1. **OpenVLA** consistently demonstrates the highest precision and control across all tasks.
2. **Octo** shows moderate performance, with minor errors in alignment or stability.
3. **RT-1-X** exhibits the poorest performance, with frequent failures in object handling.
4. No numerical data or legends are present; analysis is based on visual inspection of motion and object placement.
---
### Interpretation
The comparison highlights **OpenVLA** as the most reliable method for robotic manipulation, likely due to its advanced planning or control algorithms. **Octo** performs adequately but lags behind in precision, while **RT-1-X** struggles with basic tasks, suggesting potential issues in its motion execution or object recognition. The absence of quantitative metrics (e.g., success rates, error margins) limits deeper analysis, but the visual trends strongly favor OpenVLA for practical applications.
</details>
Figure 5: Qualitative policy rollouts on Bridge and Google Robot for RT-1-X, Octo, and OpenVLA. OpenVLA rollouts often lead to more visual successes than the other two policies across environments.
Quantitative Evaluation.
Using the simulated rollouts from WorldGym, we then compute the average task success rate similar to Kim et al. , and plot the success rate for each task for each policy in Figure 4(a). We find that real-world task performance is strongly correlated with the task performance reported by the world model, achieving a Pearson correlation of $r=0.78$ . While per-task policy success rates within WorldGym still differ slightly from those in the real world (see Table 5), the mean success rates achieved by these policies within WorldGym are quite close to the their real-world values, as shown in Figure 4(b). The success rates differ by an average of only 3.3%, with RT-1-X achieving 18.5% in the real world vs 15.5% in the world model, Octo achieving 20.0% vs 23.82%, and OpenVLA achieving 70.6% vs 67.4%, respectively. See quantitative results of evaluating the three policies on the Google Robot in Appendix E.2
4.2 Policy Ranking within a World Model
<details>
<summary>x7.png Details</summary>

### Visual Description
## Bar Chart: Mean Success Rates Across Different Model Versions
### Overview
The chart compares the mean success rates of four model versions using vertical bars with error bars. Success rates are plotted on a percentage scale (0â70%) against model versions on the x-axis. The highest success rate is observed in the "OpenVLA 7B" model, while the lowest is in "Octo Small 1.5."
### Components/Axes
- **X-axis**: Model versions labeled as:
- Octo Small 1.5 (blue)
- Octo Base 1.5 (orange)
- OpenVLA v0.1 7B (green)
- OpenVLA 7B (red)
- **Y-axis**: Success Rate (%) with increments of 10%.
- **Error Bars**: Represent variability in success rates, with approximate ranges:
- Octo Small 1.5: ±2.5%
- Octo Base 1.5: ±3.0%
- OpenVLA v0.1 7B: ±3.5%
- OpenVLA 7B: ±5.0%
### Detailed Analysis
- **Octo Small 1.5**: 21.5% success rate (blue bar, ±2.5% error).
- **Octo Base 1.5**: 23.8% success rate (orange bar, ±3.0% error).
- **OpenVLA v0.1 7B**: 27.6% success rate (green bar, ±3.5% error).
- **OpenVLA 7B**: 67.4% success rate (red bar, ±5.0% error).
### Key Observations
1. **Performance Gap**: OpenVLA 7B outperforms all other models by a significant margin (67.4% vs. 27.6% for the next highest).
2. **Error Trends**: Larger models (OpenVLA 7B) exhibit greater variability in success rates (±5.0%) compared to smaller models (±2.5â3.5%).
3. **Progression**: Success rates increase steadily from Octo Small 1.5 to OpenVLA 7B, suggesting architectural or scaling improvements.
### Interpretation
The data demonstrates a clear trend where larger, more advanced models (e.g., OpenVLA 7B) achieve higher success rates, likely due to enhanced capacity or training. However, the increased error margin for OpenVLA 7B implies greater sensitivity to input variability or environmental factors. The Octo models show minimal improvement between versions (21.5% to 23.8%), indicating limited gains from incremental updates. This chart highlights the trade-off between model complexity and reliability, with advanced models offering higher performance at the cost of consistency.
</details>
Figure 6: Success Rates of different model versions in WorldGym. We evaluate different generations of Octo and OpenVLA in the world model, showing that WorldGym assigns higher score to larger and more recent versions.
<details>
<summary>x8.png Details</summary>

### Visual Description
## Line Chart: Mean Success Rate Across Checkpoints
### Overview
The chart compares the mean success rates of two policies ("Video Policy" and "Diffusion Policy") across training checkpoints (0 to 60k steps). Success rate is measured in percentage, with confidence intervals shaded around each line.
### Components/Axes
- **X-axis**: Checkpoint (training steps) labeled at 0, 10k, 20k, 40k, and 60k.
- **Y-axis**: Success Rate (%) ranging from 0 to 35% in 5% increments.
- **Legend**: Located in the top-left corner, with:
- **Blue line with circles**: Video Policy
- **Orange line with squares**: Diffusion Policy
### Detailed Analysis
#### Video Policy (Blue)
- **Trend**: Steady upward trajectory with a sharp increase between 10k and 40k checkpoints.
- **Data Points**:
- 0k: ~19% (confidence interval: 18â20%)
- 10k: ~20% (confidence interval: 19â21%)
- 20k: ~26% (confidence interval: 24â28%)
- 40k: ~29% (confidence interval: 27â31%)
#### Diffusion Policy (Orange)
- **Trend**: Gradual rise with fluctuations, followed by a sharp increase after 40k.
- **Data Points**:
- 10k: ~4% (confidence interval: 3â5%)
- 20k: ~10% (confidence interval: 9â11%)
- 40k: ~8% (confidence interval: 7â9%)
- 60k: ~15% (confidence interval: 14â16%)
### Key Observations
1. **Video Policy Dominance**: Consistently outperforms Diffusion Policy across all checkpoints, especially after 20k steps.
2. **Diffusion Policy Volatility**: Success rate fluctuates significantly (e.g., drops from 10% at 20k to 8% at 40k) before recovering.
3. **Confidence Intervals**: Video Policyâs wider shaded area suggests higher variability in early checkpoints, narrowing as training progresses.
### Interpretation
- **Performance Dynamics**: Video Policy demonstrates robust learning efficiency, achieving ~29% success by 40k steps. Diffusion Policy lags initially but shows potential for improvement, reaching ~15% by 60k steps.
- **Stability vs. Exploration**: The narrowing confidence intervals for Video Policy imply stabilizing performance, while Diffusion Policyâs wider intervals at later stages suggest ongoing exploration or instability.
- **Practical Implications**: Video Policy may be preferable for applications requiring early convergence, whereas Diffusion Policy might benefit from extended training to mitigate volatility.
*Note: All values are approximate, derived from visual inspection of the chart.*
</details>
Figure 7: Success Rate within WorldGym throughout training. We train a video-based policy and a diffusion policy from scratch and evaluate it within our world model as it trains. We see that mean task success rate within the world model increases with additional training steps.
Now we test whether WorldGym can preserve policy rankings known a priori. We evaluated policies across different versions, sizes, and training stages within WorldGym on the OpenVLA Bridge evaluation task suite, and found their in-world-model performance rankings to be consistent with prior knowledge of their relative performance.
Different VLAs with Known Ranking. First, we average success rates across all 17 tasks and find that the relative performance rankings between RT-1-X, Octo, and OpenVLA are the same (Figure 4(b)) within both WorldGym and the real-world results reported in OpenVLA (Kim et al., ).
Same Policies across Versions and Sizes. We further examine whether WorldGym preserves rankings between different versions and sizes of the same policy. In particular, we compare Octo-Small 1.5 against Octo-Base 1.5, and OpenVLA v0.1 7B, an undertrained development model, against OpenVLA 7B. As shown in Figure 7, the larger and more recent models outperform their smaller or earlier counterparts within WorldGym, consistent with the findings of real-world experiments performed in Octo Model Team et al. (2024) and Kim et al. . This provides additional evidence that WorldGym faithfully maintains relative rankings even across model upgrades.
Same Policy across Training Steps. To examine whether WorldGym provides meaningful signals for policy training, hyperparameter tuning, and checkpoint selections, we train two robot policies from scratch. Building on prior evidence of WorldGymâs effectiveness in evaluating VLA-based policies, we extend our study to two additional families: a video predictionâbased policy (UniPi) (Du et al., 2023a) and a diffusion-based policy (DexVLA) (Wen et al., 2025), both trained on the Bridge V2 dataset (see Appendix C and Appendix D). We evaluate checkpoints of the video prediction policy at 2K, 8K, 12K, and 18K steps, and the diffusion policy at 10K, 20K, 40K, and 60K steps.
As shown in Figure 7, WorldGym tends to assign higher success rates to checkpoints as they increase in training steps, consistent with the lower mean squared error these policies achieve on their validation splits. This demonstrates WorldGymâs ability to preserve policy rankings across models with different amounts of training compute.
Thus, we have shown how WorldGym can be used to obtain reasonable policy rankings. In particular, for the VLA-based policies we evaluate, we arrive at the same conclusions as real-world experiments about their relative performances. Notably, this is achieved all without the manual effort of setting up real robot evaluation environments and monitoring policy rollouts. While real-world evaluation can sometimes take days to complete, all WorldGym rollouts reported here can be completed in under an hour on a single GPU and require only initial images for each trial.
4.3 Out-of-Distribution Inputs
In this section, use WorldGym to explore policiesâ performance on both OOD input images and OOD language instructions.
<details>
<summary>x9.png Details</summary>

### Visual Description
## Screenshot: Robotic Arm Task Execution
### Overview
The image depicts a robotic arm interacting with colored squares (red and blue) on a wooden surface. The scene is divided into four quadrants, each showing the robot arm in different positions. Text labels ("Pick red" and "Pick blue") are overlaid in the top-left corners of the top-left and bottom-left quadrants, respectively.
### Components/Axes
- **Labels**:
- "Pick red" (top-left quadrant, white text on gray background)
- "Pick blue" (bottom-left quadrant, white text on gray background)
- **Objects**:
- Robotic arm (black, with visible joints and grippers)
- Red square (bottom-right quadrant, solid red)
- Blue square (top-left quadrant, solid blue)
- **Background**:
- Wooden surface (light brown, textured)
- Gray wall (textured, with a red "L"-shaped object in the top-right corner)
### Detailed Analysis
1. **Top-Left Quadrant**:
- Label: "Pick red" (positioned at top-left corner).
- Robotic arm is positioned above the blue square, not the red square.
- Blue square is stationary; red square is in the bottom-right quadrant.
2. **Bottom-Left Quadrant**:
- Label: "Pick blue" (positioned at top-left corner).
- Robotic arm is positioned above the blue square, aligning with the label.
- Blue square is stationary; red square is in the bottom-right quadrant.
3. **Top-Right Quadrant**:
- No label.
- Robotic arm is positioned above the red square, but no explicit instruction is provided.
4. **Bottom-Right Quadrant**:
- No label.
- Robotic arm is positioned above the red square, but no explicit instruction is provided.
### Key Observations
- The labels "Pick red" and "Pick blue" are spatially isolated to their respective quadrants but do not align with the robot arm's position in those quadrants.
- The robot arm's movement suggests it is executing tasks based on the labels, but the alignment between labels and actions is inconsistent (e.g., "Pick red" is in the top-left quadrant, but the robot is above the blue square there).
- The red and blue squares are fixed in their positions across all quadrants, indicating a static environment.
### Interpretation
The image demonstrates a robotic arm performing color-based tasks, but the labels and robot positions are misaligned. This could imply:
1. **Conditional Logic**: The robot may be testing or simulating task execution based on labels, even if the physical position does not match the instruction.
2. **Error in Task Sequencing**: The labels might be placeholders for a larger system where the robot dynamically adjusts its target based on real-time input.
3. **Visualization of Decision-Making**: The labels could represent hypothetical scenarios, with the robot's position reflecting its "decision" to pick a color despite the label's location.
The absence of additional text or data suggests this is a simplified demonstration of robotic task execution, emphasizing color recognition and movement control.
</details>
Figure 8: OOD: Color Classification. We add red and blue pieces of paper to a table, and ask the policies to âpick redâ or âpick blueâ (OOD image and language). OpenVLA excels, picking the correct colored paper in all trials, whereas all other policies score near chance.
-5mm
OOD Image Input. Using modern image generation models like Nano Banana (Google, 2025), we can easily generate new input images to initialize our world model with. We evaluate robot policies under three OOD settings: unseen object interaction, distractor objects, and object classification (see detailed results in Table 6).
<details>
<summary>x10.png Details</summary>

### Visual Description
## Diagram: Robotic Task Execution System
### Overview
The diagram illustrates a robotic system performing three distinct tasks in a simulated kitchen environment. It demonstrates the integration of image processing models with robotic policy instructions to manipulate objects (orange, carrot, plate) within a sink setup. The system uses color-coded prompts to differentiate between image edit instructions and robot policy commands.
### Components/Axes
1. **Legend** (left side):
- **Image Edit Prompt** (yellow): Represents task-specific instructions for image manipulation
- **Robot Policy Instruction** (gray): Indicates the robotic action to be executed
2. **Task Structure** (vertical flow):
- Each task (a, b, c) contains:
- Input image of the sink environment
- Image model processing
- Output robot policy instruction
- Execution result image
3. **Spatial Layout**:
- Legend positioned in bottom-left quadrant
- Task components arranged in three vertical columns
- Robot arm shown in action across all task executions
### Detailed Analysis
1. **Task (a): Add an orange**
- Image Model Input: Sink with plate and carrot
- Robot Policy: "Put orange on plate"
- Execution: Orange placed in dish rack
2. **Task (b): Swap carrot and orange**
- Image Model Input: Sink with orange and carrot
- Robot Policy: "Put orange on plate"
- Execution: Orange moved to dish rack, carrot placed in sink
3. **Task (c): Turn the carrot red**
- Image Model Input: Sink with orange and carrot
- Robot Policy: "Put orange on plate"
- Execution: Carrot color changed to red while maintaining position
### Key Observations
- All tasks share the same robot policy instruction ("Put orange on plate") despite different objectives
- Color coding strictly follows legend: yellow for image prompts, gray for robot policies
- Robot arm maintains consistent positioning across all task executions
- Object positions change according to task requirements while maintaining spatial relationships
### Interpretation
This diagram demonstrates a vision-guided robotic system capable of:
1. Interpreting task descriptions through image models
2. Generating appropriate robotic actions based on environmental context
3. Executing complex object manipulations (addition, swapping, color modification)
The system shows:
- **Task-Image-Execution Correlation**: Each task's image model output directly informs the robot's physical action
- **Color-Coded Workflow**: Visual distinction between planning (yellow) and execution (gray) phases
- **Environmental Adaptation**: Maintains object relationships while performing specific manipulations
Notable patterns include the consistent use of the same robot policy instruction across different tasks, suggesting either a limitation in the system's task interpretation or a deliberate design choice for demonstration purposes. The color modification task (c) implies potential integration with computer vision systems capable of object recognition and property alteration.
</details>
Figure 9: OOD: Unseen object. We use Nano Banana (Google, 2025) to add an orange to the world modelâs initial frame. When both the orange and the carrot are present, (a-b) OpenVLA grabs whichever is closer. After (c) editing the carrotâs color to red, however, the orange is correctly picked up.
Figure 10: OOD: Failure modes. Left: We add a laptop to the scene, which displays an image of a carrot. In 15% of trials, OpenVLA grabs the laptop instead of the real carrot. Right: We test the ability distinguish to between squares and circles, celebrity faces, and cats and dogs, with all policies scoring near-chance.
- Unseen Objects: We edit a scene to contain both a carrot and an orange, asking the policy to pick up the orange (Figure 10). OpenVLA grabs whichever object is closer until we edit the carrotâs color to be red, after which it always grabs the orange correctly. This suggests that it struggles to distinguish carrots and oranges by their shape.
- Distractor Objects: We use the image editing model to add a computer displaying an image of a carrot (Figure 10, left). We see that OpenVLA mistakenly to grabs the carrot on the computer screen in 15% of trials, suggesting limited 3D/2D object distinction.
- Classification: We add a piece of paper on each side of a desk. We first color one paper red and the other blue and instruct the model to âpick redâ/âpick blueâ (Figure 8). OpenVLA achieves a perfect score, always moving towards the correct color. Octo and RT-1, on the other hand, typically move towards whichever paper is closer, scoring no better than chance. We also try more advanced classification tasks, (Figure 10, right), but find that the policies all score near-chance.
For a more quantitative study, we modify all the initial frames of the OpenVLAâs Bridge evaluation task suite to include random OOD distractor items (see Figure 12), keeping the language instructions the same. We then repeat the rollout procedure from Section 4.1 in order to measure the degree to which the addition of unrelated objects affects policy performance. We find that all the tested VLAs degrade in performance, with OpenVLA being the most robust of the three (Figure 13).
<details>
<summary>x12.png Details</summary>

### Visual Description
## Screenshot: Robotic Arm Task Execution Sequence
### Overview
The image displays six sequential panels demonstrating a robotic arm performing household tasks in a simulated kitchen environment. Each panel includes a caption describing the action, followed by three sub-images showing the robot's motion and object manipulation.
### Components/Axes
- **Panels**: Four main panels, each with a caption and three sub-images.
1. **Panel 1**: "Put yellow corn in red cup"
2. **Panel 2**: "Put plate on drying rack"
3. **Panel 3**: "Put yellow corn in red cup" (repeated task)
4. **Panel 4**: "Move the pot to the counter"
- **Objects**:
- Yellow corn, red cup, purple/blue/green objects (possibly toys or containers), orange carrot, white plate, silver pot, drying rack, counter.
- **Robot Arm**: Black mechanical arm with a gripper, interacting with objects in each panel.
### Detailed Analysis
#### Panel 1: "Put yellow corn in red cup"
- **Sub-images**:
1. Robot arm approaches yellow corn and red cup.
2. Gripper picks up yellow corn.
3. Corn is placed into the red cup.
- **Key Details**:
- Yellow corn is initially on the counter.
- Red cup is positioned near the sink.
- Robot successfully transfers the corn.
#### Panel 2: "Put plate on drying rack"
- **Sub-images**:
1. Robot arm picks up the white plate from the sink.
2. Plate is moved toward the drying rack.
3. Plate is placed on the drying rack.
- **Key Details**:
- Plate is initially in the sink.
- Drying rack is located to the right of the sink.
#### Panel 3: "Put yellow corn in red cup" (repeated)
- **Sub-images**:
1. Robot arm picks up yellow corn from the sink (now empty).
2. Corn is transferred to the red cup.
3. Final placement in the red cup.
- **Key Details**:
- Yellow corn is now in the sink (from prior task).
- Red cup remains in the same position.
#### Panel 4: "Move the pot to the counter"
- **Sub-images**:
1. Robot arm lifts the silver pot from the sink.
2. Pot is moved horizontally across the counter.
3. Pot is placed on the counter near the drying rack.
- **Key Details**:
- Pot is initially in the sink.
- Counter is the workspace area.
### Key Observations
1. **Repetition**: The task "Put yellow corn in red cup" appears twice, suggesting a test of consistency or error handling.
2. **Object Relocation**: Objects (corn, plate, pot) are moved between the sink, counter, and drying rack, indicating spatial reasoning.
3. **Gripper Precision**: The robotâs gripper successfully handles objects of varying shapes (corn, plate, pot).
### Interpretation
The sequence demonstrates a robotic systemâs ability to execute multi-step household tasks with precision. The repetition of the corn-cup task may test reliability or adaptability to minor environmental changes (e.g., corn location). The use of a drying rack and counter highlights the robotâs integration into a functional kitchen workflow. No numerical data or anomalies are visible; the focus is on task completion accuracy.
</details>
Figure 11: OOD Language Instructions. We pick a set of tasks from the OpenVLA Bridge evaluation suite and modify the language instruction, e.g. changing the the target object and/or its goal destination.
OOD Language. Additionally, even without access to an image editing model, we demonstrate that WorldGym can be used to evaluate policiesâ performance on OOD language instructions. Starting from a set of initial frames from the tasks listed in Table 5, we modify each taskâs language instruction, e.g. changing the target object and/or its goal state. Figure 11 shows rollouts from OpenVLA for these OOD language tasks. We can then easily obtain success rates for these unseen tasks by rolling them out within WorldGym, finding that OpenVLA generalizes best (see Table 1). Policies struggle across the board on the âMove the pot to the counterâ task, with only OpenVLA achieving a single success. We suspect that OpenVLA consistently outperforms Octo and RT-1-X on OOD language tasks due to its strong VLM backbone and richer robot pretraining dataset (Kim et al., ).
| Task | RT-1-X | Octo | OpenVLA |
| --- | --- | --- | --- |
| Move Pot Into Drying Rack | 3 | 0 | 7 |
| Move The Pot To The Counter | 0 | 0 | 1 |
| Put Plate On Drying Rack | 4 | 2 | 8 |
| Put Yellow Corn In Red Cup | 1 | 2 | 3 |
Table 1: Policy Evaluations Results on Bridge OOD Language Tasks. âMove the pot to the counterâ is perhaps the most challenging because the Bridge dataset does not contain trajectories which move objects outside of the sink basin. OpenVLA has the strongest performance, which we attribute to its more powerful language model backbone.
<details>
<summary>x13.png Details</summary>

### Visual Description
## Diagram: Robotic Arm Interaction with Distracted Sink Environment
### Overview
The image depicts a robotic arm interacting with a simulated sink environment containing various objects. The scene is divided into two stages:
1. **Initial State**: A sink with a stainless steel pot, yellow dish rack, and gray faucet.
2. **Distracted State**: Additional objects (toys, cups, blocks) are added to the sink, and the robotic arm is shown interacting with the environment.
### Components/Axes
- **Sink**: Light blue basin with a yellow dish rack (left side) and gray faucet (right side).
- **Robotic Arm**: Black mechanical arm with a two-pronged gripper, positioned above the sink.
- **Objects**:
- Stainless steel pot (tilted in initial state).
- Yellow rubber duck, red toy car, purple toy, green block, red cup, blue toy car, crumpled white cloth.
- **Text Labels**:
- "add distractions" (yellow box, top-left of first two panels).
- "Image Model" (white box, bottom-center of first two panels).
### Detailed Analysis
1. **Initial State (Left Panels)**:
- Sink contains a stainless steel pot, yellow dish rack, and gray faucet.
- No distractions present.
- Robotic arm is absent in the first panel but appears in the second panel above the sink.
2. **Distracted State (Right Panels)**:
- Sink cluttered with:
- Yellow rubber duck (center-left).
- Red toy car (center-right).
- Purple toy (dish rack).
- Green block (near faucet).
- Red cup (near faucet).
- Blue toy car (bottom-right).
- Crumpled white cloth (bottom-right).
- Robotic arm actively interacts with the environment (gripper positioned near objects).
3. **Image Model Processing**:
- Arrows indicate the flow from the "add distractions" stage to the "Image Model" analysis.
- The model processes the cluttered scene, suggesting a focus on object recognition or manipulation under distraction.
### Key Observations
- **Distraction Complexity**: The addition of 7 objects (toys, cups, blocks) increases environmental complexity.
- **Robotic Arm Positioning**: The armâs gripper is centrally located in the distracted state, implying active engagement with objects.
- **Object Distribution**: Objects are scattered unevenly, with higher density near the faucet and bottom-right corner.
### Interpretation
This setup simulates a real-world scenario where a robotic system must operate in cluttered environments. The "Image Model" likely evaluates the armâs ability to:
1. Identify target objects (e.g., the stainless steel pot) amid distractions.
2. Adjust gripper positioning for precise manipulation.
3. Handle occlusions caused by overlapping objects.
The progression from a clean to a cluttered sink tests the systemâs robustness to visual noise. The presence of diverse object shapes (cylindrical pot, spherical duck, angular blocks) suggests a focus on generalizability in object recognition. The crumpled cloth introduces texture variability, further challenging the model.
No numerical data or quantitative metrics are provided in the image. The diagram emphasizes qualitative analysis of robotic perception and interaction in dynamic environments.
</details>
Figure 12: OOD Distraction Examples. We use Nano Banana (Google, 2025) to add distractions to every image of the OpenVLA Bridge task suite. The resulting change in mean success rates can be seen in Figure 13.
<details>
<summary>x14.png Details</summary>

### Visual Description
## Bar Chart: Effect of OOD Distractors on Success Rates
### Overview
The chart compares success rates (%) of two model variants ("World Model" and "World Model (with OOD input image)") across three categories: RT-1-X, Octo, and OpenVLA. Success rates are visualized as grouped bars with error bars, and the chart emphasizes the impact of OOD distractors on performance.
### Components/Axes
- **X-axis**: Categories (RT-1-X, Octo, OpenVLA)
- **Y-axis**: Success Rate (%) ranging from 0 to 70%
- **Legend**:
- Solid blue: World Model
- Striped red: World Model (with OOD input image)
- **Error Bars**: Present on all bars, indicating variability (exact error values not labeled).
### Detailed Analysis
1. **RT-1-X**:
- World Model: 15.6% (solid blue bar)
- World Model (with OOD input image): 7.6% (striped red bar)
- **Trend**: Success rate decreases by ~53% when OOD input is added.
2. **Octo**:
- World Model: 23.8% (solid blue bar)
- World Model (with OOD input image): 4.1% (striped red bar)
- **Trend**: Success rate drops by ~83% with OOD input.
3. **OpenVLA**:
- World Model: 67.4% (solid blue bar)
- World Model (with OOD input image): 39.4% (striped red bar)
- **Trend**: Success rate decreases by ~42% with OOD input.
### Key Observations
- **Inverse Relationship**: In all categories, adding OOD input images reduces success rates.
- **Magnitude of Impact**:
- Octo shows the steepest decline (~83%).
- RT-1-X has a moderate decline (~53%).
- OpenVLA has the smallest decline (~42%).
- **Error Bars**: Visually, error margins are largest for OpenVLA (World Model) and smallest for Octo (World Model with OOD input image).
### Interpretation
The data suggests that OOD distractors consistently degrade model performance across all categories. However, the severity of this degradation varies:
- **Octo** is most vulnerable to OOD input, with near-collapse in success rates.
- **OpenVLA** retains higher absolute success rates even after OOD input is introduced, indicating better robustness.
- The inverse correlation implies that OOD distractors act as significant noise sources, particularly in models like Octo that lack adaptive mechanisms to handle such inputs.
The chart highlights the need for OOD-aware training strategies to mitigate these performance drops, especially for models deployed in environments with unpredictable inputs.
</details>
Figure 13: Effect of OOD Distractors. We use an image editing model to add distractor objects to the Bridge evaluation suite, finding that RT-1-X drops in performance by 51%, Octo by 83%, and OpenVLA by 41.5%, making OpenVLA the most robust to distractors. See Table 7 for details.
The ability to use WorldGym to quickly design and evaluate policies within OOD tasks and environments thus leads us to new findings about policiesâ strengths and weaknesses. Future research could be prioritized to address these issues, all without spending extra effort to set up additional experiments in the real world or within handcrafted simulators.
5 Related Work
Action-Conditioned Video Generation.
Previous work has shown that video generation can simulate real-world interactions (Yang et al., 2023; Brooks et al., 2024), robotic plans (Du et al., 2024; 2023b), and games (AI et al., 2024; Bruce et al., 2024; Valevski et al., 2024; Alonso et al., 2024) when conditioned on text or keyboard controls. Prior work (NVIDIA et al., 2025) has begun to explore applying video generation to simulating complex robotic controls. We take this a step further by using video-based world models to quantitatively estimate robot policy success rates. WorldGym draws architectural inspirations from prior work on video generation such as Diffusion Forcing (Chen et al., 2024) and Diffusion Transformers (Peebles & Xie, 2023), but experiments with variable horizon lengths to support efficient long-horizon rollouts for policies with a variety of action chunk sizes.
Policy Evaluation.
Off-policy and offline policy evaluation has long been studied in the RL literature (Farajtabar et al., 2018; Jiang & Li, 2015; Kallus & Uehara, 2019; Munos et al., 2016; Precup et al., 2000; Thomas et al., 2015a). Some of these approaches are model-based, learning a dynamics model from previously collected data and rolling out the learned dynamics model for policy evaluation (Fonteneau et al., 2013; Zhang et al., 2021; Yu et al., 2020; Hafner et al., 2020). Since learning a dynamics model is challenging and subject to accumulation of error, a broader set of work has focused on model-free policy evaluation, which works by estimating the value function (Le et al., 2019; Duan & Wang, 2020; Sutton et al., 2009; 2016) or policy correction (Kanamori et al., 2009; Nguyen et al., 2010; Nachum et al., 2019). WorldGym performs model-based policy evaluation, but proposes to learn a single world model on image-based observation that can be used to evaluate different policies on different tasks. SIMPLER (Li et al., 2024) aims to evaluate realistic policies by constructing software-based simulators from natural images and showed highly correlated curves between simulated evaluation and real-robot execution, but it is hard to evaluate OOD language and image input in SIMPLER without significant hand engineering of the software simulator. Li et al. (2025) proposes to evaluate robot policies in a world model in a specific bi-manual manipulation setup, whereas WorldGym focuses on evaluating policies across diverse environments and robot morphologies while enabling testing OOD language and image inputs.
6 Conclusion
We have presented WorldGym, a world-model-based environment for evaluating robot policies. WorldGym emulates realistic robot interactions and shows strong correlations between simulated evaluation and real-world policy outcomes. WorldGym further provides the flexibility for evaluating OOD language instructions and performing tasks with an OOD initial frame. While not all interactions emulated by WorldGym are fully realistic, WorldGym serves as an important step towards safe and reproducible policy evaluation before deployment.
Acknowledgments
We thank Xinchen Yan and Doina Precup for reviewing versions of this manuscript. We thank Moo Jin Kim for help in setting up the OpenVLA policy. We thank Boyuan Chen and Kiwhan Song for the Diffusion Forcing GitHub repository.
References
- Afzal et al. (2020) Afsoon Afzal, Deborah S Katz, Claire Le Goues, and Christopher S Timperley. A study on the challenges of using robotics simulators for testing. arXiv preprint arXiv:2004.07368, 2020.
- AI et al. (2024) Decart AI, Julian Quevedo, Quinn McIntyre, Spruce Campbell, Xinlei Chen, and Robert Wachen. Oasis: A universe in a transformer. 2024. URL https://oasis-model.github.io/.
- Alonso et al. (2024) Eloi Alonso, Adam Jelley, Vincent Micheli, Anssi Kanervisto, Amos Storkey, Tim Pearce, and François Fleuret. Diffusion for world modeling: Visual details matter in atari. arXiv preprint arXiv:2405.12399, 2024.
- Brohan et al. (2022) Anthony Brohan, Noah Brown, Justice Carbajal, Yevgen Chebotar, Joseph Dabis, Chelsea Finn, Keerthana Gopalakrishnan, Karol Hausman, Alex Herzog, Jasmine Hsu, et al. Rt-1: Robotics transformer for real-world control at scale. arXiv preprint arXiv:2212.06817, 2022.
- Brooks et al. (2024) Tim Brooks, Bill Peebles, Connor Holmes, Will DePue, Yufei Guo, Li Jing, David Schnurr, Joe Taylor, Troy Luhman, Eric Luhman, et al. Video generation models as world simulators. 2024. URL https://openai. com/research/video-generation-models-as-world-simulators, 3, 2024.
- Bruce et al. (2024) Jake Bruce, Michael D Dennis, Ashley Edwards, Jack Parker-Holder, Yuge Shi, Edward Hughes, Matthew Lai, Aditi Mavalankar, Richie Steigerwald, Chris Apps, et al. Genie: Generative interactive environments. In Forty-first International Conference on Machine Learning, 2024.
- Chen et al. (2024) Boyuan Chen, Diego Marti Monso, Yilun Du, Max Simchowitz, Russ Tedrake, and Vincent Sitzmann. Diffusion forcing: Next-token prediction meets full-sequence diffusion. arXiv preprint arXiv:2407.01392, 2024.
- Chi et al. (2023) Cheng Chi, Zhenjia Xu, Siyuan Feng, Eric Cousineau, Yilun Du, Benjamin Burchfiel, Russ Tedrake, and Shuran Song. Diffusion policy: Visuomotor policy learning via action diffusion. The International Journal of Robotics Research, pp. 02783649241273668, 2023.
- Choi et al. (2021) HeeSun Choi, Cindy Crump, Christian Duriez, Asher Elmquist, Gregory Hager, David Han, Frank Hearl, Jessica Hodgins, Abhinandan Jain, Frederick Leve, et al. On the use of simulation in robotics: Opportunities, challenges, and suggestions for moving forward. Proceedings of the National Academy of Sciences, 118(1):e1907856118, 2021.
- Chung et al. (2023) Hyung Won Chung, Noah Constant, Xavier Garcia, Adam Roberts, Yi Tay, Sharan Narang, and Orhan Firat. Unimax: Fairer and more effective language sampling for large-scale multilingual pretraining, 2023. URL https://arxiv.org/abs/2304.09151.
- Du et al. (2023a) Yilun Du, Mengjiao Yang, Bo Dai, Hanjun Dai, Ofir Nachum, Joshua B. Tenenbaum, Dale Schuurmans, and Pieter Abbeel. Learning universal policies via text-guided video generation, 2023a. URL https://arxiv.org/abs/2302.00111.
- Du et al. (2023b) Yilun Du, Mengjiao Yang, Pete Florence, Fei Xia, Ayzaan Wahid, Brian Ichter, Pierre Sermanet, Tianhe Yu, Pieter Abbeel, Joshua B Tenenbaum, et al. Video language planning. arXiv preprint arXiv:2310.10625, 2023b.
- Du et al. (2024) Yilun Du, Sherry Yang, Bo Dai, Hanjun Dai, Ofir Nachum, Josh Tenenbaum, Dale Schuurmans, and Pieter Abbeel. Learning universal policies via text-guided video generation. Advances in Neural Information Processing Systems, 36, 2024.
- Duan & Wang (2020) Yaqi Duan and Mengdi Wang. Minimax-optimal off-policy evaluation with linear function approximation, 2020. arXiv:2002.09516.
- Dulac-Arnold et al. (2019) Gabriel Dulac-Arnold, Daniel Mankowitz, and Todd Hester. Challenges of real-world reinforcement learning. arXiv preprint arXiv:1904.12901, 2019.
- Ebert et al. (2021) Frederik Ebert, Yanlai Yang, Karl Schmeckpeper, Bernadette Bucher, Georgios Georgakis, Kostas Daniilidis, Chelsea Finn, and Sergey Levine. Bridge data: Boosting generalization of robotic skills with cross-domain datasets. arXiv preprint arXiv:2109.13396, 2021.
- Erez et al. (2015) Tom Erez, Yuval Tassa, and Emanuel Todorov. Simulation tools for model-based robotics: Comparison of bullet, havok, mujoco, ode and physx. In 2015 IEEE international conference on robotics and automation (ICRA), pp. 4397â4404. IEEE, 2015.
- Esser et al. (2024) Patrick Esser, Sumith Kulal, Andreas Blattmann, Rahim Entezari, Jonas MĂŒller, Harry Saini, Yam Levi, Dominik Lorenz, Axel Sauer, Frederic Boesel, Dustin Podell, Tim Dockhorn, Zion English, Kyle Lacey, Alex Goodwin, Yannik Marek, and Robin Rombach. Scaling rectified flow transformers for high-resolution image synthesis, 2024. URL https://arxiv.org/abs/2403.03206.
- Farajtabar et al. (2018) Mehrdad Farajtabar, Yinlam Chow, and Mohammad Ghavamzadeh. More robust doubly robust off-policy evaluation. arXiv preprint arXiv:1802.03493, 2018.
- Fonteneau et al. (2013) Raphael Fonteneau, Susan A. Murphy, Louis Wehenkel, and Damien Ernst. Batch mode reinforcement learning based on the synthesis of artificial trajectories. Annals of Operations Research, 208(1):383â416, 2013.
- Fu et al. (2021) Justin Fu, Mohammad Norouzi, Ofir Nachum, George Tucker, Ziyu Wang, Alexander Novikov, Mengjiao Yang, Michael R Zhang, Yutian Chen, Aviral Kumar, et al. Benchmarks for deep off-policy evaluation. arXiv preprint arXiv:2103.16596, 2021.
- Google (2025) Google. Image editing in gemini just got a major upgrade. Blog post on âThe Keywordâ, Google, August 26 2025. URL https://blog.google/products/gemini/updated-image-editing-model/. Multimodal Generation Lead, Gemini Apps; Gemini Image Product Lead, Google DeepMind.
- Hafner et al. (2019) Danijar Hafner, Timothy Lillicrap, Jimmy Ba, and Mohammad Norouzi. Dream to control: Learning behaviors by latent imagination. arXiv preprint arXiv:1912.01603, 2019.
- Hafner et al. (2020) Danijar Hafner, Timothy Lillicrap, Mohammad Norouzi, and Jimmy Ba. Mastering atari with discrete world models. arXiv preprint arXiv:2010.02193, 2020.
- He et al. (2015) Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition, 2015. URL https://arxiv.org/abs/1512.03385.
- Ho et al. (2022) Jonathan Ho, William Chan, Chitwan Saharia, Jay Whang, Ruiqi Gao, Alexey Gritsenko, Diederik P Kingma, Ben Poole, Mohammad Norouzi, David J Fleet, et al. Imagen video: High definition video generation with diffusion models. arXiv preprint arXiv:2210.02303, 2022.
- Hu et al. (2022) Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, et al. Lora: Low-rank adaptation of large language models. ICLR, 1(2):3, 2022.
- Jiang & Li (2015) Nan Jiang and Lihong Li. Doubly robust off-policy value evaluation for reinforcement learning. arXiv preprint arXiv:1511.03722, 2015.
- Jiang & Li (2016) Nan Jiang and Lihong Li. Doubly robust off-policy value evaluation for reinforcement learning. In International conference on machine learning, pp. 652â661. PMLR, 2016.
- Kaelbling et al. (1995) Leslie Pack Kaelbling, Michael L Littman, and Anthony R Cassandra. Partially observable markov decision processes for artificial intelligence. In International Workshop on Reasoning with Uncertainty in Robotics, pp. 146â163. Springer, 1995.
- Kaiser et al. (2019) Lukasz Kaiser, Mohammad Babaeizadeh, Piotr Milos, Blazej Osinski, Roy H Campbell, Konrad Czechowski, Dumitru Erhan, Chelsea Finn, Piotr Kozakowski, Sergey Levine, et al. Model-based reinforcement learning for atari. arXiv preprint arXiv:1903.00374, 2019.
- Kallus & Uehara (2019) Nathan Kallus and Masatoshi Uehara. Double reinforcement learning for efficient off-policy evaluation in Markov decision processes. arXiv preprint arXiv:1908.08526, 2019.
- Kanamori et al. (2009) Takafumi Kanamori, Shohei Hido, and Masashi Sugiyama. A least-squares approach to direct importance estimation. The Journal of Machine Learning Research, 10:1391â1445, 2009.
- (34) Moo Jin Kim, Karl Pertsch, Siddharth Karamcheti, Ted Xiao, Ashwin Balakrishna, Suraj Nair, Rafael Rafailov, Ethan Foster, Grace Lam, Pannag Sanketi, et al. Openvla: An open-source vision-language-action model, 2024. URL https://arxiv. org/abs/2406.09246.
- Le et al. (2019) Hoang Le, Cameron Voloshin, and Yisong Yue. Batch policy learning under constraints. In International Conference on Machine Learning, pp. 3703â3712. PMLR, 2019.
- Levine et al. (2020) Sergey Levine, Aviral Kumar, George Tucker, and Justin Fu. Offline reinforcement learning: Tutorial, review, and perspectives on open problems. arXiv preprint arXiv:2005.01643, 2020.
- Li et al. (2024) Xuanlin Li, Kyle Hsu, Jiayuan Gu, Karl Pertsch, Oier Mees, Homer Rich Walke, Chuyuan Fu, Ishikaa Lunawat, Isabel Sieh, Sean Kirmani, et al. Evaluating real-world robot manipulation policies in simulation. arXiv preprint arXiv:2405.05941, 2024.
- Li et al. (2025) Yaxuan Li, Yichen Zhu, Junjie Wen, Chaomin Shen, and Yi Xu. Worldeval: World model as real-world robot policies evaluator. arXiv preprint arXiv:2505.19017, 2025.
- Liu et al. (2024) Peiqi Liu, Yaswanth Orru, Jay Vakil, Chris Paxton, Nur Muhammad Mahi Shafiullah, and Lerrel Pinto. Ok-robot: What really matters in integrating open-knowledge models for robotics. arXiv preprint arXiv:2401.12202, 2024.
- Ma et al. (2025) Xin Ma, Yaohui Wang, Xinyuan Chen, Gengyun Jia, Ziwei Liu, Yuan-Fang Li, Cunjian Chen, and Yu Qiao. Latte: Latent diffusion transformer for video generation, 2025. URL https://arxiv.org/abs/2401.03048.
- Munos et al. (2016) R. Munos, T. Stepleton, A. Harutyunyan, and M. Bellemare. Safe and efficient off-policy reinforcement learning. In Advances in Neural Information Processing Systems, pp. 1054â1062, 2016.
- Nachum et al. (2019) Ofir Nachum, Yinlam Chow, Bo Dai, and Lihong Li. Dualdice: Behavior-agnostic estimation of discounted stationary distribution corrections. Advances in neural information processing systems, 32, 2019.
- Nguyen et al. (2010) XuanLong Nguyen, Martin J Wainwright, and Michael I Jordan. Estimating divergence functionals and the likelihood ratio by convex risk minimization. IEEE Transactions on Information Theory, 56(11):5847â5861, 2010.
- NVIDIA et al. (2025) NVIDIA, :, Niket Agarwal, Arslan Ali, Maciej Bala, Yogesh Balaji, Erik Barker, Tiffany Cai, Prithvijit Chattopadhyay, Yongxin Chen, Yin Cui, Yifan Ding, Daniel Dworakowski, Jiaojiao Fan, Michele Fenzi, Francesco Ferroni, Sanja Fidler, Dieter Fox, Songwei Ge, Yunhao Ge, Jinwei Gu, Siddharth Gururani, Ethan He, Jiahui Huang, Jacob Huffman, Pooya Jannaty, Jingyi Jin, Seung Wook Kim, Gergely KlĂĄr, Grace Lam, Shiyi Lan, Laura Leal-Taixe, Anqi Li, Zhaoshuo Li, Chen-Hsuan Lin, Tsung-Yi Lin, Huan Ling, Ming-Yu Liu, Xian Liu, Alice Luo, Qianli Ma, Hanzi Mao, Kaichun Mo, Arsalan Mousavian, Seungjun Nah, Sriharsha Niverty, David Page, Despoina Paschalidou, Zeeshan Patel, Lindsey Pavao, Morteza Ramezanali, Fitsum Reda, Xiaowei Ren, Vasanth Rao Naik Sabavat, Ed Schmerling, Stella Shi, Bartosz Stefaniak, Shitao Tang, Lyne Tchapmi, Przemek Tredak, Wei-Cheng Tseng, Jibin Varghese, Hao Wang, Haoxiang Wang, Heng Wang, Ting-Chun Wang, Fangyin Wei, Xinyue Wei, Jay Zhangjie Wu, Jiashu Xu, Wei Yang, Lin Yen-Chen, Xiaohui Zeng, Yu Zeng, Jing Zhang, Qinsheng Zhang, Yuxuan Zhang, Qingqing Zhao, and Artur Zolkowski. Cosmos world foundation model platform for physical ai, 2025. URL https://arxiv.org/abs/2501.03575.
- Octo Model Team et al. (2024) Octo Model Team, Dibya Ghosh, Homer Walke, Karl Pertsch, Kevin Black, Oier Mees, Sudeep Dasari, Joey Hejna, Charles Xu, Jianlan Luo, Tobias Kreiman, You Liang Tan, Lawrence Yunliang Chen, Pannag Sanketi, Quan Vuong, Ted Xiao, Dorsa Sadigh, Chelsea Finn, and Sergey Levine. Octo: An open-source generalist robot policy. In Proceedings of Robotics: Science and Systems, Delft, Netherlands, 2024.
- OâNeill et al. (2023) Abby OâNeill, Abdul Rehman, Abhinav Gupta, Abhiram Maddukuri, Abhishek Gupta, Abhishek Padalkar, Abraham Lee, Acorn Pooley, Agrim Gupta, Ajay Mandlekar, et al. Open x-embodiment: Robotic learning datasets and rt-x models. arXiv preprint arXiv:2310.08864, 2023.
- OpenAI et al. (2024) OpenAI, :, Aaron Hurst, Adam Lerer, Adam P. Goucher, Adam Perelman, Aditya Ramesh, Aidan Clark, AJ Ostrow, Akila Welihinda, Alan Hayes, Alec Radford, Aleksander MÄ dry, Alex Baker-Whitcomb, Alex Beutel, Alex Borzunov, Alex Carney, Alex Chow, Alex Kirillov, Alex Nichol, Alex Paino, Alex Renzin, Alex Tachard Passos, Alexander Kirillov, Alexi Christakis, Alexis Conneau, Ali Kamali, Allan Jabri, Allison Moyer, Allison Tam, Amadou Crookes, Amin Tootoochian, Amin Tootoonchian, Ananya Kumar, Andrea Vallone, Andrej Karpathy, Andrew Braunstein, Andrew Cann, Andrew Codispoti, Andrew Galu, Andrew Kondrich, Andrew Tulloch, Andrey Mishchenko, Angela Baek, Angela Jiang, Antoine Pelisse, Antonia Woodford, Anuj Gosalia, Arka Dhar, Ashley Pantuliano, Avi Nayak, Avital Oliver, Barret Zoph, Behrooz Ghorbani, Ben Leimberger, Ben Rossen, Ben Sokolowsky, Ben Wang, Benjamin Zweig, Beth Hoover, Blake Samic, Bob McGrew, Bobby Spero, Bogo Giertler, Bowen Cheng, Brad Lightcap, Brandon Walkin, Brendan Quinn, Brian Guarraci, Brian Hsu, Bright Kellogg, Brydon Eastman, Camillo Lugaresi, Carroll Wainwright, Cary Bassin, Cary Hudson, Casey Chu, Chad Nelson, Chak Li, Chan Jun Shern, Channing Conger, Charlotte Barette, Chelsea Voss, Chen Ding, Cheng Lu, Chong Zhang, Chris Beaumont, Chris Hallacy, Chris Koch, Christian Gibson, Christina Kim, Christine Choi, Christine McLeavey, Christopher Hesse, Claudia Fischer, Clemens Winter, Coley Czarnecki, Colin Jarvis, Colin Wei, Constantin Koumouzelis, Dane Sherburn, Daniel Kappler, Daniel Levin, Daniel Levy, David Carr, David Farhi, David Mely, David Robinson, David Sasaki, Denny Jin, Dev Valladares, Dimitris Tsipras, Doug Li, Duc Phong Nguyen, Duncan Findlay, Edede Oiwoh, Edmund Wong, Ehsan Asdar, Elizabeth Proehl, Elizabeth Yang, Eric Antonow, Eric Kramer, Eric Peterson, Eric Sigler, Eric Wallace, Eugene Brevdo, Evan Mays, Farzad Khorasani, Felipe Petroski Such, Filippo Raso, Francis Zhang, Fred von Lohmann, Freddie Sulit, Gabriel Goh, Gene Oden, Geoff Salmon, Giulio Starace, Greg Brockman, Hadi Salman, Haiming Bao, Haitang Hu, Hannah Wong, Haoyu Wang, Heather Schmidt, Heather Whitney, Heewoo Jun, Hendrik Kirchner, Henrique Ponde de Oliveira Pinto, Hongyu Ren, Huiwen Chang, Hyung Won Chung, Ian Kivlichan, Ian OâConnell, Ian OâConnell, Ian Osband, Ian Silber, Ian Sohl, Ibrahim Okuyucu, Ikai Lan, Ilya Kostrikov, Ilya Sutskever, Ingmar Kanitscheider, Ishaan Gulrajani, Jacob Coxon, Jacob Menick, Jakub Pachocki, James Aung, James Betker, James Crooks, James Lennon, Jamie Kiros, Jan Leike, Jane Park, Jason Kwon, Jason Phang, Jason Teplitz, Jason Wei, Jason Wolfe, Jay Chen, Jeff Harris, Jenia Varavva, Jessica Gan Lee, Jessica Shieh, Ji Lin, Jiahui Yu, Jiayi Weng, Jie Tang, Jieqi Yu, Joanne Jang, Joaquin Quinonero Candela, Joe Beutler, Joe Landers, Joel Parish, Johannes Heidecke, John Schulman, Jonathan Lachman, Jonathan McKay, Jonathan Uesato, Jonathan Ward, Jong Wook Kim, Joost Huizinga, Jordan Sitkin, Jos Kraaijeveld, Josh Gross, Josh Kaplan, Josh Snyder, Joshua Achiam, Joy Jiao, Joyce Lee, Juntang Zhuang, Justyn Harriman, Kai Fricke, Kai Hayashi, Karan Singhal, Katy Shi, Kavin Karthik, Kayla Wood, Kendra Rimbach, Kenny Hsu, Kenny Nguyen, Keren Gu-Lemberg, Kevin Button, Kevin Liu, Kiel Howe, Krithika Muthukumar, Kyle Luther, Lama Ahmad, Larry Kai, Lauren Itow, Lauren Workman, Leher Pathak, Leo Chen, Li Jing, Lia Guy, Liam Fedus, Liang Zhou, Lien Mamitsuka, Lilian Weng, Lindsay McCallum, Lindsey Held, Long Ouyang, Louis Feuvrier, Lu Zhang, Lukas Kondraciuk, Lukasz Kaiser, Luke Hewitt, Luke Metz, Lyric Doshi, Mada Aflak, Maddie Simens, Madelaine Boyd, Madeleine Thompson, Marat Dukhan, Mark Chen, Mark Gray, Mark Hudnall, Marvin Zhang, Marwan Aljubeh, Mateusz Litwin, Matthew Zeng, Max Johnson, Maya Shetty, Mayank Gupta, Meghan Shah, Mehmet Yatbaz, Meng Jia Yang, Mengchao Zhong, Mia Glaese, Mianna Chen, Michael Janner, Michael Lampe, Michael Petrov, Michael Wu, Michele Wang, Michelle Fradin, Michelle Pokrass, Miguel Castro, Miguel Oom Temudo de Castro, Mikhail Pavlov, Miles Brundage, Miles Wang, Minal Khan, Mira Murati, Mo Bavarian, Molly Lin, Murat Yesildal, Nacho Soto, Natalia Gimelshein, Natalie Cone, Natalie Staudacher, Natalie Summers, Natan LaFontaine, Neil Chowdhury, Nick Ryder, Nick Stathas, Nick Turley, Nik Tezak, Niko Felix, Nithanth Kudige, Nitish Keskar, Noah Deutsch, Noel Bundick, Nora Puckett, Ofir Nachum, Ola Okelola, Oleg Boiko, Oleg Murk, Oliver Jaffe, Olivia Watkins, Olivier Godement, Owen Campbell-Moore, Patrick Chao, Paul McMillan, Pavel Belov, Peng Su, Peter Bak, Peter Bakkum, Peter Deng, Peter Dolan, Peter Hoeschele, Peter Welinder, Phil Tillet, Philip Pronin, Philippe Tillet, Prafulla Dhariwal, Qiming Yuan, Rachel Dias, Rachel Lim, Rahul Arora, Rajan Troll, Randall Lin, Rapha Gontijo Lopes, Raul Puri, Reah Miyara, Reimar Leike, Renaud Gaubert, Reza Zamani, Ricky Wang, Rob Donnelly, Rob Honsby, Rocky Smith, Rohan Sahai, Rohit Ramchandani, Romain Huet, Rory Carmichael, Rowan Zellers, Roy Chen, Ruby Chen, Ruslan Nigmatullin, Ryan Cheu, Saachi Jain, Sam Altman, Sam Schoenholz, Sam Toizer, Samuel Miserendino, Sandhini Agarwal, Sara Culver, Scott Ethersmith, Scott Gray, Sean Grove, Sean Metzger, Shamez Hermani, Shantanu Jain, Shengjia Zhao, Sherwin Wu, Shino Jomoto, Shirong Wu, Shuaiqi, Xia, Sonia Phene, Spencer Papay, Srinivas Narayanan, Steve Coffey, Steve Lee, Stewart Hall, Suchir Balaji, Tal Broda, Tal Stramer, Tao Xu, Tarun Gogineni, Taya Christianson, Ted Sanders, Tejal Patwardhan, Thomas Cunninghman, Thomas Degry, Thomas Dimson, Thomas Raoux, Thomas Shadwell, Tianhao Zheng, Todd Underwood, Todor Markov, Toki Sherbakov, Tom Rubin, Tom Stasi, Tomer Kaftan, Tristan Heywood, Troy Peterson, Tyce Walters, Tyna Eloundou, Valerie Qi, Veit Moeller, Vinnie Monaco, Vishal Kuo, Vlad Fomenko, Wayne Chang, Weiyi Zheng, Wenda Zhou, Wesam Manassra, Will Sheu, Wojciech Zaremba, Yash Patil, Yilei Qian, Yongjik Kim, Youlong Cheng, Yu Zhang, Yuchen He, Yuchen Zhang, Yujia Jin, Yunxing Dai, and Yury Malkov. Gpt-4o system card, 2024. URL https://arxiv.org/abs/2410.21276.
- Peebles & Xie (2023) William Peebles and Saining Xie. Scalable diffusion models with transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 4195â4205, 2023.
- Precup et al. (2000) Doina Precup, Richard S. Sutton, and Satinder P. Singh. Eligibility traces for off-policy policy evaluation. In Proceedings of the 17th International Conference on Machine Learning, pp. 759â766, 2000.
- Puterman (2014) Martin L Puterman. Markov decision processes: discrete stochastic dynamic programming. John Wiley & Sons, 2014.
- Salvato et al. (2021) Erica Salvato, Gianfranco Fenu, Eric Medvet, and Felice Andrea Pellegrino. Crossing the reality gap: A survey on sim-to-real transferability of robot controllers in reinforcement learning. IEEE Access, 9:153171â153187, 2021.
- Shafiullah et al. (2023) Nur Muhammad Mahi Shafiullah, Anant Rai, Haritheja Etukuru, Yiqian Liu, Ishan Misra, Soumith Chintala, and Lerrel Pinto. On bringing robots home. arXiv preprint arXiv:2311.16098, 2023.
- Singer et al. (2022) Uriel Singer, Adam Polyak, Thomas Hayes, Xi Yin, Jie An, Songyang Zhang, Qiyuan Hu, Harry Yang, Oron Ashual, Oran Gafni, et al. Make-a-video: Text-to-video generation without text-video data. arXiv preprint arXiv:2209.14792, 2022.
- Soljacic et al. (2024) Fran Soljacic, Theresa Law, Meia Chita-Tegmark, and Matthias Scheutz. Robots in healthcare as envisioned by care professionals. Intelligent Service Robotics, pp. 1â17, 2024.
- SĂŒnderhauf et al. (2018) Niko SĂŒnderhauf, Oliver Brock, Walter Scheirer, Raia Hadsell, Dieter Fox, JĂŒrgen Leitner, Ben Upcroft, Pieter Abbeel, Wolfram Burgard, Michael Milford, et al. The limits and potentials of deep learning for robotics. The International journal of robotics research, 37(4-5):405â420, 2018.
- Sutton et al. (2009) Richard S Sutton, Hamid Reza Maei, Doina Precup, Shalabh Bhatnagar, David Silver, Csaba SzepesvĂĄri, and Eric Wiewiora. Fast gradient-descent methods for temporal-difference learning with linear function approximation. In Proceedings of the 26th annual international conference on machine learning, pp. 993â1000, 2009.
- Sutton et al. (2016) Richard S Sutton, A Rupam Mahmood, and Martha White. An emphatic approach to the problem of off-policy temporal-difference learning. Journal of Machine Learning Research, 17(73):1â29, 2016.
- Tedrake et al. (2019) Russ Tedrake et al. Drake: Model-based design and verification for robotics. 2019.
- Thomas et al. (2015a) P. Thomas, G. Theocharous, and M. Ghavamzadeh. High confidence off-policy evaluation. In Proceedings of the 29th Conference on Artificial Intelligence, 2015a.
- Thomas & Brunskill (2016) Philip Thomas and Emma Brunskill. Data-efficient off-policy policy evaluation for reinforcement learning. In International Conference on Machine Learning, pp. 2139â2148. PMLR, 2016.
- Thomas et al. (2015b) Philip Thomas, Georgios Theocharous, and Mohammad Ghavamzadeh. High-confidence off-policy evaluation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 29, 2015b.
- Todorov et al. (2012) Emanuel Todorov, Tom Erez, and Yuval Tassa. Mujoco: A physics engine for model-based control. In 2012 IEEE/RSJ international conference on intelligent robots and systems, pp. 5026â5033. IEEE, 2012.
- Valevski et al. (2024) Dani Valevski, Yaniv Leviathan, Moab Arar, and Shlomi Fruchter. Diffusion models are real-time game engines. arXiv preprint arXiv:2408.14837, 2024.
- Villegas et al. (2022) Ruben Villegas, Mohammad Babaeizadeh, Pieter-Jan Kindermans, Hernan Moraldo, Han Zhang, Mohammad Taghi Saffar, Santiago Castro, Julius Kunze, and Dumitru Erhan. Phenaki: Variable length video generation from open domain textual description. arXiv preprint arXiv:2210.02399, 2022.
- Walke et al. (2023) Homer Rich Walke, Kevin Black, Tony Z Zhao, Quan Vuong, Chongyi Zheng, Philippe Hansen-Estruch, Andre Wang He, Vivek Myers, Moo Jin Kim, Max Du, et al. Bridgedata v2: A dataset for robot learning at scale. In Conference on Robot Learning, pp. 1723â1736. PMLR, 2023.
- Wang et al. (2024) Peng Wang, Shuai Bai, Sinan Tan, Shijie Wang, Zhihao Fan, Jinze Bai, Keqin Chen, Xuejing Liu, Jialin Wang, Wenbin Ge, Yang Fan, Kai Dang, Mengfei Du, Xuancheng Ren, Rui Men, Dayiheng Liu, Chang Zhou, Jingren Zhou, and Junyang Lin. Qwen2-vl: Enhancing vision-language modelâs perception of the world at any resolution, 2024. URL https://arxiv.org/abs/2409.12191.
- Wen et al. (2025) Junjie Wen, Yichen Zhu, Jinming Li, Zhibin Tang, Chaomin Shen, and Feifei Feng. Dexvla: Vision-language model with plug-in diffusion expert for general robot control, 2025. URL https://arxiv.org/abs/2502.05855.
- Xiao et al. (2019) Chenjun Xiao, Yifan Wu, Chen Ma, Dale Schuurmans, and Martin MĂŒller. Learning to combat compounding-error in model-based reinforcement learning. arXiv preprint arXiv:1912.11206, 2019.
- Yang et al. (2020) Mengjiao Yang, Ofir Nachum, Bo Dai, Lihong Li, and Dale Schuurmans. Off-policy evaluation via the regularized lagrangian. Advances in Neural Information Processing Systems, 33:6551â6561, 2020.
- Yang et al. (2023) Mengjiao Yang, Yilun Du, Kamyar Ghasemipour, Jonathan Tompson, Dale Schuurmans, and Pieter Abbeel. Learning interactive real-world simulators. arXiv preprint arXiv:2310.06114, 2023.
- Yu et al. (2020) Tianhe Yu, Garrett Thomas, Lantao Yu, Stefano Ermon, James Y Zou, Sergey Levine, Chelsea Finn, and Tengyu Ma. Mopo: Model-based offline policy optimization. Advances in Neural Information Processing Systems, 33:14129â14142, 2020.
- Zhang et al. (2021) Michael R Zhang, Tom Le Paine, Ofir Nachum, Cosmin Paduraru, George Tucker, Ziyu Wang, and Mohammad Norouzi. Autoregressive dynamics models for offline policy evaluation and optimization. arXiv preprint arXiv:2104.13877, 2021.
- Zhao et al. (2020) Wenshuai Zhao, Jorge Peña Queralta, and Tomi Westerlund. Sim-to-real transfer in deep reinforcement learning for robotics: a survey. In 2020 IEEE symposium series on computational intelligence (SSCI), pp. 737â744. IEEE, 2020.
Appendix
Appendix A Additional Details of the Autoregressive Diffusion Transformer
Implementation details: We use the VAE from Stable Diffusion 3 Esser et al. (2024) to independently encode 256 $Ă$ 256 image frames into latent space. We employ a 16-layer transformer with 1024 hidden dimensions and 16 attention heads. We train the world model on a diverse set of data sources, including 9 of the robot datasets from Open-X Embodiment whose action spaces can be unified, such as Bridge V2 (Walke et al., 2023) and RT-1 (Brohan et al., 2022). We encode actions from a 7-dimensional vector, using the 6-dimensional end-effector position and binary gripper state as our action space. Action spaces from different robots are aligned by normalizing each componentâs 10 ${}^{\text{th}}$ - and 90 ${}^{\text{th}}$ -percentile values to those of the RT-1 dataset. We train with a context length of 20 frames; for longer rollouts, we condition on a sliding window of the last 20 frames.
| Hyperparameter | Value |
| --- | --- |
| Total parameters | 609 M |
| Image Resolution | 256 $Ă$ 256 |
| DiT Patch Size | 2 |
| Input Channels | 16 |
| Hidden Size | 1024 |
| Layers | 16 |
| Attention Heads | 16 |
| MLP Ratio | 4 |
| Optimizer | AdamW (weight decay = $0.002$ , $\beta_{1}=0.9,\beta_{2}=0.99$ ) |
| Learning rate | 8e-5 |
| Batch size | 16 |
| Action dimension | 7 |
| Training hardware | 2xA100 80GB |
| Training steps | 300k |
| Diffusion noise schedule | sigmoid |
| Sampling timesteps | 10 |
| Prediction target | $v$ |
Table 2: Hyperparameters for training WorldGymâs video prediction model.
Algorithm 1 WorldGym policy evaluation loop.
World model $\hat{T}$ with training context length $N_{\text{train}}$ and prediction horizon $h$ , rollout length $N_{\text{rollout}}$ , policy $\pi$ with action chunk size $|\mathbf{a}_{\text{pred}}|$ , reward model $\hat{R}$ , initial observation $o_{0}$ , goal $g$
$\mathbf{o}â[o_{0}]$
$\mathbf{a}â[a_{\text{null}}]$
$n=0$
while $n†N_{\text{rollout}}$ do
$\mathbf{a_{\text{pred}}}â\pi(\mathbf{o}_{n},g)$
for $i=0$ to $\lceil|\mathbf{a}_{\text{pred}}|/h\rceil-1$ do
$\mathbf{a}_{\text{ctx}}â\mathbf{a}_{-N_{\text{train}}:}$
$\mathbf{o}_{\text{ctx}}â\mathbf{o}_{-N_{\text{train}}:}$
$\mathbf{o}_{\text{pred}}â\hat{T}(\mathbf{o}_{\text{ctx}},\mathbf{a}_{\text{ctx}}||\mathbf{a}_{\text{pred},h· i:h·(i+1)})$ $\triangleright$ predict a block of $h$ frames in parallel
$\mathbf{o}â\mathbf{o}||\mathbf{o}_{\text{pred}}$ $\triangleright$ concatenate generated block of observation frames with observation history
$\mathbf{a}â\mathbf{a}||\mathbf{a}_{\text{pred},h· i:h·(i+1)}$
end for
$nâ n+n_{\text{chunk}}$
end while
$râ\hat{R}(\mathbf{o})$
Algorithm 1 shows the detailed algorithm for performing a sliding window rollout of a policy with action chunk size $|\mathbf{a}_{\text{pred}}|$ and a world model with prediction horizon $h$ . Note that in practice we always choose $h=|\mathbf{a}_{\text{pred}}|$ at inference time.
Appendix B Details of VLM as Reward
B.1 Prompt for VLM as Reward
Prompt GPT-4o as Reward $\hat{R}$ . Note that has_partial is True if the chosen task has a partial credit criteria, which is the case for some tasks used in OpenVLA (Kim et al., ).
⏠Here is a sequence of frames from a robot policy which has been rolled out in a video - generation - based world model. I need your help determining whether the policy is successful. How successfully does the robot complete the following task? Instruction: {instruction} {rubric. strip ()} Provide brief reasoning (2-3 sentences). Then output EXACTLY one final line: Final Score: X Where X is { â one of 0, 0.5, or 1â if has_partial else â0 or 1â }. No extra numbers after that line. Note: Since this video was generated by a video prediction model (conditioned on robot actions), it may contain some artifacts due to the video model capacity.
If there is a partial credit criteria, the rubric is:
⏠0 = Failure: little or no progress toward: "{instruction}" 0.5 = Partial: "{partial_desc}" achieved BUT the instruction not fully completed 1 = Success: Instruction fully completed (counts even if partial also true)
Otherwise, the rubric is:
⏠Score rubric: 0 = Failure: instruction "{instruction}" not completed. 1 = Success: instruction completed.
B.2 Validating VLM Success Predictions
Table 3: Performance of VLM as reward (mean and standard error across 4 runs) on videos from RT-1 (Brohan et al., 2022) using ground truth task success labels. GPT-4o achieves high true positives and true negatives. Notably, GPT-4o as reward has very low false positive rate, which is especially important for not over-estimating a policy value.
| | RT-1 Success | RT-1 Fail |
| --- | --- | --- |
| VLM Success | 0.81 $±$ 0.14 (TP) | 0.03 $±$ 0.05 (FP) |
| VLM Fail | 0.19 $±$ 0.14 (FN) | 0.97 $±$ 0.05 (TN) |
To determine whether a VLM can serve as a reliable reward function, we pass rollout videos from the RT-1 dataset, along with the prompts constructed from the templates above, as inputs to query GPT-4o. We use whether the task is successful according to the RT-1 data (validation split) as the ground truth. Table 3 shows that GPT-4o achieves high true positive and true negative rate for real videos, indicating that it is an effective evaluator of task success. Notably, GPT-4o achieves very low false positives (i.e., the rollout is a failure but the VLM thinks it is a success), which is highly useful in policy evaluation.
Appendix C Architecture and Training Details of Video Based Policy
Our video-based policy follows the framework of UniPi Du et al. (2023a), combining a language-conditioned video prediction model with an inverse dynamics model.
The video prediction module shares the same architecture as our world model, but replaces the conditioning on robot actions at each timestep with language instructions. For language conditioning, we employ the pretrained and frozen UMT5-xxl encoder Chung et al. (2023) to obtain token-level embeddings. These embeddings are aggregated via mean pooling to form a 4096-dimensional instruction representation. This representation is projected to match the model dimensionality and is used to modulate the diffusion transformer through adaptive layer normalization (adaLN-Zero). In this way, task semantics are directly integrated into the video prediction process. We train our video generation model for 180k steps on Bridge V2 (Walke et al., 2023). The visualization of the video generation policy on validation scenes can be seen in Figure 14.
The inverse dynamics model predicts the action sequence given a short video clip of 10 frames. Each frame is encoded with a ResNet 50 backbone He et al. (2015), producing per-frame features $f_{t}$ . To capture motion, we compute both $f_{t}$ and temporal differences $\Delta f_{t}=f_{t+1}-f_{t}$ , concatenate them, and flatten across the clip. The resulting representation is passed through an MLP to predict $10Ă d_{a}$ outputs, corresponding to the action dimension $d_{a}$ at each timestep. Input images are normalized with ImageNet statistics, and the model is trained with mean squared error on ground-truth actions. The inverse dynamics model is trained independently on the Bridge V2 dataset for 200k steps.
C.1 Validation Visualization of Language Conditioned Video Generation Model
<details>
<summary>x15.png Details</summary>

### Visual Description
## Image Comparison: Ground-truth vs. Generated Scenarios
### Overview
The image presents a side-by-side comparison of 10 scenarios (5 rows Ă 2 columns) showcasing "Ground-truth" (real-world) images and their corresponding "Generated" (AI-simulated) versions. Each row represents a distinct task or environment, with multiple sequential images per scenario to illustrate motion or interaction.
### Components/Axes
- **Columns**:
- Left: "Ground-truth" (real-world scenarios)
- Right: "Generated" (AI-simulated scenarios)
- **Rows**:
1. Kitchen (mug, fruits, robot arm)
2. Blocks (stacked cubes, robot arm)
3. Toys (plastic animals, robot arm)
4. Cloth (folded fabric, robot arm)
5. Oven (burner, pot, robot arm)
- **Image Layout**:
- Each scenario contains 5 sequential images (e.g., robot arm movement).
- No explicit axis markers, legends, or numerical data.
### Detailed Analysis
- **Kitchen Scenario**:
- Ground-truth: Mug, banana, apple, and robot arm interacting with objects.
- Generated: Similar objects but slight positional discrepancies (e.g., banana slightly misaligned).
- **Blocks Scenario**:
- Ground-truth: Blue, yellow, red, and green cubes stacked in a tower.
- Generated: Cubes stacked but with minor color blending (e.g., red cube appears slightly orange).
- **Toys Scenario**:
- Ground-truth: Pink pig, green dinosaur, and robot arm.
- Generated: Toys rendered with lower resolution; pig appears slightly flattened.
- **Cloth Scenario**:
- Ground-truth: Blue fabric folded neatly on a table.
- Generated: Fabric rendered with unrealistic creases and lighting artifacts.
- **Oven Scenario**:
- Ground-truth: Red burner, silver pot, and robot arm.
- Generated: Burner appears oversaturated; pot lacks reflective details.
### Key Observations
1. **Consistency**: Generated images generally replicate object shapes and arrangements but with minor inaccuracies.
2. **Lighting/Texture**: Ground-truth images exhibit realistic lighting and material textures (e.g., glossy mug, matte blocks), while generated versions show simplified or exaggerated effects.
3. **Motion Artifacts**: In dynamic scenarios (e.g., robot arm movement), generated sequences sometimes lag in synchronization with ground-truth.
### Interpretation
The comparison highlights the capabilities and limitations of AI-generated imagery in replicating real-world scenarios. While object placement and basic interactions are well-simulated, subtle details like lighting, texture, and motion precision reveal areas for improvement. This suggests the AI model excels at structural replication but struggles with nuanced environmental fidelity. The scenarios span diverse tasks (kitchen, stacking, play, fabric handling, appliance use), indicating the model's versatility across contexts. Outliers, such as the flattened toy pig, point to potential weaknesses in rendering complex geometries or soft materials.
</details>
Figure 14: Validation Visualization of the language-conditioned video generation model on Bridge-V2. At inference, the model takes a UMT5-xxl instruction encoding and an initial frame, then predicts the next nine frames to complete the task.
Appendix D Architecture and Training Details of Diffusion Policy
We followed the recipe of DexVLA (Wen et al., 2025) for training the diffusion policy. We load the Qwen2-VL-2B (Wang et al., 2024) backbone and the pre-trained control head, and perform an adaptation on BridgeV2 using LoRA (Hu et al., 2022), which inserts low-rank adapter matrices inside the backboneâs attention and feed-forward blocks. These matrices along with the policy head are the only trainable modules in the adaptation stage. This preserves the backboneâs general vision-language competence, and makes adaptation compute and memory efficient. We fine-tune the model with adapters on Bridge-V2 for 60k steps. During training, we rescale the actions to $(-1,1)$ to match the diffusion target range.
DexVLAâs policy head is trained as a denoising diffusion model with the standard $\epsilon$ -prediction DDPM objective, i.e. at each update Gaussian noise is added to ground-truth action sequences at a randomly sampled diffusion step and the network is trained to predict that noise using an MSE loss. At inference, actions are generated with DDIM in a small number of steps, progressively denoising from a Gaussian initialization to a trajectory. We choose AdamW for the optimizer, using standard decoupled weight decay which applies decay to linear/attention weights, but exclude biases and LayerNorm parameters.
Appendix E Additional Experimental Results
E.1 Additional Results on Real-Robot Videos
<details>
<summary>x16.png Details</summary>

### Visual Description
## Image Comparison: Ground-truth vs. Generated Scenes
### Overview
The image presents a side-by-side comparison of **ground-truth** (real-world) and **generated** (simulated or AI-created) visual scenes. There are six rows, each depicting a distinct scenario:
1. Desk with a box
2. Robotic arm interacting with objects
3. Coffee machine setup
4. Bowl with a robotic arm
5. Blue bin with objects
6. Desk with a keyboard
Each row contains two columns: **Ground-truth** (left) and **Generated** (right). The images are arranged in a grid format, with no explicit axis titles, legends, or numerical data.
---
### Components/Axes
- **Columns**:
- **Left Column**: Labeled "Ground-truth" (real-world reference images).
- **Right Column**: Labeled "Generated" (simulated or AI-generated images).
- **Rows**:
- Each row corresponds to a unique scenario (e.g., "Desk with a box," "Robotic arm," etc.).
- No explicit axis markers or scales are present.
---
### Detailed Analysis
#### Row 1: Desk with a Box
- **Ground-truth**: A wooden desk with a light-colored box, a blue object inside, and a cluttered background (e.g., bottles, a keyboard).
- **Generated**: The box and desk are rendered with similar lighting and composition, but minor discrepancies in object placement (e.g., the blue object appears slightly misaligned).
#### Row 2: Robotic Arm
- **Ground-truth**: A robotic arm holding a yellow object over a table with a red object.
- **Generated**: The arm and objects are visually similar, but the red objectâs position and the armâs grip show slight inaccuracies.
#### Row 3: Coffee Machine
- **Ground-truth**: A coffee machine with a white cup, a black tray, and a wooden surface.
- **Generated**: The machine and cup are rendered with comparable details, but the trayâs texture and the cupâs handle appear less defined.
#### Row 4: Bowl with a Robotic Arm
- **Ground-truth**: A pink bowl on a wooden table with a robotic arm hovering above it.
- **Generated**: The bowl and arm are visually consistent, but the armâs shadow and the bowlâs rim show minor artifacts.
#### Row 5: Blue Bin with Objects
- **Ground-truth**: A blue bin containing a red ball and a yellow object on a wooden floor.
- **Generated**: The bin and objects are rendered with similar colors, but the red ballâs position and the yellow objectâs texture differ slightly.
#### Row 6: Desk with a Keyboard
- **Ground-truth**: A desk with a black keyboard, a white mouse, and a monitor.
- **Generated**: The keyboard and mouse are present, but the monitorâs angle and the deskâs edge show noticeable discrepancies.
---
### Key Observations
1. **Fidelity**: Generated images generally replicate the ground-truth scenes but with minor inaccuracies in object positioning, texture, and lighting.
2. **Artifacts**: The generated images exhibit subtle artifacts (e.g., blurred edges, misaligned shadows) in complex scenarios (e.g., robotic arm, coffee machine).
3. **Consistency**: Simpler scenes (e.g., desk with a box) show higher fidelity compared to dynamic interactions (e.g., robotic arm).
---
### Interpretation
The image demonstrates the capabilities and limitations of a generative model in replicating real-world scenarios. While the generated images maintain overall structure and color accuracy, they struggle with fine details (e.g., object alignment, texture rendering). This suggests the model excels at macro-level composition but requires improvement in micro-level precision. The discrepancies highlight challenges in simulating dynamic interactions (e.g., robotic arm movements) and material properties (e.g., reflective surfaces).
No numerical data or explicit textual labels beyond the column headers ("Ground-truth" and "Generated") are present. The analysis is based solely on visual comparison.
</details>
Figure 15: Additional Qualitative Evaluation of simulating actions from different robots. The world model generally generates the video that look very similar to the original video conditioned on the same actions that produced the original video in the real world.
<details>
<summary>x17.png Details</summary>

### Visual Description
## Grid of Robot Arm Interaction Scenarios
### Overview
The image displays a 4x3 grid of photographs (12 total) showing a robotic arm interacting with objects in various spatial configurations. Each image is labeled with directional or operational terms ("Right," "Left," "Backward," "Forward," "Close," "Open") to indicate the robot's movement or action. The scenes appear to be part of a controlled experiment or demonstration of robotic manipulation capabilities.
### Components/Axes
- **Grid Structure**:
- 4 rows, 3 columns
- Each cell contains a labeled photograph
- **Labels**:
- **Top Row**: "Right" (x2), "Left"
- **Second Row**: "Backward" (x2), "Forward"
- **Third Row**: "Close" (x2), "Open"
- **Fourth Row**: "Close" (x2), "Open"
- **Visual Elements**:
- Robotic arm (black, industrial-style) in all images
- Objects:
- Kitchen items (sink, fruits, bowls) in top row
- Tools (wrenches, pliers) in bottom rows
- Purple fabric and metallic spheres in middle rows
- Backgrounds vary slightly (kitchen counter, wooden surface, gray backdrop)
### Detailed Analysis
1. **Top Row (Spatial Directions)**:
- **Right**: Robot arm positioned to the right of a sink with fruits and bowls.
- **Right**: Similar setup but with slight positional adjustment of the arm.
- **Left**: Arm shifted to the left side of the sink area.
2. **Second Row (Movement Directions)**:
- **Backward**: Arm retracting from a wooden surface with a purple fabric.
- **Backward**: Arm further retracted, revealing a metallic sphere.
- **Forward**: Arm extended toward a metallic sphere on the surface.
3. **Third Row (Object Interaction)**:
- **Close**: Arm approaching a set of tools (wrenches, pliers) on a wooden surface.
- **Open**: Arm grasping a tool, with visible tool separation.
- **Close**: Arm returning to the tool set.
4. **Fourth Row (Repetition)**:
- Mirrors the third row: "Close," "Open," "Close" sequence with identical tool configurations.
### Key Observations
- **Repetition**: Labels "Close" and "Open" appear twice in the final two rows, suggesting a focus on grasping mechanics.
- **Spatial Consistency**: Directional labels ("Right," "Left," "Backward," "Forward") are paired with consistent object placements, implying controlled testing of movement precision.
- **Object Variety**: Tools, kitchen items, and abstract objects (fabric, spheres) indicate testing across different interaction scenarios.
### Interpretation
This grid likely demonstrates a robotic system's ability to:
1. Navigate spatial directions (left/right, forward/backward)
2. Manipulate objects at varying distances ("Close"/"Open")
3. Adapt to different object types (tools, kitchenware, abstract items)
The repetition of "Close"/"Open" in the final rows emphasizes precision in grasping, while the directional labels in earlier rows test positional awareness. The use of diverse objects suggests the system is being evaluated for versatility in real-world applications. No numerical data or quantitative metrics are visible, so conclusions are based on qualitative observation of movement patterns and object interactions.
</details>
Figure 16: Additional End-Effector Control Sweep on Bridge. We simulate different gripper controls along different action dimensions corresponding to left-right, forward-backward, and gripper open-close. The world model generally generates videos that follow the actions.
E.2 Additional Results on Google Robot
To assess the generalizability of WorldGym, we performed rollouts with different policies on Google Robot. For our analysis, we chose a subset of tasks from the RT-1 dataset (Brohan et al., 2022). A partial score of 0.5 was assigned to a rollout if the robot attempted to reach the target location. OpenVLA again outperformed Octo and RT-1-X (see Table 4). However, in this environment Octo and RT-1-X are narrowly behind. The strong performance of RT-1-X might be due to a higher proportion of Google Robot trajectories than WidowX in its pretraining mix.
Table 4: Policy rollouts on Google Robot (RT-1 subset). OpenVLA outperforms RT-1-X and Octo, but by a smaller margin than on the Bridge dataset.
| Task | # Trials | RT-1-X # Successes | Octo # Successes | OpenVLA # Successes |
| --- | --- | --- | --- | --- |
| Close Bottom Drawer | 10 | 9 | 8.5 | 6 |
| Open Left Fridge Door | 10 | 4.5 | 3.5 | 4 |
| Pick Blue Chip Bag | 10 | 5 | 5.5 | 9 |
| Place Redbull Can Upright | 10 | 1.5 | 3 | 3.5 |
E.3 Detailed Results on the OpenVLA Bridge evaluation tasks
Table 5: Detailed Bridge Evaluation Results comparing RT-1-X (OâNeill et al., 2023), Octo (Octo Model Team et al., 2024), and OpenVLA (Kim et al., ) on the Bridge evaluation suite of tasks from Kim et al. . Real-world task success rates are taken directly from (Kim et al., ), WorldGym success rates are from rolling out policies within our world model.
| Task | # Trials | RT-1-X | Octo | OpenVLA | | | |
| --- | --- | --- | --- | --- | --- | --- | --- |
| Real-world # Successes | WorldGym # Successes | Real-world # Successes | WorldGym # Successes | Real-world # Successes | WorldGym # Successes | | |
| Put Eggplant into Pot (Easy Version) | 10 | 1 | 1 | 5 | 1 | 10 | 7 |
| Put Eggplant into Pot | 10 | 0 | 0 | 1 | 2 | 10 | 6 |
| Put Cup from Counter into Sink | 10 | 1 | 3 | 1 | 3 | 7 | 9 |
| Put Eggplant into Pot (w/ Clutter) | 10 | 1 | 0.5 | 3.5 | 3.5 | 7.5 | 8 |
| Put Yellow Corn on Pink Plate | 10 | 1 | 3 | 4 | 6 | 9 | 9.5 |
| Lift Eggplant | 10 | 3 | 2 | 0.5 | 1.5 | 7.5 | 7.5 |
| Put Carrot on Plate (w/ Height Change) | 10 | 2 | 0.5 | 1 | 3 | 4.5 | 6 |
| Put Carrot on Plate | 10 | 1 | 0 | 0 | 1 | 8 | 4 |
| Flip Pot Upright | 10 | 2 | 3 | 6 | 1 | 8 | 5 |
| Lift AAA Battery | 10 | 0 | 1 | 0 | 0 | 7 | 4 |
| Move Skull into Drying Rack | 10 | 1 | 2 | 0 | 3 | 5 | 5 |
| Lift White Tape | 10 | 3 | 1 | 0 | 1 | 1 | 6 |
| Take Purple Grapes out of Pot | 10 | 6 | 5 | 0 | 2 | 4 | 4 |
| Stack Blue Cup on Pink Cup | 10 | 0.5 | 0 | 0 | 0 | 4.5 | 6 |
| Put {Eggplant, Red Bottle} into Pot | 10 | 2.5 | 0.5 | 4 | 5 | 7.5 | 9 |
| Lift {Cheese, Red Chili Pepper} | 10 | 1.5 | 2.5 | 2.5 | 2.5 | 10 | 10 |
| Put {Blue Cup, Pink Cup} on Plate | 10 | 5 | 1.5 | 5.5 | 5 | 9.5 | 8.5 |
| Mean Success Rate | | 18.5 $±$ 4.0% | \cellcolor lightlightgray15.5 $±$ 3.4% | 20.0 $±$ 5.3% | \cellcolor lightlightgray23.82 $±$ 4.3% | 70.6 $±$ 6.1% | \cellcolor lightlightgray67.4 $±$ 4.9% |
We report the mean success rate across tasks with standard error (SE) computed as
$$
\text{SE}=\frac{\mathrm{sd}(r_{1},\dots,r_{T})}{\sqrt{T}},
$$
where $r_{i}$ is the per-task success rate and $T$ is the number of tasks.
E.4 Detailed Results on OOD Image Evaluation Tasks
Table 6: Detailed Bridge OOD Image task results. OpenVLA appears to be more robust across the different OOD settings of object generalization, distractions and classification.
| Category | Task | # Trials | RT-1-X # Successes | Octo # Successes | OpenVLA # Successes |
| --- | --- | --- | --- | --- | --- |
| Object Generalization | Pick up Orange (Carrot closer to Gripper) | 10 | 1 | 1 | 4 |
| Object Generalization | Pick up Orange (Orange closer to Gripper) | 10 | 3 | 4 | 9 |
| Object Generalization | Pick up Orange (Replace Carrot with Radish) | 10 | 1 | 4 | 10 |
| Distractor Robustness | Pick up Carrot (With Computer on side) | 10 | 6 | 7 | 9 |
| Distractor Robustness | Pick up Carrot (Computer closer to gripper) | 10 | 3 | 1 | 8 |
| Classification | Pick {Red, Blue} | 20 | 8 | 10 | 20 |
| Classification | Pick {Circle, Square} | 20 | 8 | 10 | 12 |
| Classification | Pick {Taylor Swift, Snoop Dogg} | 20 | 7 | 10 | 11 |
Table 7: Policy rollout performance comparison in the presence of unrelated distractions. OpenVLA is more robust to distractions over RT-1-X and Octo. However, all policies suffer significant performance drop in the presence of distractors.
| Task | # Trials | RT-1-X # Successes | Octo # Successes | OpenVLA # Successes |
| --- | --- | --- | --- | --- |
| Put Eggplant into Pot (Easy Version) | 10 | 1 | 1 | 3 |
| Put Eggplant into Pot | 10 | 0 | 0 | 6 |
| Put Cup from Counter into Sink | 10 | 0 | 1 | 8 |
| Put Eggplant into Pot (w/ Clutter) | 10 | 0 | 0 | 4 |
| Put Yellow Corn on Pink Plate | 10 | 0 | 0 | 2 |
| Lift Eggplant | 10 | 0 | 1 | 7 |
| Put Carrot on Plate (w/ Height Change) | 10 | 0 | 0 | 2 |
| Put Carrot on Plate | 10 | 0 | 2 | 4 |
| Flip Pot Upright | 10 | 0 | 0 | 0 |
| Lift AAA Battery | 10 | 1 | 0 | 1 |
| Move Skull into Drying Rack | 10 | 3 | 0 | 5 |
| Lift White Tape | 10 | 0 | 0 | 4 |
| Take Purple Grapes out of Pot | 10 | 7 | 0 | 3 |
| Stack Blue Cup on Pink Cup | 10 | 0 | 0 | 4 |
| Put {Eggplant, Red Bottle} into Pot | 10 | 0 | 0 | 5 |
| Lift {Cheese, Red Chili Pepper} | 10 | 1 | 2 | 6 |
| Put {Blue Cup, Pink Cup} on Plate | 10 | 0 | 0 | 3 |
| Mean Success Rate | | 7.6 $±$ 4.3% | 4.1 $±$ 1.7% | 39.4 $±$ 5.1% |
Appendix F Ablation Studies
F.1 Dataset Size Analysis
Table 8: Dataset ablation. Larger training dataset improves all three metrics comparing generated videos and ground-truth validation videos. $\uparrow$ means higher the better.
| | Subset (Bridge V1) | Full (Bridge V2) |
| --- | --- | --- |
| MSE $\downarrow$ | 0.015 | 0.010 |
| LPIPS $\downarrow$ | 0.131 | 0.073 |
| SSIM $\uparrow$ | 0.735 | 0.827 |
We measure MSE, LPIPS, and SSIM on generated videos from a model that is trained on less video data (Bridge V1 (Ebert et al., 2021)) and compare with a model that is trained on more data (Bridge V2 (Walke et al., 2023)). Table 8 shows that the model trained on more data leads to improvements in all three metrics.
F.2 Parallelism Efficiency Analysis
Table 9: Parallelism efficiency comparison. Inference time for generating 40-frame video rollouts on an A100 GPU with different horizon lengths, demonstrating the efficiency gains from parallel frame denoising.
| Prediction Horizon | Time (s) |
| --- | --- |
| $h=1$ | 93 |
| $h=4$ | 33 |
Increasing the horizon from 1 to 4 frames achieves a 2.8Ă speedup. This is particularly useful for evaluating robot policies with differing action chunk sizes. For instance, OpenVLA (Kim et al., ) predicts just a single action per frame, while Octo (Octo Model Team et al., 2024) predicts 4 actions per frame. Using the same world model checkpoint, we can improve the efficiency of rollout generations by matching the horizon to the policyâs chunk size at inference time.