MultiWorld: Scalable Multi-Agent Multi-View Video World Models

Summary (Overview)

  • Key Contribution: Introduces MultiWorld, a unified framework for scalable multi-agent, multi-view video world modeling that addresses three core challenges: multi-agent controllability, multi-view consistency, and framework scalability.
  • Core Modules: Proposes the Multi-Agent Condition Module (MACM) for precise control of multiple agents via Agent Identity Embedding and Adaptive Action Weighting, and the Global State Encoder (GSE) using a VGGT backbone to ensure multi-view consistency by encoding a shared 3D-aware environment state.
  • Scalability: Framework supports arbitrary numbers of agents and camera views without architectural changes, enabling parallel view generation and stable long-horizon autoregressive simulation.
  • Empirical Validation: Demonstrates superior performance on multi-player game (It Takes Two) and multi-robot manipulation (RoboFactory) datasets, outperforming baselines in video fidelity (FVD), action-following ability, and multi-view consistency (RPE).
  • Practical Application: Shows capability to simulate realistic failure trajectories and long-horizon scenarios (up to 4× training context) for collaborative robotics, highlighting potential for downstream tasks like planning and simulation.

Introduction and Theoretical Foundation

Video world models, typically formulated as action-conditioned video generation models, have succeeded in simulating environmental dynamics. However, existing models implicitly assume a single-agent scenario, ignoring the complex interactions and interdependencies inherent in real-world multi-agent systems (e.g., collaborative robotics, multi-player games). Furthermore, multi-agent simulation inherently requires generating consistent observations across multiple views, as each agent perceives the shared environment from its own viewpoint. Previous models fail to preserve this multi-view consistency.

Extending world models to multi-agent, multi-view scenarios introduces three key challenges:

  1. Multi-Agent Controllability: Controlling multiple agents requires associating specific actions with corresponding agents and synchronizing their executions. Simple stacking of actions leads to identity ambiguity (e.g., "mirror actions").
  2. Multi-View Consistency: The model must synthesize visually coherent videos across diverse perspectives, ensuring observations from multiple agents remain geometrically consistent within a shared environment.
  3. Framework Scalability: Real-world environments involve variable numbers of agents and views. Previous works assume fixed counts or predefined setups, limiting applicability.

MultiWorld addresses these challenges by proposing a scalable framework that enables flexible scaling of agent and view counts. The theoretical foundation combines:

  • Flow Matching (FM) as the generative backbone for conditional video distribution.
  • Transformer architecture for sequential modeling with causal masking to prevent future information leakage.
  • Relative positional embeddings (RoPE) for agent identity distinction.
  • 3D-aware latent representations from a pre-trained reconstruction model (VGGT) to encode a global environment state.

Methodology

3.1 Backbone and Notation

The problem is formulated as a collection of CC image-action-conditioned video generation tasks, where CC is the number of camera views. Videos from different views are synthesized in parallel conditioned on a shared global environment state.

  • Notation: KK agents, CC camera views (independent). Let ai=(ai1,...,aiK)a_i = (a_i^1, ..., a_i^K) denote the joint action of all KK agents at frame ii, and a={a0,...,aI}a = \{a_0, ..., a_I\} the full action sequence. The video from camera cc is xcx^c. The environment observation is o={oc}c=1Co = \{o^c\}_{c=1}^C, where oco^c is the initial frame of xcx^c.
  • Flow Matching (FM): For each view cc, the conditional distribution of future frames is modeled. Sampling tU(0,1)t \sim U(0,1) and noise ϵN(0,I)\epsilon \sim N(0,I), the noisy observation and target velocity are:
xtc=(1t)xc+tϵ,x_t^c = (1 - t) x^c + t \epsilon, u=ϵxc.u = \epsilon - x^c.

The velocity network is parameterized as vθ(xtc,t,a,o)v_\theta(x_t^c, t, a, o), trained to predict uu.

  • Causal Masking: A frame-wise causal mask is applied to action cross-attention to ensure tokens at frame ii attend only to actions from frames {0,...,i}\{0, ..., i\}, supporting stable autoregressive generation.

3.2 Multi-Agent Condition Module (MACM)

MACM processes multi-agent actions to resolve identity ambiguity and prioritize dynamic agents.

Agent Identity Embedding (AIE) To support variable agent counts and eliminate identity confusion, AIE uses Rotary Position Embedding (RoPE). Given an action latent at frame ff, afRK×Da_f \in \mathbb{R}^{K \times D} (DD is latent dimension), for each agent i{1,...,K}i \in \{1, ..., K\}, a rotation matrix RΘ,iR_{\Theta,i} is applied:

AIE(ai,i)=RΘ,iai\text{AIE}(a^i, i) = R_{\Theta,i} a^i

where RΘ,iR_{\Theta,i} is defined by pre-computed frequencies θj=b2j/D\theta_j = b^{-2j/D}, with bb as the base frequency constant. For each pair of dimensions (2j,2j+1)(2j, 2j+1):

[a(2j)a(2j+1)]out=[cos(iθj)sin(iθj)sin(iθj)cos(iθj)][a(2j)a(2j+1)]in\begin{bmatrix} a^{(2j)} \\ a^{(2j+1)} \end{bmatrix}_{\text{out}} = \begin{bmatrix} \cos(i\theta_j) & -\sin(i\theta_j) \\ \sin(i\theta_j) & \cos(i\theta_j) \end{bmatrix} \begin{bmatrix} a^{(2j)} \\ a^{(2j+1)} \end{bmatrix}_{\text{in}}

The attention between agents mm and nn becomes:

(Rmam)(Rnan)=amRmRnan=amRnman.(R_m a_m)^\top (R_n a_n) = a_m^\top R_m^\top R_n a_n = a_m^\top R_{n-m} a_n.

This provides relative identity encoding.

Adaptive Action Weighting (AAW) An MLP predicts adaptive weighting factors for each action token. Tokens are multiplied by their weights and summed into a unified representation per frame, allowing the model to focus on active agents driving environmental change.

3.3 Global State Encoder (GSE)

GSE aggregates multi-view observations into a compact, 3D-aware global environment state to ensure consistency.

  • Process: Given multi-view observation set O={Oc}c=1CO = \{O^c\}_{c=1}^C with CC images OcR3×H×WO^c \in \mathbb{R}^{3 \times H \times W}, a frozen VGGT backbone (an end-to-end 3D reconstruction model) encodes them:
Hvggt=VGGT(O),HvggtRC×n×dH_{\text{vggt}} = \text{VGGT}(O), \quad H_{\text{vggt}} \in \mathbb{R}^{C \times n \times d}

where nn is tokens per image, dd is latent dimension.

  • Alignment: An MLP aligns features with the DiT backbone: H=MLP(Hvggt)H = \text{MLP}(H_{\text{vggt}}). HH is injected via cross-attention to condition generation.
  • Advantages: Improves multi-view spatial consistency, supports arbitrary view counts by compression, and enables parallel view generation.

3.4 Scalable Framework

  • Agent Scalability: AIE's relative embeddings can be extrapolated, accommodating arbitrary KK without architectural changes.
  • View Scalability: GSE compresses variable-length observations into a unified global state, decoupling computation from CC. Parallel generation of different views keeps inference latency nearly constant with scaled resources (~1.5× speedup over sequential for double-view).
  • Autoregressive Long-Horizon Simulation: Videos are generated chunk-wise. The last frames of a chunk update the global state via GSE for the next chunk, enabling stable generation beyond training context length (up to 2× with minimal degradation, extendable to 4×).

Empirical Validation / Results

Datasets & Metrics

  • Multi-player game: 500 hours (100 hours cleaned) of real-player data from It Takes Two, 2560×1440 resolution, over 21M frames.
  • Multi-robot manipulation: Constructed using RoboFactory, tasks with 2-4 agents, variable camera viewpoints.
  • Metrics:
    • Visual quality: FVD, PSNR, SSIM, LPIPS.
    • Multi-view consistency: Reprojection Error (RPE).
    • Action-following: Inverse Dynamics Model (IDM) accuracy.

Baselines

  1. Standard Image-Action-to-Video Model: Treats each view independently.
  2. Concatenated-View Video World Model: Combines views into a single video (fixed view count, memory-intensive).
  3. COMBO: Compositional multi-agent model (two-stage: train single-agent models, then combine). Neglects inter-agent interactions.

Implementation Details

  • Backbone: Wan2.2-5B model.
  • Resolution: 320×320 (game), 320×256 (robotics).
  • Training: 40,000 iterations, learning rate 5e-5, cosine scheduler, batch size 64 on 8 NVIDIA A800 GPUs (~4 days).

Main Results

Table 1: Quantitative comparison across scenarios

MethodFVD ↓LPIPS ↓SSIM ↑PSNR ↑Action ↑RPE ↓
Multi-Player Video Game
Standard2450.360.5017.4888.40.75
Concat-View2150.360.4917.5489.10.74
Combo2070.340.5117.8289.30.72
Ours1790.350.5117.7289.80.67
Multi-Robot Manipulation
Standard1000.070.9026.3988.21.60
Concat-View*1060.060.9027.4492.00.82
Combo990.080.9026.4988.51.54
Ours960.070.9026.6088.71.52

*Concat-View trained only on two camera views per episode, not directly comparable.

MultiWorld achieves best or second-best performance across most metrics in both domains, demonstrating effectiveness and generalization.

Ablation Studies

Table 2: Ablation of main architectural components

ConfigFVD ↓LPIPS ↓SSIM ↑PSNR ↑Action ↑RPE ↓
Standard2450.360.5017.4888.40.75
+ MACM2280.360.5117.5689.70.76
Both1790.350.5117.7289.80.67

MACM improves action controllability; GSE improves multi-view consistency; together they enhance visual quality.

Table 3: Ablation on Agent Identity Embedding base frequency

Config (base)FVD ↓PSNR ↑Action ↑
base=10k23417.5389.2
base=2022817.5689.7

Lower base frequency (20 vs default 10000) provides better angular separation for agent identities, improving performance.

Table 4: Ablation on Adaptive Action Weighting

ConfigFVD ↓PSNR ↑Action ↑
w/o AAW24517.4888.4
w/ AAW23617.5288.6

AAW enhances visual quality and action-following by prioritizing active agents.

Table 5: Ablation on Global State Encoder backbone

Global State EncoderFVD ↓LPIPS ↓SSIM ↑PSNR ↑RPE ↓
w/o Global State2280.360.5117.560.75
Wan VAE2560.360.5017.380.71
DINOv22320.360.5017.480.72
VGGT (Ours)1790.350.5117.720.67

VGGT, as a 3D reconstruction model, best captures shared 3D environment state, leading to superior multi-view consistency and visual quality.

Qualitative Results

  • Multi-player game: MultiWorld achieves more accurate action following and better multi-view consistency compared to baselines, mitigating failure modes like inaccurate action execution, agent disappearance, and view inconsistency.
  • Multi-robot failure trajectory simulation: MultiWorld can simulate realistic failure modes (e.g., inter-robot collisions) by enhancing failure trajectory data, useful for safe robotics training.
  • Long-horizon generation: Autoregressive simulation of three robots stacking cubes fluently up to 2× training context without significant degradation, extendable to 4×.

Theoretical and Practical Implications

  • Theoretical: Advances video world modeling by formally addressing multi-agent controllability and multi-view consistency, introducing scalable architectures (MACM, GSE) that generalize beyond fixed configurations. Provides a framework for modeling inter-agent interactions and shared 3D environment states.
  • Practical:
    • Robotics: Enables simulation of multi-robot collaborative tasks, including failure trajectories, reducing physical risk and data collection burden.
    • Gaming: Supports realistic multi-player environment simulation with consistent viewpoints.
    • Planning & Embodied AI: Can serve as a generalizable virtual environment for training multi-agent reinforcement learning or vision-language-action models.
    • Scalability: The parallel view generation and autoregressive capabilities allow efficient simulation of complex, long-horizon multi-agent scenarios.

Conclusion

MultiWorld presents a scalable framework for multi-agent, multi-view video world modeling, successfully addressing controllability, consistency, and scalability challenges through MACM and GSE. Extensive experiments on diverse datasets demonstrate superior performance in video quality, action adherence, and multi-view coherence.

Limitations: Current scale is limited by computational constraints; large-scale training remains unexplored.

Future Directions:

  • Investigate real-time multi-agent generation for downstream tasks.
  • Explore memory mechanisms for ultra-long multi-agent simulation to handle spatial/temporal demands.
  • Extend to even larger numbers of agents and views with optimized architectures.