On August 12, 2025, Skywork AI released Matrix-Game 2.0, an open-source "interactive world model” that generates video in real time while responding to a player’s keyboard and mouse. The code and model weights are public under the MIT license. The aim: make long-horizon, controllable AI worlds something you can run, not just watch.
Unlike earlier generative video demos that produced short, non-interactive clips, Matrix-Game 2.0 is built for streaming play: it targets ~25 frames per second and can sustain minute-level sequences, binding user inputs directly into each predicted frame. The release lands a week after DeepMind’s Genie 3 preview and positions Skywork’s model as a fully open alternative for researchers and builders.
Code & weights are public under MIT. Labs and builders can fork, replicate, and extend without waiting on approval.
Keyboard/mouse signals are embedded per frame, steering each predicted step in real time.
Documents the data pipeline (Unreal + GTA5, ~1,200 hours), dependencies, and quick start. Designed for reproducibility and benchmarking.
Targets ~25 FPS streaming and minute-long sequences; throughput varies by GPU and setup (FlashAttention + NVIDIA Apex recommended).
Access is limited; external teams typically cannot run weights locally or modify internals.
Interactive, but details emphasize higher-level control; explicit frame-level action injection is not the primary focus in public materials.
Shares high-level research framing; limited documentation on data pipeline, dependencies, or runnable quick start for the public.
Coverage highlights real-time interactivity (often cited around 720p/24 FPS in media); specifics may vary and are subject to change.
#!/bin/bash
# Clone repo
git clone <REPO_URL>
cd matrix-game-2.0
# Create env & install
python -m venv .venv && source .venv/bin/activate
pip install -r requirements.txt
# Install performance deps (GPU)
pip install flash-attn --no-build-isolation
pip install ninja git+https://github.com/NVIDIA/apex.git
# Run: batch video
python inference.py \
--config configs/base.yaml \
--ckpt checkpoints/model.safetensors \
--vae checkpoints/vae.safetensors \
--input prompts.txt \
--out out/
python inference_streaming.py \
--config configs/stream.yaml \
--ckpt checkpoints/model.safetensors \
--vae checkpoints/vae.safetensors \
--resolution 720 \
--target_fps 25
# Genie 3 — public quick start is limited # Typical path: request access; follow provider docs / SDK when available.
inference.py for batch video or inference_streaming.py for real-time output (provide config, checkpoint, VAE path).Watch for independent evaluations on the GameWorld benchmark, community ports that improve UX, and experiments coupling Matrix-Game 2.0 with agent training loops. If the open baseline holds up, expect rapid iteration—fine-tunes for domains, better latency-resolution trade-offs, and toolchains that make "AI worlds” a routine lab artifact