Runway Launches Physics-Aware World Model for Video and Robotics
Full Transcript
Runway has launched its first world model, GWM-1, which operates through frame-by-frame prediction, allowing it to create simulations with an understanding of physics and real-world behaviors. The model is designed to facilitate reasoning, planning, and action without requiring exhaustive training on every real-life scenario.
Runway's CTO, Anastasis Germanidis, emphasized that the development of a robust video model was crucial for building the world model, stating that teaching models to predict pixels directly is essential for achieving general-purpose simulation.
The GWM-1 is considered more general than Google's Genie-3 and aims to support simulations for robotics and life sciences. Runway has introduced specific variants of the GWM-1, including GWM-Worlds, GWM-Robotics, and GWM-Avatars.
GWM-Worlds allows users to create interactive projects with simulated environments that incorporate geometry, physics, and lighting at 24 frames per second and 720p resolution. GWM-Robotics focuses on using enriched synthetic data to train robots, considering varying conditions and obstacles, while GWM-Avatars aims to simulate realistic human behavior.
Additionally, Runway has updated its Gen 4.5 video model to incorporate native audio and multi-shot generation capabilities, enabling users to create one-minute videos with consistent characters, dialogue, and background audio.
The Gen 4.5 model update brings Runway closer to competitors like Klings and positions video generation tools as production-ready. GWM-Robotics will be available through an SDK, with ongoing discussions with robotics firms for potential applications.