Google DeepMind has announced Genie 3, a next-generation world model that can create interactive 3D environments from text prompts in real time. The model generates 720p environments at 24 frames per second, enabling users to navigate dynamic worlds with consistent physics and visuals for several minutes. This advancement builds on DeepMind’s decade of research in simulated environments and follows previous models, Genie 1 and Genie 2, as well as breakthroughs in video generation with Veo 2 and Veo 3.
Genie 3 expands possibilities in virtual simulations, modeling natural phenomena, vibrant ecosystems, historical settings, and imaginative animated worlds. It supports “promptable world events,” allowing real-time changes like weather shifts or new objects, and maintains environment consistency over longer horizons than its predecessors. These capabilities are expected to accelerate embodied AI research, allowing agents to train in rich, interactive virtual environments with extended goal-based tasks.
The technology could have applications in robotics training, education, generative media, and research toward general AI. However, DeepMind acknowledges limitations, including constrained agent action spaces, difficulty simulating multiple independent agents, imperfect geographic accuracy, challenges with text rendering, and limited continuous interaction duration.
Due to its powerful and open-ended nature, Genie 3 is being released as a limited research preview to select academics and creators to gather feedback and explore responsible development. DeepMind emphasizes its commitment to safe AI deployment, highlighting collaboration with its Responsible Development & Innovation Team to address potential risks.
Genie 3 marks a step toward immersive, AI-generated worlds with real-time interactivity and controllability, potentially transforming how agents and humans interact with virtual environments. DeepMind plans to expand testing access and explore broader applications in training, evaluation, and creative fields while continuing to assess the societal impacts of such technology.