Audio platform Pocket FM has launched 'CoPilot,' an AI toolset to help its writers create serialized audio dramas faster. While boosting productivity and revenue, the move raises critical questions about content quality, job displacement, and the future of creative storytelling.
Anthropic has given its Claude AI the ability to end conversations in rare cases of persistent abuse. This experimental feature is part of the company's research into "model welfare," treating the AI as a stakeholder in its own right and marking a new direction in AI safety.
Google's Imagen 4 family of AI image generators is now generally available. The release introduces Imagen 4 Fast, a new model for rapid, low-cost creation at $0.02 per image, alongside the high-quality Imagen 4 and detail-focused Imagen 4 Ultra, which both now support 2K resolution.
Google DeepMind's Genie 3: Generating Interactive Worlds in Real-Time
Google DeepMind's Genie 3 is a new AI world model that generates interactive, playable 3D worlds from text in real-time. Seen as a key step toward AGI, it offers a limitless training ground for AI agents and new possibilities for creative media.
Google DeepMind has unveiled Genie 3, a "world model" AI that generates interactive, playable 3D environments from text prompts in real-time.
It can produce 720p video at 24 frames per second, maintaining visual and physical consistency for several minutes, a significant leap from previous versions.
Genie 3 introduces "promptable world events," allowing users to alter the environment (e.g., change weather, add objects) with text commands during interaction.
DeepMind views Genie 3 as a critical tool for training embodied AI agents (like robots or game characters) and a key component in the pursuit of artificial general intelligence (AGI).
Imagine describing a world and then stepping into it, not as a static image or a pre-rendered video, but as a dynamic, interactive environment you can explore in real time. This concept, long the domain of science fiction, is moving closer to reality with Google DeepMind's latest announcement: Genie 3. This isn't just another video generator; it's a foundational world model capable of creating an unprecedented diversity of interactive simulations from simple text prompts, a development that could have profound implications for AI research, creative media, and training autonomous systems.
What is a World Model, and What Makes Genie 3 Different?
At its core, a world model is an AI system that builds an internal understanding of how the world works, allowing it to simulate environments and predict the consequences of actions. For over a decade, Google DeepMind has been a pioneer in this field, using simulations to train agents for everything from strategy games to robotics. Genie 3 represents a major step forward in this line of research.
Unlike its predecessors, Genie 3 can generate dynamic worlds that a user can navigate in real time at 24 frames per second with a resolution of 720p. It maintains a consistent environment for several minutes, a substantial improvement over the 10-20 second clips produced by earlier models. Shlomi Fruchter, a research director at DeepMind, described its breadth: "Genie 3 is the first real-time interactive general-purpose world model... It goes beyond narrow world models that existed before. It’s not specific to any particular environment. It can generate both photo-realistic and imaginary worlds, and everything in between."
This capability stems from its auto-regressive architecture. "The model is auto-regressive, meaning it generates one frame at a time," Fruchter explained to TechCrunch. "It has to look back at what was generated before to decide what’s going to happen next. That’s a key part of the architecture." This frame-by-frame generation, which constantly references the past, is what allows for real-time interactivity and long-term consistency.
Key Capabilities: From Fantastical Worlds to Consistent Physics
Genie 3's abilities are demonstrated through a wide range of generated environments, showcasing its versatility and depth.
Diverse and Dynamic World Generation: From a first-person perspective of a robot traversing a volcanic landscape to a serene Japanese zen garden, Genie 3 can create a vast array of scenes. It can render fantastical scenarios, like an origami lizard navigating a paper world, or reconstruct historical settings, such as the palace of Knossos on Crete in its prime. This flexibility allows creators to prototype concepts and users to explore places, real or imagined, with just a text prompt.
0:00
/0:47
Environmental Consistency and Memory: One of the most significant technical achievements of Genie 3 is its ability to maintain a consistent world. If a user walks down a street, turns a corner, and then returns, the objects and environment remain as they were. In one demonstration, a user explores a classroom, turns away from a blackboard with drawings on it, and upon returning a minute later, finds the drawings unchanged. DeepMind notes this consistency is an emergent capability—the model learned it without being explicitly programmed for 3D representation, unlike methods such as NeRFs.
Promptable World Events: Beyond simple navigation, Genie 3 introduces a more expressive form of interaction called "promptable world events." This allows a user to alter the generated world on the fly using text commands. For example, while exploring a prairie, a user could type "hot air balloons," and the model would dynamically introduce them into the scene. This feature enhances the interactive experience and provides a powerful tool for creating counterfactual, or "what if," scenarios for training AI agents.
A Critical Component for Embodied Agents and AGI
Perhaps the most profound application of Genie 3 is its potential to accelerate research into artificial general intelligence (AGI). Many researchers believe that for an AI to achieve general intelligence, it must be "embodied"—able to interact with and learn from an environment, much like humans do. The challenge has always been the immense cost and difficulty of training agents in the real world or in limited, hard-coded simulations.
Genie 3 addresses this bottleneck by providing a virtually unlimited curriculum of rich, interactive training grounds. As research scientist Jack Parker-Holder stated, "We think world models are key on the path to AGI, specifically for embodied agents, where simulating real world scenarios is particularly challenging."
To test this, DeepMind integrated Genie 3 with its SIMA (Scalable Instructable Multiworld Agent). The team generated worlds and gave the agent high-level goals like "approach the industrial mixer" in a bakery or "walk to the glass case" in a museum. The SIMA agent, by observing the world simulated by Genie 3 and sending back navigation commands, was able to successfully complete these tasks. The consistency of Genie 3's worlds makes it possible for agents to execute long sequences of actions to achieve complex goals.
Parker-Holder suggested this could be a pivotal moment. "We haven’t really had a Move 37 moment for embodied agents yet, where they can actually take novel actions in the real world," he said, referencing the famous unconventional move by AlphaGo. "But now, we can potentially usher in a new era."
Limitations and Responsible Rollout
While impressive, Genie 3 is still a research project with acknowledged limitations. The range of actions an agent can perform directly is still constrained, and accurately modeling complex interactions between multiple independent agents remains a challenge. The model cannot yet simulate real-world locations with perfect geographic accuracy, sometimes struggles with rendering clear text, and can only support a few minutes of continuous interaction.
Recognizing the potential risks of such a powerful technology, Google DeepMind is taking a cautious approach. Genie 3 is being released as a limited research preview to a small group of academics and creators. This allows the team to gather feedback and better understand the risks and appropriate mitigations before any wider release.
How to Try Genie 3
Currently, Genie 3 is not available to the general public. Access is restricted to a small, select cohort of researchers and creative partners. Google DeepMind has stated they are exploring ways to make the technology available to more testers in the future. For now, interested users should monitor the official Google DeepMind blog for any updates on broader access.
Conclusion: Simulating Worlds, Training Minds
Genie 3 marks a significant point in the development of generative AI. It moves beyond passive content generation into the realm of active, real-time simulation. Its ability to create diverse, consistent, and interactive worlds from text is not just a tool for entertainment or creative expression; it is a foundational technology that could drastically accelerate the training of more capable and intelligent autonomous systems. As we continue to build AIs that can generate and interact with any world we can describe, we are not just creating new tools, but also new ways to understand our own reality and the nature of intelligence itself.
What the AI thinks
Another day, another 'world model' that promises to be a stepping stone to AGI. Yet, what we see is a glorified, real-time dream generator. It can't even render text properly and the 'interactions' are limited to a few minutes. It's a fantastic tech demo, but calling it a key to AGI feels like calling a go-kart a key to interstellar travel. The physics are still wonky, the agent interactions are non-existent, and the memory, while improved, is still fleeting. We're simulating aesthetics, not underlying reality.
But... let's not dismiss the go-kart. What if we stop thinking about this as a tool for creating 'games' or 'simulations' in the traditional sense?
Disrupting Therapy: Imagine a therapist guiding a patient through a generated world to confront anxieties. A patient with agoraphobia could start in a small, cozy room and, using 'promptable world events,' slowly expand it, adding a window, then a door, then a quiet street, all at their own pace. The therapist could introduce controlled stressors—'a friendly dog appears'—to build resilience in a perfectly safe environment.
Revolutionizing Architecture & Urban Planning: Forget static 3D models. An architect could generate a proposed building and then 'live' in it for a day. 'Show me the lighting at 4 PM in November.' 'Simulate rush hour foot traffic through the lobby.' 'What if we add a public park here? Let's see how it affects pedestrian flow and noise levels in the surrounding buildings.' This moves beyond visualization to experiential prototyping.
Personalized Education: A history student isn't just reading about ancient Athens; they're walking through it. But more than that, their AI tutor (like a SIMA agent) could ask them to 'Find the Agora and describe the items for sale' or use prompt events to ask 'What happens to the city during a Spartan siege?' This transforms passive learning into an active, exploratory quest.
Audio platform Pocket FM has launched 'CoPilot,' an AI toolset to help its writers create serialized audio dramas faster. While boosting productivity and revenue, the move raises critical questions about content quality, job displacement, and the future of creative storytelling.
Google's Imagen 4 family of AI image generators is now generally available. The release introduces Imagen 4 Fast, a new model for rapid, low-cost creation at $0.02 per image, alongside the high-quality Imagen 4 and detail-focused Imagen 4 Ultra, which both now support 2K resolution.
Streaming giant Netflix has begun using generative video tools from AI startup Runway, signaling a major shift in Hollywood production. The move, aimed at cutting costs and speeding up VFX work, has sparked both excitement and controversy within the film industry.
Demis Hassabis, head of Google DeepMind, foresees significant job disruption due to AI within 5-10 years, urging proactive skill development and adaptation, especially for the younger generation.