The Google DeepMind team unveiled Genie 2, the second version of a fundamental AI model capable of generating new interactive digital environments, or game worlds, on the fly.
Image source: Google
To recap, the original Genie was released in February and could generate 2D virtual worlds from synthesized images. The Genie 2 is capable of doing this in 3D and based on text commands.
The user can describe the desired world, select a suitable rendering and step into a new environment. At each step, a person/agent performs an action (moving a mouse, pressing a key on a keyboard), and Genie 2 simulates its consequences.
Each example is based on an image generated by Imagen 3’s AI model based on a text prompt.
According to Google DeepMind, Genie 2 can generate sequential interactive worlds lasting about a minute, although most examples shown (see video below) last 10-20 seconds.
Compared to the first version of Genie 2:
According to Google DeepMind, Genie 2 demonstrates the potential of fundamental models of the world to create a variety of three-dimensional environments and speed up the training/testing of AI agents (like SIMA).
Google DeepMind clarifies that the research is at an early stage and requires significant improvements in the areas of agent capabilities and environment generation, but already sees Genie 2 as a solution to the structural problem of safely training AI agents.
Nissan Leaf can rightfully be considered a long-liver of the electric car market, since the…
OpenAI, the market leader in generative artificial intelligence systems, remains nominally a startup, its financial…
OpenAI has been forced to delay the release of ChatGPT's built-in image generator for free…
Xiaomi continues to update its Redmi G27Q gaming monitor every year. The model was first…
Android device makers can significantly customize the look and feel of the operating system, but…
In China, scammers have started selling GeForce RTX 3090 graphics cards, passing them off as…