The Google DeepMind team unveiled Genie 2, the second version of a fundamental AI model capable of generating new interactive digital environments, or game worlds, on the fly.
Image source: Google
To recap, the original Genie was released in February and could generate 2D virtual worlds from synthesized images. The Genie 2 is capable of doing this in 3D and based on text commands.
The user can describe the desired world, select a suitable rendering and step into a new environment. At each step, a person/agent performs an action (moving a mouse, pressing a key on a keyboard), and Genie 2 simulates its consequences.
Each example is based on an image generated by Imagen 3’s AI model based on a text prompt.
According to Google DeepMind, Genie 2 can generate sequential interactive worlds lasting about a minute, although most examples shown (see video below) last 10-20 seconds.
Compared to the first version of Genie 2:
According to Google DeepMind, Genie 2 demonstrates the potential of fundamental models of the world to create a variety of three-dimensional environments and speed up the training/testing of AI agents (like SIMA).
Google DeepMind clarifies that the research is at an early stage and requires significant improvements in the areas of agent capabilities and environment generation, but already sees Genie 2 as a solution to the structural problem of safely training AI agents.
Apple is preparing to launch updated 13- and 15-inch versions of the MacBook Air laptop,…
The VideoCardz portal writes that AMD held a closed briefing for journalists this week, where…
Bonn, Germany, is in dire need of kindergartens, so they are sometimes placed in the…
According to online sources, Apple will focus more on improving video recording in the new…
It was previously reported that some GeForce RTX 5090/RTX 5090D graphics cards, and as it…
A group of researchers from China has developed a technology that will restore the capacity…