By demonstrating transferable learning across unseen 3D worlds, DeepMind’s CIMA 2 brings AI closer to human‑level adaptability, unlocking new business opportunities in simulation, robotics and gaming while reshaping the competitive landscape of general‑intelligence research.
DeepMind unveiled CIMA 2, a multimodal game‑playing AI that learns to navigate a suite of modern 3D environments using only raw pixel data and standard keyboard‑mouse inputs, extending the field beyond the Atari‑centric agents of a decade ago.
The system ingests voice commands, hand‑drawn sketches, emojis and even reverse‑psychology prompts, planning actions up to about ten seconds ahead and transferring skills from one title to another. In head‑to‑head tests on an unseen game—Minecraft—the success rate leapt from near 0% with CIMA 1 to roughly 14‑20%, and the agent also adapted to procedurally generated worlds generated by other AI models.
Demo footage showed the AI locating distress signals, building a camp, and navigating novel art styles after a brief practice period, while the presenter highlighted quantitative gains and candidly noted limitations such as modest absolute success rates and the use of sped‑up video for illustration.
The breakthrough points to agents that acquire generalizable, curiosity‑driven competencies, foreshadowing commercial applications in simulation‑based training, autonomous robotics, and interactive content creation, and it reinforces DeepMind’s leadership in the race toward more human‑like artificial intelligence.
Comments
Want to join the conversation?
Loading comments...