DeepMind today unveiled a new multi-modal AI system capable of performing more than 600 different tasks. Dubbed Gato, it’s arguably the most impressive all-in-one machine learning kit the world’s seen yet. According to a DeepMind blog post: The agent, which we refer to as Gato, works as a multi-modal, multi-task, multi-embodiment generalist policy. The same network with the same weights can play Atari, caption images, chat, stack blocks with a real robot arm and much more, deciding based on its context whether to output text, joint torques, button presses, or other tokens. Credit: DeepMind And while it remains to be…
This story continues at The Next Web
No comments:
Post a Comment