Reasoning Robots: Google DeepMind and Boston Dynamics Evolve Spot’s Intelligence

Boston Dynamics and Google DeepMind are integrating Large Language Models (LLMs) with the Spot robot. This allows the quadruped to 'reason' through tasks and respond to natural language commands in real-world environments.

Share
Reasoning Robots: Google DeepMind and Boston Dynamics Evolve Spot’s Intelligence

From Code to Conversation: DeepMind Gives Spot a Brain

In a landmark collaboration, Boston Dynamics and Google DeepMind have demonstrated a version of the Spot robot that can reason using natural language. For years, robots were programmed with rigid scripts—if x happens, do y. This new integration allows Spot to interpret vague human instructions, such as 'find the leak and notify the technician,' by breaking the goal down into sub-tasks autonomously.

DeepMind’s visual-language models allow the robot to perceive the world not just as a 3D map of obstacles, but as a collection of semantic objects. When Spot 'sees' a fire extinguisher, it now understands what it is and its potential utility in a safety context. This 'Reasoning' layer sits atop the 'Athletic' layer that Boston Dynamics has perfected, creating a machine that is as smart as it is agile.

The implications for industrial robotics are profound. Instead of needing a specialized engineer to write thousands of lines of code for every new task, a warehouse manager could simply 'talk' the robot through its new duties. This move toward 'Socially Assistive' and 'Reasonably Autonomous' robotics marks the transition of robots from simple tools to collaborative agents capable of navigating the unpredictability of human-dominated spaces.


Source: IEEE Spectrum