Google's Most Advanced Robot Brain Just Got a Body
Google's Most Advanced Robot Brain Just Got a Body
8:11

Google's Most Advanced Robot Brain Just Got a Body

Tech
Speaker 1: Google wants to make robots smarter by teaching them to understand human language and then acting on it in the real world, melding the physical capabilities of walking roaming robots and giving them the kind of intuitive AI powers that you'd expect from a voice assistant or a smart speaker. It's a new technology called palms can, and it takes Google's smarts in natural language processing and machine learning and bakes them into robots, built by a company called everyday robots. And it's [00:00:30] something we haven't seen before. This robot doesn't need to be programmed with really specific instructions. Like if this, then that it can take vague instructions like I'm hungry or I'm thirsty, and then work out the steps it needs to take to solve that problem up until now, we've seen robots out in the real world doing park or, and really physical activities. And we've seen conversational AI driven voice assistance, [00:01:00] but now Google has combined the two. This is a huge deal for the future of robotics and human assistance. So we thought for this week's episode of what the future, we would try something a little bit different. I have my colleagues, Steven Shanklin here to tell me why it's such a game changer now, shanks, you and I were both at this Google demo. It was kind of impressive to see. Can you gimme the basic rundown of what Google was doing? Speaker 2: Sure. This is a technology called Palm Sayan and it combines two very different technologies. The first one is called [00:01:30] Palm, which is Google's very sophisticated, very complicated, natural language processing engine. So this is an AI system that's trained on millions of documents, mostly from the internet. And that is combined with the physical abilities of a robot. They have trained a robot to take a number of actions like moving around a kitchen, grasping objects, recognizing objects. They start with this language model. You can give some, you can give the robot a natural language command, like I've spilled my drink. I need [00:02:00] some help. The robot comes up with a number of possible actions, but then it grounds those possible actions and what the robot actually knows how to do so the marriage of the language model and the real world abilities is what's interesting here. Speaker 1: We saw these great demos of, um, a robot picking up a number of different like balls and blocks in different colors. And it knew that the yellow ball stood for the desert and the blue ball stood for the ocean. How is it recognizing those things? Speaker 2: This is what it learns from [00:02:30] the real world language information that it's been trained on. It knows sort of at a metaphorical level, that green means jungle blue means ocean and yellow means desert. So for example, by reading the novel dune, it can learn that the yellow desert, it might be a phrase that shows up somewhere so it can learn to associate these things. So it actually attains sort of a metaphorical reasoning level. That's much more humanlike than what we've seen in most robots, which are extremely literal, extremely, precisely scripted, and strictly programmed to do a very narrow [00:03:00] set of operations. So this is much more open ended. Speaker 1: Yeah. I remember with that hamburger demo, they showed us a couple of demonstrations of stacking blocks and bowls. But then I asked whether they could ask the robot to make a hamburger and it just picked up the pieces and put them in the order. It did put an entire bottle of ketchup in the middle of a hamburger, which was peak robot behavior. But I loved that. You don't actually have to say, put hamburger, Patty, put lettuce on top of hamburger, Patty, if lettuce then tomato, it [00:03:30] kind of just knows how to do that all at once. Speaker 2: Yeah. So a traditional industrial robot that's maybe installing windshield wipers or soldering capacitors onto a circuit board. That's a very specific, very scripted activity. This is very open ended. And because it's learned from this incredible wealth of knowledge, that's on the internet, it knows what the components of a hamburger might be. It was a pretty interesting demonstration and it was, it was not something that Google had planned out in advance. That was your random in the moment question. So this was, [00:04:00] you know, a good ex example, a good illustration of how this robot can, you know, be more improvisational. Speaker 1: We've seen plenty of robots before from the likes of Boston dynamics, you know, running over obstacles. Or I saw the Amicka robot at CES, which has this very humanoid face and was able to respond with natural language. But those are kind of examples of like physical, real world robot and then natural language in a kind of a humanlike suit, right? This is something that's quite different to [00:04:30] those. Speaker 2: One of the reasons such an interesting demonstration is it combines the brains and the bra. It's got the, it's got the AI language processing and it's got some physical ability to actually go out in the real world. The robots themselves were designed by an alphabet subsidiary called everyday robots. And they want to just build everyday robots that will show up in your house or your workplace. And so they they're designed to actually, you know, move around and grasp things and they have, you know, digital vision. And so with that combined with the Google framework [00:05:00] is, you know, something that's potentially more useful in the house, if they can actually, you know, develop this for another few years to get it out of the research, uh, lab and into your home. Speaker 1: Yeah. So, I mean, we've seen robots like say Astro from Amazon, which is a little home helper, you know, can bring you a can of Coke from the fridge and wheel it into your, into your bathtub. I saw that demo from our smart home team. What would be the future of this kind of robot in the home context compared to some of the other home helpers we've seen before? Speaker 2: If you look at a lot [00:05:30] of these other alternatives, it's, you know, kind of a smartphone with a bit of navigation glued on top. So, you know, Amazon asteroids, you know, it's impressive, it's a first step, but this is, you know, another level entirely when it comes to understanding what humans want and understanding what the robot itself can do. It's much more potentially open ended and therefore much more versatile. I guess, one of the interesting things here that, uh, I saw from the robot demonstration at Google is, is this is, uh, [00:06:00] designed for the chaos and unpredictability of the real world. If you compare it to Boston dynamics, they have very impressive physical, real world navigation abilities. You know, the Atlas robot can do parkour can do flips the spot dogs that can go up and down stairs deal with very complicated terrain. Um, but those don't really have a lot of abilities in terms of actually executing commands. They can go places, but they can't do things. The Google robot is a combination of going places and doing things. Speaker 1: Yeah. I feel like you're kind of combining [00:06:30] like the football team with the chess club into one robot. So if you think about where this goes in the future, maybe 5, 10, 20 years from now, what could the future of this kind of technology bring us? Obviously it's very early days, but it's pretty exciting, right? Speaker 2: Yeah. So what we've seen with the AI revolution is a, is a complete transformation of the computer industry from, uh, machines that could do a very specific task to machines that could handle really complicated, uh, real world situations. Some of those things [00:07:00] are very difficult, like driving a car in a street, incredible number of unpredictable events that could happen in that situation. But AI technology is good enough that it can start to deal with this really, really complicated landscape instead of something, you know, very limited like driving a shuttle bus down a track and back and down a track and back, right? So this is, this is what AI opens up. When you build that into a robot, it's very complicated. And, and you, I think you're, you know, 10 or 20 year time horizon is more likely what we're [00:07:30] looking at here. But when you combine that AI with this physical ability to navigate the real world and take actions, then that's potentially very transformative. Speaker 1: So there you have it, but I'm interested to know what you think. Is this the future of robotics or is it kind of terrifying or is it both because sometimes robotics and technology is like that. Let me know when the comments down below, and while you're here, throw us alike and subscribe for plenty more. What the future videos. We've got amazing stuff on robotics, flying machines, everything you [00:08:00] could possibly want. All right, until next time I'm Claire Riley for CNET bringing you the world of tomorrow today.

Up Next

If Apple Makes Siri Like ChatGPT or Gemini, I'm Done
240516-site-hey-siri-lets-talk

Up Next

If Apple Makes Siri Like ChatGPT or Gemini, I'm Done

How Many Times Did Google Say AI at I/O 2024?
Every AI Thumbnail

How Many Times Did Google Say AI at I/O 2024?

Everything Google Just Announced at I/O 2024
240513-site-google-io-supercut-thumbnail-v2

Everything Google Just Announced at I/O 2024

Google Introduces Gemini AI Upgrades to Gmail and Chat
google-io-gemini-gmail-chip

Google Introduces Gemini AI Upgrades to Gmail and Chat

Ask Photos Uses AI to Search Your Google Gallery
googlephotos

Ask Photos Uses AI to Search Your Google Gallery

Everything OpenAI Announced at Its Spring Update Event
openai-event-recording-13-may-2024-09-56-17-am-00-01-00-20-still006

Everything OpenAI Announced at Its Spring Update Event

Apple's iPad Pro Event Was a Flop
240509-site-omt-apple-flopped-this-ipad-event

Apple's iPad Pro Event Was a Flop

What to Expect at Apple's iPad Event
240501-yt-ipad-wildcards-v07.jpg

What to Expect at Apple's iPad Event

What AirPods Rumors Tell Us About Apple's Health Ambitions
240424-omt-next-airpods-v07

What AirPods Rumors Tell Us About Apple's Health Ambitions

What is the Fediverse?
240418-fediverse-winged

What is the Fediverse?

Tech Shows

The Apple Core
apple-core.png

The Apple Core

Alphabet City
alphabet-city.png

Alphabet City

CNET Top 5
top-5.png

CNET Top 5

The Daily Charge
the-daily-charge.png

The Daily Charge

What the Future
what-the-future.png

What the Future

Tech Today
tech-today.png

Tech Today

Cooley On Cars
on-cars.png

Cooley On Cars

Carfection
carfection.png

Carfection

Latest News

If Apple Makes Siri Like ChatGPT or Gemini, I'm Done
240516-site-hey-siri-lets-talk

If Apple Makes Siri Like ChatGPT or Gemini, I'm Done

Bose SoundLink Max Review: How Does It Compare to the Cheaper SoundLink Flex?
240514-site-bose-soundlink-max-2

Bose SoundLink Max Review: How Does It Compare to the Cheaper SoundLink Flex?

Hands-On With Huawei's Pura 70 Ultra
240515-winged-pura-70-ultra-00-01-30-05-still001.jpg

Hands-On With Huawei's Pura 70 Ultra

I Tried Three Fitness Apps to Help My Postpartum Recovery
cs-ai-coaching-seq-00-07-22-19-still001

I Tried Three Fitness Apps to Help My Postpartum Recovery

How Many Times Did Google Say AI at I/O 2024?
Every AI Thumbnail

How Many Times Did Google Say AI at I/O 2024?

I Tried Google's Project Astra
240513-site-google-project-astra-hands-on-v3

I Tried Google's Project Astra

Everything Google Just Announced at I/O 2024
240513-site-google-io-supercut-thumbnail-v2

Everything Google Just Announced at I/O 2024

Google Introduces Gemini AI Upgrades to Gmail and Chat
google-io-gemini-gmail-chip

Google Introduces Gemini AI Upgrades to Gmail and Chat

Google Brings Multistep Reasoning to Search
screenshot-2024-05-14-at-11-16-37am.png

Google Brings Multistep Reasoning to Search

Project Astra Revealed at Google I/O
screenshot-2024-05-14-at-1-42-51pm.png

Project Astra Revealed at Google I/O