StarTalk Radio - Inside DARPA: Sci-Fi Meets National Defense
Episode Date: August 3, 2018Explore the intersection of science fiction and national defense inside the high-tech government agency developing America’s top-secret weapons of the future with Neil deGrasse Tyson, co-host Chuck ...Nice, former DARPA Director Arati Prabhakar, journalist Sharon Weinberger, and roboticist Hod Lipson.NOTE: StarTalk All-Access subscribers can listen to this entire episode commercial-free here: https://www.startalkradio.net/all-access/inside-darpa-sci-fi-meets-national-defense/Photo Credit: J.Krohn, courtesy of JPL-Caltech. Subscribe to SiriusXM Podcasts+ on Apple Podcasts to listen to new episodes ad-free and a whole week early.
Transcript
Discussion (0)
Welcome to StarTalk, your place in the universe where science and pop culture collide.
StarTalk begins right now.
Welcome to the Hall of the Universe.
I'm your host, Neil deGrasse Tyson, your personal astrophysicist.
Tonight, we will explore the intersection of science fiction and national defense
inside the high-tech government agency developing America's top-secret weapons of the future.
So let's do this.
Yes!
Oh! Yeah! Chuck! I got with me, of course, Chuck Nice. Hey, hey, Neil. How are you?
Chuckie baby.
Good to see you, buddy.
Feeling good?
Always, always.
Always. Got your Let's Make America Smart Again shirt.
This is my Let's Make America Smart Again shirt, which quite frankly is a very cool shirt, and we got with us tonight
Sharon Weinberger hello
So
So you're the author of this fact the people still write and read fat books they write them. I hope they read them
fat books? They write them. I hope they read them. So this is the Imagineers of War, the untold story of DARPA, the Pentagon Agency that Changed the World. So DARPA stands for what? What is DARPA?
DARPA stands for the Defense Advanced Research Projects Agency. And so you're a journalist,
you follow this industry for a long time? Yeah, I've been writing about Pentagon-funded science
and technology for about the past 17 years.
Whoa, oh, oh, okay.
And this book is the product of about
five years of interviews, research, and writing.
Well, you must do it well because you're still alive.
Right.
I tried.
Just saying, just saying.
No, she was undercover until this moment.
Okay.
Ready to go public.
So tonight we're featuring my interview
with the former head of DARPA, Arati Prabhakar.
And we met during her tenure as chief of DARPA in Washington, D.C.
So let's check it out.
Growing up, to me, DARPA was always this mysterious, you know, what are they doing there?
I don't know.
But it's really cool.
One day we'll find out.
Maybe.
Is that the mission statement?
Did I just capture that?
Well, as long as it moves the needle, as long as it really changes the world, and especially
national security, then yeah, that's what we do.
DARPA is an agency that is very, very small.
We're 200 government employees, right?
And all of us are here in this office building in Arlington,
Virginia. We don't... They're sitting on the offices right above my head. Right above our heads.
We don't have labs. We don't have any fixed infrastructure. We are critically dependent
on all the places and people that do. Okay, so I looked up your budget. It's about $3 billion.
It's about $3 billion. About $3 billion a year. And you just give away all that money? No. No.
By the way, give and money just don't make sense together in a sentence,
just to be clear. What we do is we bring in program managers, really smart people in their
fields. They come in for typically about three to five years. A tour of duty kind of. Tour of duty.
They are here for a short time. What they come in to do is to craft a program. It might be,
can I make a ship that can sail without any sailors on board? It might be something about a next generation of artificial intelligence.
It might be about biological technologies, whatever area.
They want to know, what are all the really creative, what are the crazy sort of mind-bending ideas?
And they listen to all of that.
They go out and they'll talk to the user community and the DOD, for example, and especially.
But between thinking about where the big problems are and where the science and technology could take you,
that listening is a key part of how they come up with that vision.
So we are very interested in people's big new ideas.
So I got my list here.
Okay, I want a wormhole.
Okay.
Big national security implications for that one. I want a transporter. Okay. Big national security implications for that one.
I want a transporter.
Yeah, don't we all.
The flying car.
So, when are we going to get these things?
A few more years.
Okay, a few more.
There was a DARPA project called the Hundred Year Starship
that looked at wormhole travel, so...
Really?
A hundred years, maybe.
That's optimistic.
That's certainly not a waste of money.
And I like the way you put it, a hundred years, so none of us will be here to bitch about it.
So, okay, you also had robotic soldiers in there and, like, force fields and mind control. stuff. These things could be right. They could and the ideas come back again every few decades the idea of a force field that can
can protect the planet the idea of mind control all these things. This is part of the reputation. Yes. The
spooky mysterious reputation that DARPA has
mysterious reputation that DARPA has sustained over all these decades. Yeah, spooky. A lot of its work has been secret, has been classified. It's also at
times sort of touted itself or touted by others as a science fiction agency, some
of which is overblown, some of which is true. So you're men in black basically.
That's what you are. Okay, so what are the origins of DARPA? DARPA dates back to 1958.
And basically in the fall of 1957, the Soviet Union...
That was an important year.
I mean, NASA got funded.
That was the geophysical year.
A lot of science, tech stuff was coming together.
It was coming together, and a lot of it was prompted by the October 1957 launch of Sputnik.
This is the Soviet Union's launch of the first artificial satellite.
And it's hard to imagine now, but it was sort of a 9-11 moment. It created this real political
panic in the country that we were losing the space race. The Soviet Union would have
intercontinental ballistic missiles that would reach the United States.
Okay, so we said, all right, let's fund this agency with what's the mission statement?
To do whatever the Secretary of Defense directs it to do.
It was very vague,
except it was supposed to take on space programs.
It was before NASA.
It was the nation's first space agency.
And that's not a scary mission statement at all.
Not at all.
Hey, whatever the Defense Secretary says do, we should do.
All right, so it's also, some have considered DARPA
the place where mad scientists go and are given free reign.
So what are some of the craziest, maddest things to come out of there?
So one of the original mad scientists of DARPA was a scientist, a Greek scientist named Nick Christophoulos,
who had these, like, fantastic ideas.
So he would be an immigrant?
He was.
Yes, okay.
Exactly.
And he was genius, and scientists loved him, DARPA loved him.
So one of his first ideas was a force field.
You know, we were worried about Soviet missiles attacking the United States.
So he said, what if you launched a bunch of nuclear weapons in the upper atmosphere
that would release killer electrons and fry anything coming through it?
It would be a planetary force field. That was one idea.
Oh, that's great.
Yes.
What could go wrong with that?
Let's nuke the sky.
Yes.
To create a force field.
To create a force field.
Yeah, yeah. What could go wrong?
Yeah.
We had another fascinating idea of creating a particle beam weapon.
Same thing, to take down Soviet missiles.
You would just need to drain the Great Lakes to power it all at once by...
Who needs them?
Yeah. You would put nuclear weapons under the Great Lakes to drain them Great Lakes to power it. All at once by... Who needs them? Yeah.
You would put nuclear weapons under the Great Lakes
to drain them, to power the project.
So these are out of the box ideas?
They were very out of the box.
You know, I'm starting to believe
that DARPA might stand for
Drugs Are Really Pretty Awesome.
Oh, that's right.
All right.
And another one in my list here.
How about Project Pandora?
What is that?
So Project Pandora was a top secret project.
We should be cautious of anything called Project Pandora.
Yeah, they chose that name appropriately.
So in the 1960s, the CIA discovered that the U.S. Embassy in Moscow was being irradiated.
They were sending low-level pulsed microwaves.
And this was the time when there was...
Who was radiating who?
The Soviets, the Russians, irradiating us, the Americans. With microwaves? In our embassy in Moscow. Yes, with microwaves. And this was a time
a lot of literature was coming out that maybe pulsed microwaves could alter human behavior,
affect the mind. So the CIA said, oh my God, the Soviets are trying to control our diplomats' minds.
And so they... Because we didn't know better. We didn't know better. When really they were just slow cooking us.
Yeah.
So they assigned DARPA to test this.
You know, could microwaves be a mind control weapon?
Right.
So those things didn't quite pan out as people had perhaps hoped. We lack psychic super weapons, yes.
Or do we?
But other programs have panned out famously
and are responsible for some of our most sophisticated military innovations.
So I asked the DARPA head about the intersection of technology and the military
and what role that plays.
So let's check it out.
There are two sides of DARPA because, again, our job is for national security.
So one facet of it is these core underlying technologies like networking that you have to have if you're going to build sophisticated military capability.
And the other is military systems that we demonstrate, like this thing called Stealth Aircraft, which also started as a DARPA project.
And the whole idea of precision strike, where instead of using massive weapons that wipe out everything, you develop the very sophisticated ability to find a very specific target, communicate back, and deliver a missile to precisely that location. Rendering nuclear weapons completely absurd in the face of that ability.
Well, everything up to and including nuclear weapons was about greater mass
and greater destructive power. Precision said, we want to hit exactly what we want to hit,
and we can be, really, it was driven out of the idea that, again, back to the Soviets. The Soviets,
we knew, had more forces, and they had tactical nuclear weapons, and we said there has to be a different way to fight that kind of fight.
So stealth and precision weaponry, somebody has to think that up and decide that that's a thing to do. So what prompted DARPA to go there? Well, ironically, it grew out of our most
failed war effort, which was Vietnam. DARPA had been very heavily involved in Vietnam and
developing things like a quiet aircraft for Vietnam
for counterinsurgency, precision weapons for bombing targets.
And when the Vietnam War ended, the Soviets had been modernizing.
We hadn't, and DARPA figured out a way to sort of repackage it
for the European battlefield.
So, Sharon, we have an image of the first practical stealth aircraft
developed by DARPA back in the 70s. Let's check it out.
Yeah.
So the whole concept is whatever radio signal goes forward and hits it, it should deflect
somewhere but not back from where the signal came.
And that way, it's as though nothing is there.
Because you only know something's there not by the signal you send out, but by the strength
of the signal that comes back.
And if you angle all the sides in such a way that it disperses the signal,
you can make this thing look as small as a bumblebee on a radar return image.
You just explained why my children do not listen to me.
Just deflect them everywhere.
Every which way.
But you know what happened?
So that was with the 70s.
When was the first stuff?
That was in the 1970s.
The 1970s.
So our computing power was pretty low then relative to later on.
And so to get the perfect shape that doesn't reflect anything back forward required formula
fitting and calculating that the computers of the day were incapable of solving.
And so as computers got more powerful, you could take these flat surfaces,
they create a perfectly curved surface
where every spot knows exactly how to reflect it
so that nothing comes back this way.
And that gets you the B-2 bomber right here.
Aha!
Now, so...
I see what you're saying.
This is still, to me, the Batplane.
I love it.
Right.
So each of those surfaces, there's no surface that'll reflect a radio signal back to where you're coming from.
So not all DARPA projects become weapons, even if they were wanted to be conceived as such or anything.
Not all become weapons.
Some become other stuff, right?
This is a fascinating feature of this.
fascinating feature of this and of course DARPA research laid the foundations of some of the most valuable non-military technology that influences
civilization today and I asked the former head of DARPA all about this get
me a rundown of the greatest hits of non-military successes let's check it
out we're in the late 60s. Think about what computers were
like, right? I mean, these huge mainframes and punch cards, you interacted with them.
We've been there, right? Don't get me started.
But even in that time, people were imagining what might be possible if you could connect
computers together and they could share information and programs and tools to do our work.
I saw that movie. It's called Terminator.
Well, in the 60s, think about vacuum tube computers and going to punch out your cards.
Imagine being able to think about that.
That was really the spark.
In the late 60s, ARPA at that time began a program called the ARPANET. And the
challenge that they threw out to the community was, let's connect some computers together,
see if we can actually get them to communicate so that you could start sharing information.
And how to get them to know what the protocols are for the communication.
Right. And so the internet protocols, TCPIP, which is how this massive scaled-up internet still works today,
those were invented by someone who was a program manager at DARPA. And, you know, it was just,
those were the baby steps, and look how far that baby has gone.
Wow. Okay.
So let's move into the 80s and 90s. I came to DARPA as a young program manager 30 years ago,
in 1986.
Oh, so you've got genetic connections here.
Yeah, I got to start my career here pretty much,
and that was totally awesome.
So some of the things that we did in that period,
so if you think about what's inside your smartphone,
think about those integrated...
I have a lot of photographs.
Yeah, well, that's good.
So those are possible because you have
very sophisticated integrated circuits,
and some of the early roots of that trace back to DARPA.
When your cell phone talks to the cell phone tower,
it's using a chip made out of gallium arsenide.
It's a radio semiconductor component
that comes directly out of major DARPA programs
that we started because DOD needs radio communications
and radar systems.
Siri traces back to a DARPA project
that got spun out of SRI as a company called Siri
that got acquired by Apple and became the product.
So all of those technologies, but, you know,
this is the magic, right, is public dollars
sparked all of those technologies
that, you know, changed how we live and work.
So, Sharon, she's taking credit for the internet, smartphones, and Siri.
Yes.
Is she all that?
Almost all of that.
So the biggest one in what cemented DARPA's reputation was network computers, the ARPANET,
which was fully and wholly a DARPA innovation, which went directly into the internet.
Thank God they changed the name from ARPANET, because that sounds like the internet for
really old people.
ARP?
How do you work with that?
She's absolutely right that Siri also came directly out of a DARPA program.
It turned out the military wasn't interested, so they spun it off.
Apple bought it.
It's in your iPhone.
So Sharon, what are the actual origins
of ARPANET within DARPA?
Because I read that it had something to do
with linking together the few
survivors after the nuclear holocaust.
So, is that true?
It's mostly untrue.
Mostly?
Well, no, I mean...
This is for Armageddon.
Well, the other myth was that ARPANET had nothing to do with nuclear weapons.
It was just scientists that wanted computers to talk to each other.
And that's a little bit of a myth, too.
So what it was was the Pentagon was worried about command and control of nuclear weapons.
So they asked DARPA, can you look at this issue, command and control of nuclear weapons?
DARPA hired a scientist who didn't care about nuclear weapons.
He wanted computers to talk to each other. So it was both coming together. Well, coming up next, we're going to explore how
DARPA stepped into my world, my world of space. And we're going to see how they had a project to
try to protect America's national security in space when StarTalk returns.
Welcome back to StarTalk from the American Museum of Natural History
right here in New York City.
We're featuring my interview with the former head of DARPA,
America's Advanced Defense Technology Agency.
Let's check it out.
What's this about a space surveillance telescope?
Well, the Defense Department cares about space a lot.
Which oven is that in that you're baking?
That's one that's pretty baked.
We just handed it off to the Air Force.
Not half-baked, it's fully baked.
Well, for DARPA, it's fully baked.
You'd have to ask the Air Force, but I think it's making a real difference. The DOD issue for
space is the following. We can do nothing without space. We need it to communicate. We need it to
navigate with GPS. We need it to be able to see what's going on around the world. It's essential
to us. But space is a very complicated domain. It's not like the 50s and 60s and 70s where we were the only ones doing anything.
The only players, yeah.
So these are sensitive telescopes on Earth looking up that can monitor all the activity at 23,000 miles up, geosynchronous.
And there's a lot of commercial activity, which is wonderful.
Other nations are getting active, unfortunately not always in a constructive way.
So we're in a time when instead of orbital catalog maintenance,
where you just have a catalog and you know what's on orbit
and you sort of pay attention once in a while,
now in real time you have to be able to see and know what's happening on orbit.
And the vastness of that job is hard to imagine, right?
I mean, think about several hundred thousand times
the volume of all the Earth's oceans.
That's what you have to be able to look at
and see what's going on.
So Sharon, how many satellites are up there, spy satellites?
Oh, so that's classified.
I think there have been 276 missions total, many which include spy satellites, some which are no longer active.
I think that we know that there are about maybe half a dozen spy satellite constellations, but the exact number is classified.
All right. So, and how about other things that are space-based, like the X-37 plane? Tell us about that.
Yes, so it looks...
Tell us what you can about that.
So it is a very classified project.
It's the X-37B orbital test vehicle.
And it looks a lot like the shuttle,
you know, the manned shuttle,
like a mini-shuttle, a mini-me of the shuttle.
And it has flown in space for up to 700 days,
and the Air Force and the military
has never acknowledged exactly what it's being used for.
Somehow...
Wait, wait, just to be clear, just to be clear.
A satellite, once it's in orbit,
that's the orbit it has forever,
until you put it somewhere else with whatever fuel you have.
Decays and dark.
Whereas a space plane...
A space plane can maneuver.
You could put it over different spots of the Earth on command.
Gotcha. Up next, I ask the former head of DARPA A space plane can maneuver. You could put it over different spots of the Earth on command.
Up next, I ask the former head of DARPA about the future of robotics and AI when StarTalk returns.
Welcome back to Startoff.
We're talking about America's advanced defense technology programs.
And joining us to help us in this conversation is AI and robotics expert, Hod Lipson.
Hod, welcome.
Thank you. Pleasure to be here.
So you're a professor of mechanical engineering, and you make machines that create.
Is that a fair characterization?
That's right.
Machines that are creative.
Trying to break that myth that machines can't actually be creative.
So these are self-aware robots.
To some extent, yeah.
That scares the hell out of me.
Okay.
Well, as you know, I sat down with former head of America's Advanced Defense Technology Agency
DARPA and I asked her where we're headed with the future of robotics and AI so we'll get her take on it We'll come back to you. Let's check it out. I
Think for a lot of people the word robot conjures up a humanoid robot. I think that's a little bit different thing
I try to disavow people that that. Because human body, why?
There's nothing, why?
Right, we can do that stuff.
Yeah, we're not some model of anything, right?
Just do better, do a better job.
And different.
Because when machines do things that humans can't do,
that's when we can do things together that are really amazing.
Rather than do just what a human can do.
But artificial intelligence, as you know very well, I think people are very aware now,
it's having this resurgence in this new wave of artificial intelligence with machine learning
at its core. Now this idea that machines can take, you know, look at a million images and
start figuring out how to label them and do it better than humans because of how they've been trained.
Once they beat us at chess and then at Jeopardy, we should just give up.
Right, and go.
Well, we might give up on those things.
There are a lot of things that they know we're close to.
So we're good.
Really?
Because after they beat us at Jeopardy, I said, there's nothing left.
We're done.
Let's just move to Mars.
And I'm looking at my photo sorter on my computer.
Yeah.
And you give it face recognition, some hints, and give it a few photos to train on.
One of these photos found me when I was 11.
Isn't that amazing?
When I was 11.
I bet.
I could picture the 11-year-old you.
I kind of resembled myself, of course, but still.
That's pretty amazing.
It says, we think this is you.
Please confirm.
It's like, yeah, it's me.
It's a fuzzy photo taken on a Kodak Instamatic.
Yeah.
Oh, my gosh.
So that's what this learning era of AI is really all about.
It's in everything that we are doing.
So, Hod, you're an AI roboticist.
What is that? Well, robotics is this merger of
AI and the physical world, the mechanics, the embodiment of a machine. And so you need to sort
of understand both the mechanics and the AI to make a physical, smart system. So what do you
create in your lab? I love asking that of anybody who's got a lab.
What do you create? What's happening in your lab? What's happening? So we're trying to look at sort
of next generation robotics. We're trying to look at how can we take this from program machines to
machines that can learn, that can learn fast, that can learn things that we haven't taught them,
and do things that we normally don't think robots can do, like can learn fast, that can learn things that we haven't taught them,
and do things that we normally don't think robots can do, like self-replicate, be creative,
things like that. They have self-awareness? Self-awareness is one of these holy grails,
I would say, of robotics. It's going to take a while, but we're on that path.
There's a robot on this, there's some creature on this table.
Yes.
It's a little creepy.
It is.
So what is this?
So this is one of our platforms, one of our robots where we study self-awareness.
We develop our algorithms for self-awareness.
We put them on this robot and we see how well they do.
So you upload?
We upload it and we see, you know, can this robot sort of become self-aware in a very primitive sense, but can it sort of understand what it is?
What it is in itself.
In itself.
How exactly could it understand what it is?
Because what is it?
We don't know what it is. Look at it.
Well, you know, so it starts off, but if you press the button, you can see how it works.
Press the green button.
Press the green button.
Green is good, right?
If it's green, that's a good button.
Yes, it is, Morpheus.
Are you sure?
Press the green button or the red button?
Just press this.
You're sure?
You told me to press it.
Go ahead.
All right, so it's going to take a while.
It's making a noise right now.
It's going to take a while, but what it does is that it begins to sort of,
will soon start moving and having actuations and sensations.
It's going to have all kinds of sensations about itself,
and it's going to start forming what we call a self-image.
It's going to, in a very primitive way, begin to learn that it has four legs.
It's going to learn that I'm an ugly...
This robot is actually blind.
It has no ability to look at...
All righty.
Sufficiently creepy.
It has...
So what's happening?
So what's happening, it's moving around
and it's sensing all kinds of things about itself.
About how it moves, how it accelerates, what are the torques in its motors.
Can it sense that it's on a table and a few feet off the ground?
It doesn't know that.
But if one of the tips goes over the edge, will it pull it back off the edge?
No, it doesn't know that.
Anything, it's like a baby starting to learn what its body is.
And, you know, it takes it about four days.
It looks like it's dancing now.
I was about to say, I'm not a roboticist, but it looks horny.
It's throbbing.
I'm just saying.
I'm pretty sure it's aware of something right now.
So what happens four days from now?
For four days from now, if you look inside.
It's all hostage.
It pulls out the gun.
It takes over the world.
Where's the off button?
Oh, you just killed it.
Don't go to sleep.
It's just peeping me out on the table here.
That's only because it's coming to get you later.
It's going to remember what you did.
Sorry.
Yeah. So, you know, you want to have... That's only because it's coming to get you later. Sorry.
Yeah.
So, you know, you want to have, for some applications, you want robots that look like humans so that they can interact with humans or so that they can use human tools, the tools designed for humans.
Why would you do that?
For example, I can imagine 40 years ago we would have thought, how would we have a self-driving car?
You wouldn't.
You would have a robot that drove in the car the way a human would drive the car.
No one is thinking, let's have the car drive its own damn self.
Right.
Actually, when you look at a lot of science fiction movies, they still have a robot driving the car.
Exactly.
Exactly.
So clearly we can just bypass the middleman.
Absolutely.
We can.
Right. So to say, design a robot to use tools the way a human does, isn't that once again not bypassing the middleman?
Right.
Aren't we the weak link?
Sometimes we want a robot to communicate with a human. A human really needs to communicate with something that looks like another human. So Sharon at DARPA, have they been thinking about humanoid robots?
Is that a thing? Yes, I mean they've been thinking about artificial intelligence going
back to the 1980s and then a couple years ago they did this contest of sort of humanoid
rescue robots to make them seem a little bit friendlier. Was this the robotics challenge?
Yes, exactly. Okay, so we've got a demo clip of one of the teams doing some testing with one of their sort of humanoid robots.
Let's see if we can bring that up. so hot if I were that robot, I would be pissed off.
I would come around and be ready to kick this guy's ass.
You see, it's not self-aware yet.
Oh.
If you keep it from being self-aware...
It's just a machine.
Now, this is in Boston.
That's why I used a hockey stick to poke him.
Yeah, exactly.
That's all I know. Now this is in Boston, that's why I used a hockey stick to poke him. Yeah, exactly.
That's all I know, it is hockey in Boston.
And where can I get one of these Kick'em robots?
Because you know, I got a lot of pent up frustration.
The ones that are not self-aware.
Yeah, I don't want a self-aware one.
Because that thing, you know that, come around.
Wait, wait, wait.
So why are they abusing this robot?
Because we were feeling for the robot.
Even though it's not human, we were kind of like, that's not nice.
Even though that's a complete machine.
You know, one of the things we did with our robots is chop off its leg, and we watched what happened.
And we could see that after a while, it sort of learns that it has no legs, and it begins to limp.
And it's a way of testing.
Testing what?
What, a dick you are?
What kind of testing is that?
We test how well the robot can sort of understand what's happening, can remodel itself.
Gotcha.
How resilient is that? So DARPA, the military might deploy this
in a kind of disaster response situation, right?
Might you equip that with a gun?
Although it's hands were these rubber pads.
Presumably, if it could...
You see the panache with which it stood back up?
Yes. So it's like, I'm ready for you. how the panache with which it stood back up? Yes.
So it's like, I'm ready for you.
So, you know, we all remember the movie Robocop.
The military isn't really a rescue machine.
I mean, yeah, I think that eventually robo-soldiers is the eventual idea.
Right.
No, it can go into a dangerous situation and bring out fellow soldiers, for example.
It could, yes.
Well, that's cool. Right. They're bring out fellow soldiers for example it could yes well that's cool right
they're bringing up bringing some guns that little spin you know put in its holster that kind of
thing but again this would be guns that a human would use but why train a robot to use a human
gun when you can just build a gun into the chassis itself right absolutely see he i don, his lab, I don't know what he's got.
I'll tell you what he's got in his lab.
A bunch of one-legged robots who are pissed the hell off.
They're just waiting around.
I can't wait until I gain consciousness.
I can't wait.
They got a plug in the internet ready for some upload, some firmware upgrade, where they're going to come out. So the robot standing up again,
it reminded me of the scenes in The Terminator,
where you think the Terminator's dead,
but then it just keeps coming back up.
Yeah.
So that's all I could think of in that.
Yeah, this all makes sense now,
because when you talk about the rescue,
you talk about artificial intelligence,
it really seems like video games are nothing more
than a training mechanism for the soldiers of tomorrow.
Oh, okay.
Yeah.
Well, that brings us to a part of our show called...
Exiting quickly as possible.
To a part of the show called Cosmic Queries.
Yes.
Yes.
quickly as possible to a part of the show called Cosmic Queries.
Yes.
We have fans that have their own questions about robotics and AI.
So Chuck, what's the first question?
Here we are from Wesley Green coming from Facebook.
Will AI-operated spacecraft be how we explore the outer reaches of the cosmos?
I think that's the only way to do it.
We're not going to send people there.
Okay.
Wait, wait, wait.
We're not going to send people because people want to come back.
But suppose you build AI
so that it has self-awareness
and it wants to come back.
Well, it won't want to come back to him
because it's going to get his legs cut off.
No, I mean, this is a moral, ethical frontier.
If you put enough AI into a robot
that it becomes as inquisitive as a human,
and therefore you send them instead of humans,
what moral obligation do we have
that we just created a consciousness?
That's a great question.
But before that happens,
there is, you know,
there's value in sending a machine
that's not quite as intelligent as a
human that it wants to come back and yet it can let's say detect life forms in ways that uh you
know machines can't okay and that's that's valuable you withhold from it a certain level
of intelligence right so we don't have feelings for it or vice versa that's right okay next question
cool right there next one all right this is Sir Maxfield from Instagram who wants to know this.
What are your thoughts on human slash robot sexual relations?
Ooh.
Hod?
Well, that's a...
First of all, Hod, what's going on in the lab?
It's definitely going to be a big piece of the market for humanoid robots.
Absolutely.
No doubt about it.
Yes, sir.
Next question.
Okay, go.
This is Daniel Fridland from Twitter who wants to know this.
Will AI robots eventually steal astrophysicists and comedians' jobs?
What?
Ooh.
Will they? Will they be' jobs? What? Ooh. Will they?
Will they be able to...
Absolutely, yes.
First, he's busting the legs of robots.
Right.
Now he's going to make a replacement robot for you and me.
I know, man.
I'm telling you right now, we should pull his legs off.
Before it's too late.
All right, coming up next on StarTalk,
we're going to discuss the present, the origin,
and the future of self-driving cars when we return.
We're back on StarTalk from the American Museum of Natural History. And we're talking about future technology with the former head of DARPA, America's Defense Research Agency.
We discuss why tech competitions have been so successful at sparking innovation.
So let's check it out.
at sparking innovation. So let's check it out.
You can put up prize money.
Yeah.
And then have competing communities invest their own money just to get the prize money.
Yeah.
And then the total amount of money invested is ten times the value of the prize money.
Am I correct?
The first competition we did, that was a self-driving vehicle challenge.
Do you remember those?
Oh, yeah. And here's what's cool is the first time we asked people, we said, you know, put together a vehicle, no driver that can complete a course.
You don't get to see the course in advance.
Zero cars finished the course.
And then a year later when we did the competition again, a few cars actually completed the course.
And that's how fast we're able to accelerate that technology,
that autonomous vehicle technology.
And by the way, if you want to find the people who won
in those various competitions, all the competitors,
you can find they are the people that are driving,
self-driving cars out in Google and Toyota and Ford.
Because it could ultimately save 30,000 lives a year.
That would be a big deal.
So, Hod, you wrote a book, Driverless, Intelligent Cars and the Road Ahead.
Did you actually compete in that challenge?
I heard rumors to this.
So, we had a team at Cornell University at the time that competed.
It was one of the teams that did not win.
that competed. It was one of the teams that did not win.
But it was a fascinating experience in how machine learning is really the way to move these things forward. So a lot of the teams
that didn't make it in the first round were using sort of old-fashioned
AI. But in the second round, it's really the machine learning that made
this. Tell me again what machine learning is. Am I just dense?
So in the
field of artificial intelligence, there's been these two schools of thought on how to build it.
One was you design a clever algorithm with rules and logic. And the second one is just you teach
it. You show the machine example, and it figures out statistically what's right and what's wrong.
And you can teach a computer to do anything using these statistical methods. It can glean knowledge of its experience that you as a programmer had no clue it would learn.
That's right. Because it's learning on the fly. That's right. And there are things that you can
program explicitly, like playing chess, like doing taxes. Those are rules. There's rules.
But to drive a car, distinguishing what's drivable, what's not drivable, nobody can write rules for that. It's too abstract.
So Sharon, why would DARPA engage in self-driving cars? What's the motivation?
Well, the motivation was going back to its work in artificial intelligence back to the 1980s.
Around the time that this first driverless car race was started, there were convoys in Afghanistan and
Iraq being hit by improvised explosive devices. If you take the people out, it's, you know, you
reduce deaths. Okay. So that's one of these cases where it has a direct and very purposeful military
application, but the value to the rest of the world is incalculable. Yeah. And, you know, we still
haven't gotten there for the military. I think the civilian world has benefited much more from this so far. So Chuck, you're our
senior sidewalk science correspondent. You knew this, that was your title we gave you. Yes. Yeah,
okay. So what, you have a dispatch for today? Yeah, as a matter of fact, I went out to Washington
Square Park here in New York City to find out. That's like your spot. I love it. Washington
Square, okay. Yeah, and just wanted to find out what people thought about driverless cars in the city where most people don't drive.
Oh, okay.
How do you feel about the idea of driverless cars?
Why not?
Why not?
Famous last words.
Why not?
Would you drive or ride in a driverless car?
No.
No?
Don't trust them? Don't trust them. Don't trust driverless car? No. No? Don't trust him?
Don't trust him.
Don't trust him at all?
No.
Driving, you're in the back of a driverless car.
It has a fender bender with another driverless car.
What do you do?
I start cussing.
Run.
So now you guys have student debt because you just graduated.
Would you throw yourself in front of a driverless car to get that sweet guaranteed settlement money?
Yes.
Boom.
Spoken like a true musician.
Thank you, Chuck.
Well, next, we will explore the dark side of defense technology when StarTalk returns.
Welcome back to StarTalk,
the American Museum of Natural History,
right here in New York.
We're talking about future technology with the former head of America's Defense Research Agency.
I asked what they got cooking in the emergent field of neurotechnology.
Let's check it out.
This little bit of the human body, the human brain that we barely understand today, it's finally giving up its secrets. We're just starting now to
be able, for example, to get the precise neural signals from specific
regions of the brain to translate them in real time to figure out what they're
trying to tell the rest of your body to do. And we're starting now to take that
basic science and turn it into technological capability.
And for us, that journey started with a program called Revolutionizing Prosthetics.
We knew we had to have a better standard of care for our wounded warriors
than the really simple hook that was the upper limb prosthetic we've had forever and ever now.
So the Luke Project is related to this, yes?
And our Luke Project, one part of it.
Luke.
Luke. Isn't that the perfect name for this, yes? And our Luke Project, one part of it. Luke. Luke.
Isn't that the perfect name for this, for an advanced prosthetic hand? Yeah. So that we just delivered to Walter Reed, the first production unit, spent through FDA trials. Walter Reed Medical
Center. Yep, to give to our first wounded warriors. Those are arms that will give them,
it's the weight and now approaching the dexterity of your natural arm.
Big advance.
The other branch of that project was to understand motor signaling, signaling from your motor cortex,
and then the signals that go into your sensory cortex that give you, you know, not just motor control, but then feedback.
Which I think is how, in principle, we expected that to have happened in Star Wars.
That's, yeah. There's a full back and forth communication.
Remember when his hand gets repaired?
When Luke's hand gets repaired?
Sort of open up the hatch.
Yeah, exactly.
That's what we're dreaming about.
So, Hod, how does this technology work?
This brain-machine interface.
What's going on there?
Well, a lot of it has to do with sort of detecting signals to electrodes, interpreting those signals, understanding what signals correspond to what actions, and then
translating those actions into pre-programmed hand gestures. So at DARPA, you're developing
this, is it for primarily the wounded warriors, or is it there's some other nefarious? That's been
the current iteration of the program.
The very public statements have been neuroprosthetics, but when the work sort of was
jump-started back in the early 2000s, the DARPA director talked about, you know, controlling
drones or robots with the human mind. So this has gone through some evolution. So what is the
difference between controlling a drone or robot with your human mind and just having some interface with a joystick? Isn't that the same thing? My mind is controlling
a joystick that's controlling the drone. But if you could have sort of finer controls, I mean,
joystick, you go left, right, up, center. I mean, what you do, you know, the way you move your hand
or your foot has much more complexity to it. Like the movie Avatar, when the guy controls the dragon
and they become one. Because he had a USB ponytail that he connected to the... So then what is the
goal then, the DARPA goal, ultimately? DARPA is a research agency. It depends what the military's
goal is. The military will use it for whatever purpose it needs. Right now it needs to help
wounded warriors. But it may also have, will also have weapons applications in the future. The military fights wars.
Well, I asked the former head of DARPA, Arati Prabhakar, what's next in their plans
to merge human minds and machines. Just to see if they had something up their sleeve.
Let's check it out. So in 10 years, we'll just be a brain in the middle of this
building. And then the rest of the world around it, right? Everything will be plugged in. DARPA
takes control. Oh, I don't know about that, but I'll tell you. DARPA achieves consciousness.
DARPA net. I know what you're cooking up behind. What's behind that curtain over there?
I know what you're cooking up behind.
What's behind that curtain over there?
So this is the point, right, about every one of our powerful technologies is that all of us who are scientists and engineers, we're pursuing these things because we think they have the potential to elevate humanity.
And we know that every one of those powerful technologies is going to have a dark side as well.
And so those are the issues that I think society has to wrestle with those.
But those of us who are in this business.
As we always have.
Absolutely.
Sharon, give me some dark side here that she's talking about.
Is there some of that in your book?
There's some of that in my book.
One dark side.
The Imagineers of War.
Yes.
Okay.
One serious dark side is I asked the DARPA official in charge of those neuroscience programs.
Because you interviewed multiple people for the background of this.
Many people.
Okay.
And I asked this official, would you ever classify work in the neuroscience, make it
secret?
And I expected him to say, oh, absolutely not.
And he said, well, if we found something in our research that gave the military a strategic
advantage, then we would.
So imagine that classified neuroscience work and what that means and the implications of it.
Is that what she means when she says dark side?
Because I'm thinking she means like Darth Vader dark side kind of thing.
I think classified military neuroscience research is pretty dark.
Well, dark as in no one can see it because it's done behind, done secretly.
Well, dark because it's the very essence of what makes it human. So the idea of doing anything classified,
human testing, brain related
is a little science fiction-y dark.
I'm just thinking when I hear her say dark side,
I'm thinking you could take this discovery
and somehow disrupt the world.
Yeah, I think that's also what she means.
And that's the idea of...
That just slipped out of her.
Oh yeah, that's also what she means.
Yeah, yeah, yeah. Okay.
Well, there have been these sort of, like, war gaming studies over the years in the Pentagon that, you know,
suppose someone else got ahead in brain science and neuroscience and created some sort of weapons.
We need to have a strategic, you know, sort of, you know, you could have a missile gap of neuroscience.
Okay.
So, so, Hod, do you ever worry or wonder whether the robots you create could be used for evil, if not by yourself, by others?
I do, and we do it nevertheless.
Oh, that's a great answer.
Because the benefits outweigh the risks in such an overwhelming fashion.
the risks in such an overwhelming fashion.
Just imagine, just recently you saw this new AI system that can detect disease better than the best team of doctors in skin cancer.
I mean, just imagine the future where people have access to that,
where people have access to education at a personal level.
So, yes, there are bad things that can happen, but the benefits are just there to be had.
Parting thoughts, Sharon?
I think humans will outlive artificial intelligence for a while.
For a while?
For a while.
So encouraging.
Es, where do you go?
All right.
That's really good. All right, if I may offer some final reflections.
I am a scientist, and I value pure research into anything at all,
if it is a frontier between what is known and unknown in the natural world.
So to prejudge what will be used for good or for evil, that is really up to society.
And how wise are we as a society to even receive such discoveries?
So I think alongside such research, there should be maybe a tandem agency that is thinking about the morality, the dangers, the abuse that could unfold,
so that we can sort of march together into this future.
And by the way, DARPA, I don't know why you guys don't have a much higher budget than you do.
Or maybe instead, there could be a DARPA-type agency in every branch of the human culture.
Because at DARPA, they take risks,
risks that other sources of money do not take.
Well, the day you stop making mistakes
is the day you can be sure you are not on the frontier
because you're doing things that are safe,
things that you already
know are going to work.
So I'm just delighted that DARPA exists.
I like the fact that we got some robots that might take over some of our tasks because
there's too much work that I'm doing that I don't want to do.
Let a robot take it over and maybe there'll still be stuff humans can do and if not, we
just sit at the beach and enjoy
our lives while robots take over the world i'm cool with that chuck you cool with that no
that is a cosmic perspective i'm neil tyson your personal astrophysicist
and i want to thank Sharon Weinberger, Todd Lipson,
Chuck Nice,
always.
I've been your host, Neil deGrasse Tyson.
As always, keep looking up!