StarTalk Radio - Gazing into the Future with Ray Kurzweil
Episode Date: March 4, 2016SEASON PREMIERE: Where is humanity going, and what will we be like when we get there? Explore artificial intelligence and “the Singularity” with Neil Tyson and futurist Ray Kurzweil. Chuck Nice co...-hosts, with guest neuroscientist Gary Marcus in studio. Subscribe to SiriusXM Podcasts+ on Apple Podcasts to listen to new episodes ad-free and a whole week early.
Transcript
Discussion (0)
Welcome to StarTalk, your place in the universe where science and pop culture collide.
StarTalk begins right now.
This is StarTalk, and I'm your host, Neil deGrasse Tyson, your personal astrophysicist.
And I'm the director of New York City's Hayden Planetarium, right at the American Museum of Natural History.
Today, my co-host, Chuck Nice.
What's happening, Neil?
Chuck, always good having you. Love you, man.
Love you, too.
Love you, man.
Today, we're featuring my interview with inventor, author, futurist, all-around crazy future guy, Ray Kurzweil.
And we'll be talking about artificial intelligence, of course.
And where does that lead?
To discussions of the future of the human race.
And our guest today is neuroscientist, Dr. Gary Marcus.
Gary, thanks for being on StarTalk.
Thanks for having me.
Oh, you Vanna White here.
We got Chuck holding up his book, The Future of the Brain.
Gary Marcus.
So this is the essays collected together.
It came out a couple of years ago.
It came out a year and a half ago.
It's about to come out in paperback.
It has some Nobel Prize winners from a couple of years ago.
Christophe Koch from the Allen Brain Institute.
All kinds of top people talking about the future of the brain.
How are we going to figure it out?
When will we figure it out? What will it be like? And what is the future of the brain. How are we going to figure it out? When will we figure it out?
What will it be like?
And what is the future of the brain?
Really, really hard.
How are we doing?
How are we doing with the brain?
The brain is working pretty well, but our projects to understand it are hard.
Well, you founded a company called Geometric Intelligence.
That's right.
That's kind of audacious.
You're bringing the math in there.
So what is Geometric Intelligence?
It's a machine learning company. It's an AI company, basically. You founded it. math in there. So what is Geometric Intelligence? It's a machine learning company.
It's an AI company, basically.
You founded it.
I founded it.
I co-founded it with some friends.
And we're inventing new algorithms, trying to learn more efficiently.
So you think about deep learning people are talking about all the time.
That's all about big data or really big data.
And sometimes you can't get big data.
I have two little kids.
They learn from small data.
So I want to have machines that can learn as efficiently as kids and not need it.
Okay, so you're basing a machine, you're imagining a future where you can create learning inspired by the way kids learn.
That's right.
Okay, that's good.
That's great.
Yeah, because don't do it by the way adults learn because you'll fail.
Even worse, the way that computers learn.
Computers are very inefficient.
So they can learn to recognize white males asking search
queries in quiet rooms. But then you have somebody else speak into these machines and
they don't really recognize it. I remember that news story.
That sounds like they've learned pretty well. That's very specific.
It's a little too specific. That's part of the problem. Kids are much more flexible.
That's the mystery of cognitive science is why kids are more flexible than
even the best machines that Google can put together. Well, Ray Kurzweil, he's perhaps best known for popularizing the concept of the
singularity. I never forgave him for taking one of our words, singularity from black.
Totally ripping you off, man. Totally ripping you off.
It's like a hypothetical day. Well, it's real to him and others who follow this line of thinking
that around 2045, when artificial intelligence overtakes the human
cognitive capacities then the world will change irreversibly like walking through a proscenium
through which we can never return oh where the machines that's a good place the humans
and he was the author of the singularity is near does that sound like apocalyptic? Yeah. Why didn't he just name
it Run For Your Life? And he invented many things. Ray Kurzweil, he invented many things.
I got the list here. Text scanning, music synthesizer, voice synthesizers,
speech recognition technologies. He was director of Google Engineering in 2012.
He worked on the team developing machine intelligence. He also received the National Medal
of Technology under President Clinton. It's one of the two highest honors that the White House gives.
One of them is a National Medal of Science. And I was once on the committee to award that
under President Bush. So it's a very highly regarded tradition that we now have in the
United States. And he's author of seven books,
five of which were bestsellers. He's got 20 honorary doctorates and honors from three
presidents. So the man, the man... Yeah, he's done some things.
He makes me feel bad. I feel like I need a nap just hearing that.
So I just want to hear, let's go to my first clip of my interview with him and just find out what
exactly does he predict here and now let's find out going back to a book i wrote in the 1990s
age of spiritual machines i predicted that the age of spiritual machines yeah okay uh by 2029
computers would have all of the intellectual and emotional capabilities of humans.
And so they would be spiritual machines.
And conversely, we are spiritual machines because our brains are composed of 300 million modules that recognize patterns.
Each of those modules has 100 neurons.
Those neurons have ion channels and dendrites and exons.
And basically, you can describe based on physics.
So each neuron is a machine.
Very few people would disagree with that.
Well, then the 30 billion neurons in our neocortex is a machine also.
And actually we can understand it.
My most recent book, How to Create a Mind,
talks about how that machine works.
That sounds diabolical, how to create a mind. What's that machine works it sounds diabolical how to
create a mind what what's going on in your basement you can tell me nobody's looking tell me
well we are understanding how human intelligence works i mean these things are brain extenders
already yeah we are the smartphone that you point to uh yeah our our brain is already expanded
in the cloud which is not a physics cloud.
It's a computational cloud, but we're smarter as a result.
We will directly connect our neocortex in the 2030s to the cloud and expand it.
The last time we did that was 2 million years ago when we went from primates to humanoids.
We got these large foreheads, and we had the frontal cortex which
is where we do language and art and science and physics and radio shows no other species does that
but it's not qualitatively different the frontal cortex is actually was just an additional quantity
of neocortex the neocortex is organized in a hierarchy and we took that additional neocortex
and put it at the top of the hierarchy.
And as you go up the hierarchy, things get more abstract.
So things like music and poetry exist at the top of the neocortical hierarchy.
And we're going to add to it again by connecting our neocortex to the cloud, to synthetic neocortex, and expanding the way we did 2 million years ago.
Except this time it won't be
limited by a fixed enclosure we'll be able to expand it without limit
harry what can you summarize for me what is the difference between high level computing
and the human brain and how is it possible to close that gap erase that gap entirely well i
don't think we really know yet i mean ray is talking with a lot of precision like he says
there's 300 million modules in the brain.
And he says that all these things are at the top of the hierarchy.
That accountable number?
We don't even know what the number is to count.
So there's a lot of precision there that I think is maybe not justified.
Oh, my goodness.
Wait a minute.
I just, you know, because I'm a very polite person.
I tried to say that diplomatically.
I'm a very polite person.
So I didn't catch that precision meant bullshit.
Oh, that was
cool. That was your award, not mine.
There's a lot of...
Just to be clear,
speaking of excrement,
Douglas Hofstadter,
the Pulitzer Prize winning
writer, good old Escherbach,
I've loved his writings, and I still
do. He was even harsher on it.
He said that, quote, I'm quoting here, okay?
If you read Kurzweil's books, what I find is that it's a very bizarre mixture of ideas that are solid and good with ideas that are crazy.
Okay.
It's as if you took a lot of very good food and some dog excrement, blended it all up so that you can't possibly figure out
what's good or bad.
By the way, that's tonight's special at the restaurant.
So maybe dog excrement is a milder form of bullshit.
So please continue.
Go on.
I don't think I can touch that.
I mean, I think I listened to Kurzweil and I know the famous
Hofstetter quote, and I see a lot of stuff that's accurate and a lot of stuff where there's this
false precision where he says it's going to be exactly 2029. And as a scientist, I know that
we have confidence intervals around everything. We know it could be in this range or this range.
There's some kind of statistical uncertainty, same thing with a number like 300 million. And
Kurzweil asserts these things as fact.
And then he asserts some things that actually are fact.
And he kind of...
But wait, he's got more inventions than you.
He has way more inventions.
Okay, so just let the record show.
I'm just saying.
But could that not be because he's, you know,
he's one, passionate.
Two, he's selling books.
You know, and that sells books.
When you make... He didn't start life saying, I want to be a best selling books. And that sells books. When you make very...
He didn't start life saying, I want to be a bestseller.
I met the guy.
He's really an honest guy.
He's into this.
Right, right.
He's for real about this.
He's not a manic cult leader where you look in their eyes and they want to take over the world.
He's really just a regular guy who happens to be really smart.
Now, he goes beyond just what will computers do to match your human brain.
He wants to put nanobots in your brain.
In your brain.
Let's find out where that goes.
Okay.
Computers are getting smaller and smaller.
We'll have nanorobots at the size of blood cells that have computers in them that go
into the brain through the capillaries and communicate with our neurons.
We already know how to do that.
People with Parkinson's disease already have computer connections into their brain.
So my view is we're going to become a hybrid, partly biological, partly non-biological. However,
the non-biological part is subject to what I call the law of accelerating returns.
It's going to expand exponentially. The cloud is expanding exponentially.
It's getting about twice as powerful every year.
Our biological thinking is relatively fixed.
I mean, there have been a few genetic changes in the last thousand years, but for the most
part it hasn't changed much.
And it's not going to expand because we have this fixed skull that constrains it.
And it actually runs in a very slow substrate that's a million times slower than electronic circuits but then why invoke the brain
machine connection at that point who cares you got the machine because it's a much faster interface
i mean our fingers are very slow compared to we could have you know millions of i didn't know the
world was going too slowly for you you want to speed speed it up. Okay. Well, it is. I mean, how long does it take you to read and write? The Brothers Karamazov.
It takes, you know, months. So you're suggesting that you can get these nanobots the size of your
neurosynapses, let's say, and one will be preloaded with War and Peace, the novel,
and would somehow inject it into your neurosynaptic memory banks and then you're done
you've got just like in the matrix they would load memory programs into you yeah well i mean
we will connect to neocortical hierarchies in the cloud some of that could have pre-loaded knowledge
we could communicate with each other more intimately that way. But basically,
we'll expand the size of our neocortex. Gary, do you guys go there in your book?
We do, actually. Christoph Koch and I, in the epilogue, talk about where neuroscience will
be 50 years from now. And we think nanobots probably will be here.
As part of in the brain?
In the brain. But at the same time, we think it's a lot harder than Ray might be suggesting. So
Ray is right that there are already neural interfaces where you can directly hook up to
the brain and read some information, but then he makes the leap so you'll be able to do that in
general. Nobody knows how to do that, for example, with language. So nobody knows how to read out a
sentence from- That never stopped him from making an accurate prediction. He said, nobody knows how
to make a faster computer, but at the rate we're going, it will be faster. And sure enough, it was
faster. So why is your argument an argument?
Well, we made exponential progress in hardware. He's quite right about that. No one would argue
it. It's not a law, so it's not guaranteed. They call it Moore's Law, but it's a trend.
It's really Moore's.
Moore's trend, yeah, yeah.
But in software, it's not really the same. So if you think about AGI,
Artificial General Intelligence, we haven't made exponential progress there were in chess. We've made exponential progress. AI is
basically like idiot savants. We build some very narrow machine. It gets better and better. It's
true for chess, true for a lot of things, but general intelligence, like we had Eliza in 1965.
I remember Eliza.
And you could tell it your problems by telling it.
I had a conversation with Eliza.
Maybe even you believed it was real.
That's how old I am. Back when you had to type, I said, Eliza,
I'm not feeling well. And Eliza would reply, why do you think you're not feeling well?
It would be very sort of turn the question back around. Now the verbs and nouns and make a,
and it would be sound like a, like a, you're on a shrink. Yeah. I was going to say she could
charge $75 an hour for that. And now 55 years later, um you have Siri, and people have the same kind of conversations
with Siri, and Siri is just as fake
as Eliza. So Eliza didn't really understand
what your problems were, even though Eliza
asked you. Are you telling me Siri doesn't
know what I'm really saying?
Sorry to break it to you. I don't know what to believe anymore,
man!
So we haven't made as much progress there. So some things
are exponential, and some are closer to linear.
And Ray's not really...
When I talk to Siri, I say, Siri, where's the nearest Starbucks?
It's going to find a coordinate grid.
It's going to get an address.
It's going to hand it to me.
So don't tell me it's not thinking.
You don't want to call it thinking.
But if Siri showed up 40 years ago, your mind would explode, right?
So here's a point that he made sort of, you know, when I was just chilling with
Ray, it was that every time there's a new kind of AI, people say, oh, it's not really AI. The
real AI is that, whereas anything we're talking about would have blown away anyone's concept of
AI decades ago. So I think we have a moving baseline for how we're judging.
I mean, that's a familiar argument. People say say if it doesn't work, then it's AI.
When it finally works, it's just engineering.
And I think AI researchers have a right to complain about that.
But I still think if you're talking about general purpose AI, we just haven't made the progress.
So, yeah, you can ask a few things.
Like every time Apple rolls out a new OS, there's a new category of things you can ask.
Sports scores, Starbucks, and so forth.
But imagine if you have a kid. I have a three-year-old kid. Imagine if it was like every six months he
rolled out a new conversational topic and really stuck to three others. I would take him to a
speech pathologist or a neurologist. I'd be really worried. Well, maybe some of it is just it's
missing the emotion. So what I really wanted to know is how does Ray get his precision and how
he's calculating when the singularity happens, when we become slaves to robots?
That's what I really wanted to know.
He told me.
Check it out.
There's one thing that's surprisingly predictable, which is the pace of the exponential growth of information technology.
technology, the price performance and capacity of information technologies like computation,
communication, now biological technologies like sequencing and also simulating biological processes, the amount of data we're getting on the brain, many different measures, not of everything,
but of information technology follow amazingly precise trajectories and they're exponential.
So I started with the common wisdom that you
cannot predict the future, and I made a surprising discovery. Lots of things are unpredictable, but
if you look at the price performance and capacity of information technology,
it follows an amazingly predictable trajectory, going back to the 1890 American census.
trajectory, going back to the 1890 American census. And one of the reasons you might be skeptical is it's not intuitive. Our intuition is linear. And a linear progression, that's our intuition,
goes one, two, three. An exponential one, that's the reality of information technology, is one,
two, four. It doesn't sound that different. Except by the time you get to step 30,
the linear progression, our intuition is at 30.
The exponential one is at a billion.
So you're right on the curve as time goes on.
This gives you confidence in your modeling.
And so now you say 2029, what happens?
So at 2029, we'll be able to match human intelligence in the computer.
As we go through the 2030s, we'll be able to multiply our intelligence, you know, some number fold.
go through the 2030s we'll be able to multiply our intelligence you know some number fold you get to 2045 and according to my calculations we'll be able to multiply our capacity for thinking
a billion fold that's such a profound transformation it's such a singular transformation that we
borrow this term from you guys uh thank you very much that that's really all i wanted out of it was to thank the astrophysicist
for handing you this term it actually came from you borrowed it from math and it's an infinite
level but the most significant thing about a black hole in physics and singularity in physics is
the event horizon going past the event horizon is a little mysterious and so there's a certain
amount of mystery about going beyond the event horizon.
And that's the same.
That's the really the metaphor.
Okay.
Going past that event horizon is hard to predict because it's so transformative.
So, Gary, this precise talk, it's very apocalyptic.
And it's influenced people in exactly the way cult leaders
Influence people saying the world the end of the world is near. Well, I mean Ray takes a positive
View of it. I mean he thinks the beginning of the world. He thinks it's nerd rapture. We're all gonna ascend
So, I mean you'd be honorary nerd with us because you do the show with me
Well, I appreciate that because I'm ready to go
us because you do the show with me. Well, I appreciate that because I'm ready to go.
We'll all want to ascend when we get the chance, I guess. But I mean, again, there's the kind of false precision. So first of all, Moore's Law goes back to before Ray was on the scene talking
about these things. It's been around for a long time. But second of all, you can get seduced.
You said since the 1890 census, information. There's data going back there. I mean, a question about where the idea came from.
But anyway, you could look at something like cloning.
And if you were, you know, in 1995, you'd say it's going to be here tomorrow.
We've got Dolly.
And then the next thing you know, like, it's not really a practical technique.
All the clones, you know, died early and so forth.
We'll do more of that when we come back from the break.
But also, we're going to find out what Ray Kurzweil thought of the movie Transcendence
and what he has to say about the existential
risks associated with
AI.
Welcome back to Star Talk. I'm with Chuck
Nice, co-host, professional
comedian. Yes.
Reading a Chuck Nice comic. I love it when you say professional.
I can't.
Okay.
I do get paid for this.
I'm stealing money.
Okay.
I am stealing.
Happy to have my money stolen by you.
Well, not from you.
I'm saying when I do comedy.
And to help us out here, because we're talking about Ray Kurzweil and concepts of artificial
intelligence and the brain and the future and computers and the fate of the human race, we have Gary Marcus, who's a neuroscientist at New York university. Thanks for being on star
talk. And so, you know, Ray's views reminded me of the movie transcendence. I don't know if you
saw that. That's the one that had Johnny Depp. Yeah. I don't know how many, how was a kind of a
takeoff on lawnmower man. Oh, is that right?
Yeah, which was the Keanu Reeves version of Transcendence.
All right, all right.
You analyzing these way on a level that I don't know.
It's, you know, it's so weird that I don't have a life.
So I brought up the movie just to get his reaction to it.
Let's find out where he goes.
In the film Transcendence, that's what i imagined your scenario
would describe just for people to remember in transcendence johnny depp the lead character
lifts his mind into the network of the world's computers and that becomes his life and his body's
gone at that point and it becomes dangerous and that's a typical dystopian view of these
technologies i mean that's not how
these things happen did mobile phones or the web or the cloud just leap overnight it starts out at
a point where it doesn't really work and it's kind of very flaky and people dismiss it because that
doesn't really work by the time it works it's been around for a long time and we'll get used to it
okay so let it be that but now let's assume we get to that point
is that a natural progression of the scenario that you have predicted i mean i think that's
part of the human experience we're already a human technological society but in that case he's not
even a body he has an existence in neurosynaptic yeah well we are going to be able to project
our minds on into virtual environments.
We're already starting with virtual reality.
That's a little crude today, but it'll become very realistic.
It'll ultimately actually go inside the nervous system
so that we could...
Maybe like total recall.
In total recall, your vacation was just what they embedded in your mind.
A couple could become each other.
You can inhabit fantastic virtual environments.
You can have a different body in a virtual environment. So we're already separating
physical reality from its consequences. And we're going to increasingly do that.
So Gary, where do you brain people take all this? Where are you going with this?
Or not?
I had a friend who said that all neuroscientists are closet upload people,
which is to say that all neuroscientists want to achieve immortality by uploading their brains to the web.
But none of the neuroscientists.
Yes, we knew that.
None of the neuroscientists I know really want to do that.
All the neuroscientists I know think it's wildly unrealistic.
It's like 50, 100, 200 years away. They're going to have to. Now, wait a minute, because when you think about it, is, are we without these little
electronic impulses inside of us, these little synaptic jumps of electricity between nerve
endings, you know, outside of that, that really determines everything that we think and touch.
And so if we can experience so if we can take that
and recreate it artificially.
Why is that 200 years away?
Why wouldn't that be the same
as being who we are right here?
Just I'm feeling this,
but I'm just not actually doing it physically.
I gave you a confidence interval.
I said 50, 100, 200 years away.
I mean, I think it's somewhere in between.
I do think it's possible to upload brains.
I mean, if your brain is uploaded,
that's a copy of you.
So those are really weaselly error bars there. Well, it's not. Somewhere between? I mean, if your brain is uploaded, that's a copy of you, not you. So those are really weaselly error bars there.
Well, it's not.
Somewhere between 50 and 1,000 years from now.
Actually, I wrote a letter to the Times, actually.
It's like, we'll be there somewhere in between 8 and 5 p.m. Tuesday and Friday.
It is like cable.
That's the point.
It's like waiting for the cable guy.
We don't know exactly when he's going to come.
Wait for the cable guy.
Go around saying he's going to come at 4 o'clock.
We're misleading people. That's the point.
But you must agree that
if it goes the way he describes,
there's tremendous utility to this.
You agree?
Well, I mean, it makes a copy of you, and it depends
on how narcissistic you are for whether you want a lot
of copies of you floating around
in the cloud. It's not actually going to make you immortal.
Okay, but fine. Let there be copies. But I'm also talking about, there are a lot of brain
disorders that brain people study. If we can map your brain, can't we find out where stuff goes?
That's one of the things we talk about in the epilogue to this book, The Future of the Brain,
is the difference between, for example, if you have a brain injury, maybe I can make a backup
of your brain before the injury. I'll do regular backups of your brain.
I'll be able to restore the backup.
Like a time machine for the brain.
That's pretty cool.
But on the other hand, a congenital disorder.
That would be an Apple product that you just plugged there, by the way.
Yes, I did.
I want to cut.
No, I just realized that I did that.
I was like, oh, like everybody knows Apple Time Machine is your backup.
Time machine pre-exists as a thing.
Right.
Apple backup scenario. There's going to be some problems with the Ethernet. It's not always going to work right. machine is your backup time machine pre-exist as a thing right before you know it's a backup
scenario right there's gonna be some problems with the ethernet it's not always gonna work
right sometimes they're gonna restore the wrong version but at least it's possible in principle
on the other hand like if you have a congenital disorder maybe we won't know how to fix that
because we don't understand enough about how the brain works so even if you didn't understand the
operation of the brain if you simulated every molecule etc then maybe maybe you could at least
do a restore and maybe that'll be one of the first applications.
Is this the first step to immortality?
It's the immortality of a copy of you.
You know the Woody Allen line about preferring to live on in his apartment rather than through his work.
I mean, it's another way of living on through your work.
That's what I'm saying.
But it's a pretty cool copy of you because it's like you're like a sim now.
So you can upload
yourself. Your own avatar. Your own avatar.
Totally. I can upload myself and I'm
still there if you're able to
actually upload the consciousness. You'll be able to do
roughly that long before then
with VR. You won't be able to do
it at the neuron by neuron level, but you'll be able to make
a pretty faithful avatar
that'll get all the motion caption stuff,
get your voice right, and so forth.
That's actually scary because you know there's going to be some guy who's going to make an avatar of himself as a girl
and then go into virtual reality chat rooms and screw up the guys.
That's already happening.
Oh, right.
Yeah, isn't that kind of already happening?
Is that happening?
Plus, why do you know so much detail about that that can happen?
I have an active imagination.
Okay.
I have an active imagination.
Okay.
Well, okay.
So whether it's AI taking over us or we controlling AI for evil, nefarious purposes, there's no doubt that there are existential threats that this technology can bring.
And I asked Ray, of course.
I had to ask Ray about this.
Check it out.
Technology's been a double-edged sword ever since.
It's ever been fire i mean fire kept
the swamp but burned down our houses and every technology can be used for creative and destructive
purposes we have an actually a new technology that has an existential risk already which is
biotechnology the existential risk from artificial intelligence or nanotechnology is off in the future and we can
debate is it 10 years away or 50 years away but it's not here yet but the ability for someone to
take a benign virus like a cold virus and turn it into a super weapon that make it more deadly
more communicable more stealthy exists right now That could be done in a biotechnology lab probably a few blocks from here.
So that was recognized actually 30 years ago.
They had a conference called the Asilomar Conference to come up with guidelines.
How can we keep this safe and reap the promise without the peril?
And they came up with the Asilomar Guidelines.
Those have been made more sophisticated over time.
And they've worked very well.
We're now reaping the benefits of
biotechnology the number of incidents either intentional or accidental but there's been harm
from biotechnology so far is near zero it's zero yeah now that doesn't mean we can cross it off our
worry list because the technology keeps getting more sophisticated but nonetheless it's actually
a good model for how to keep these technologies safe. Well, plus, you know, when we know that fire exists, and so we have fire codes,
right? This is how you build a stairwell. And this is how you escape. And this is how
and we have a moral imperative to use fire or artificial intelligence or biotechnology to
overcome the problems, you know, that humans have, there's still a lot of human suffering.
And we're using AI to
diagnose disease and come up with new cures and clean up the environment to reduce poverty. And
we have a moral imperative to continue that way while we have ethical guidelines to keep the
technologies as safe as possible. So Gary, you're quite prolific on this topic in the popular media, even not only professionally.
So one of your articles for The New Yorker said why we should think about the threat of artificial intelligence.
So this sounds very lauditic.
I'm no laudite.
In fact, I'm just launching an organization called AIforgood.org, which is about what positive outcomes we can get from AI.
But there are also risks, too.
And it's a tradeoff, right?
But are ethical guidelines enough to just guide this?
Because we have ethical guidelines for everything else that could possibly kill us.
Probably need regulation, too.
Yeah, regulation.
But we do that.
We don't say let's not have airplanes because they could crash.
We have regulations to make them as safe as possible.
And they still crash, but we accept that risk.
That's right.
And, I mean, we do some kind of calculus to decide whether it's worth it.
And it may be that 200 years from now, people look at us and, like,
why did they use cars before they had computers in them to make them safe?
They've lost so many people.
And so people may look back at us now and say,
the ways in which we handled AI in 2150, the early days were really pretty poor. And so I don't know what
the regulations are going to be. It's probably going to be iterative. One of the things that
I think we all worry about is that the pace could be fast and we don't have enough time
to take care of it. Do you share the total concern that the famous trinity of Elon Musk,
Bill Gates, and Stephen Hawking have shared?
No, I have a milder view.
I mean, I think that real strong AI is—
Just to be clear, they're, like, freaking out.
Right.
They think, basically, machines are going to take over and kill us all.
Kill us all.
And then the future of the world is a world of machines.
A world of machines.
Right, right.
I don't think it's right.
The terminator.
The machines are look, they're coming.
Okay.
The honest truth is that Skynet is not going to be here tomorrow.
I have an Austrian accent.
All the machines will have an Austrian accent.
Look at this, I'm going to kill you.
I'm sorry, go ahead.
I don't think that the Schwarzenegger version of it's going to be here anytime soon.
I don't think the computers care about us so far.
You get computers that are exponentially smarter at playing chess, and they don't give a shit about us at all.
I don't know if I can say that on the air.
But at the same time, I still think we need to be worried.
As computers get more and more embedded in our lives, they have more and more power to change things.
So they're going to start driving our cars, for example.
And if the AI isn't right, then there's a risk there.
So it's really not about them taking over. It's about the risk of
failure. Like I see the real risk is, okay, all the cars are being driven by computers. And then
all of a sudden we have some kind of magnetic pulse and boom, everybody crashes all at once
more than they're just going to be like, yo man, I'm not taking you where you want to go.
Exactly. I don't care where you want to go. I'm tired. An ornery car.
I would not worry about the ornery car. An ornery car.
That's right.
I would not worry about the ornery car, but I worry about security and programmability.
So if you've ever programmed anything, it doesn't always work.
Of course, security, programmability, we got that.
But what of this quote from Nick Bostrom?
Are you familiar with his writings? And he says here that the creation of AI with human-level intelligence will be followed
immediately by an almost omnipotent
superintelligence, with the consequences
that would be disastrous.
We've had human-level intelligence without
omnipotence for a long time.
The word immediate is particularly
perplexing. I put median in here
probably. Sorry, that's my word in here.
Followed by. Followed soon after.
Yeah. And nobody knows how
soon. Nobody really knows what's gonna happen
I think it's worth spending just unplug the damn machine
Yeah, I'm gonna say like well, you know people think about that 30 off six
That's the end of the machine like that one. It's one look we shoot each other for less
Tremendous capacity for violence
They don't know who they are screwing with man
They don't know who they are Screwing with man
Are you kidding me
I will bust you
Bust a cap in your silicon wafers
Compute this
Mother
So wait
Get back to my interview with Ray Kurzweil
I had to ask him what did he think was the most realistic AI film.
He thinks about this.
We all think about it.
Let me get, like, the expert's view on that.
Let's check it out.
Of all the films that project sort of AI in the future,
which would you say was the most realistic?
Well, I liked her because there wasn't one AI.
Like, everybody had their OS, their operating system.
And Theodore's human girlfriend also had a relationship with her operating system.
And so that's more realistic.
I thought it was unrealistic that Samantha didn't have a body because I think it'll actually
be easier to provide her with a virtual body than to provide her with her human level mind.
So by the time we have Samantha's mind, we'll have virtual reality,
and she'll have a body.
And then the sex market will completely drive that.
I mean, the sex market does.
It drives the bandwidth today on the Internet.
A lot of technologies.
Isn't half the bandwidth of the Internet is porn?
Videotape and even books.
Gutenberg's first book was the bible but then
there followed a century of more purient material so we're in the future of the brain
we should not be surprised that sex drives decision making as it's not going to change
and uh so And so in another
StarTalk, we interviewed Dan Savage, and he
made it very clear, something we all knew, but I hadn't quite
put it that way, that sex predates
humans by half a billion years.
So... It's kind of the most powerful
force on the planet. It's a force of nature
unto itself. So how much of that
is driving what the
neurosynapses are in the first place,
plus what we'll do with the
power over the brain in the second place.
Well, I think all of evolution
is shaped by sex in part. I mean, there's no
question about that. I think that
in terms of are we going to have sex bots
and things like that, for sure.
And are we going to have... Chuck! Chuck!
Chuck goes double thumbs up on that one.
God, yes! It's actually an interesting question.
What's going to happen to society as sex bots
become, you know,
more and more authentic
in some way?
Fine with me,
as long as they program them
not to get jealous
because I plan on having three.
But wait, but here's...
This is a polyamorous roboticist.
Okay.
Is that a sentence?
Yes, love it.
Polyamorous.
Oh, wow.
But I can tell you this.
Whatever goes on, you don't want any... you don't want to catch a computer virus.
That's the worst kind.
You don't want these things to cross worlds.
That's what I'm saying.
You cross species.
I just fixed the gonorrhea.
Now I got a computer virus I got to deal with.
But I don't know if it's a centerpiece of your conversations or not. I mean,
I look at efforts that are going on here.
In my notes, there's
a campaign launched in
2015 for a
ban on the development of sex robots.
I think that's silly.
That's silly. Okay, good. I think we should get rid of
all the vibrators, too. Like, why?
Yeah, they're called vibrators.
I was going to say, we just don't have them for men,
so why are they trying to shut us out?
I mean, come on.
Equal time, people.
Equal time.
And there's a company.
I got a true companion, a company, which has designed what they call the first sex robot called Roxy.
Roxy with three X's.
Triple X Roxy.
It's got three X's, of course.
Look at that.
That's so clever.
Of course.
Male sex robot also in development.
That's totally useless.
Male sex robot.
Completely useless.
Every man is a male sex robot.
We are programmed to want sex and to have it no matter what, and you will never have to beg us for it.
Okay, suppose you program a sex drive into AI.
Because then that's just making it more human.
That would be honest, right?
Well, I guess there's a question about what it means to program a sex drive into an AI.
Like, do you mean you're going to program the robot to have sex with people?
Or to actually want it?
I don't want it.
I mean, probably nobody.
If you want to be a real human person.
Probably nobody's going to bother with that.
It's easier just to program it to do certain actions and so
forth. I don't know if you need that
level of conscious reflection.
Here it is. You program it to like it. That's what's interesting about her
though. Like or want. See,
you got it. That's it. Because here's the
thing. Otherwise, you're just going through the motions
and that's called marriage. Nobody needs
that.
No, no. There's a subtle distinction
here. There's a subtle distinction.
So one of them is, it likes it.
Right.
If it happens to it.
Right.
But if it wants it, then it could be waiting for you on the street corner.
Yeah, you're right.
You know what?
Talking about your tight ass.
I will not be ignored.
I'm still here.
I think I've calibrated the language now.
The real lesson of the Turing test, the famous Turing test, is that people are very easily fooled. You don't have to go to great lengths in order to fool a
person that a machine is taking interest in them and so forth. And so it's not a growth industry
in that growth industry to solve these finer points. You can get away with easier. And of
course, the movie Ex Machina was basically how to use sex as a force in anything that happened in the storytelling.
He taught her too well.
You saw that movie.
He taught her far too well.
Taught her real good.
Spoiler alert.
More of my interview with Ray Kurzweil when StarTalk returns.
Welcome back to StarTalk.
I'm with my co-host, Chuck Nice.
That's right.
And our guest, neuroscientist, Gary Marcus.
And this is a tweeting neuroscientist, apparently.
You're tweeting at Gary Marcus.
That's right.
M-A-R-C-U-S.
So now I know you're out there.
I'm going to find out what you got to say in the Twitterverse.
We're featuring my interview with the inventor, futurist, all-around, deep-thinking dude, Ray Kurzweil.
And so far, we've been talking about AI and robots.
But I also brought up the topic of 3D printing during that interview.
And what he really thinks about where that technology is headed, because if you can 3D print stuff, maybe one day we can 3d print, uh,
organs or, or,
uh,
physiological forms.
Let's just find out where he,
where he thinks this will go.
Check it out.
Cool.
What's the distant future of 3d printing?
Are we going to be able to one day just print a human being?
And then we're in the hype phase now,
kind of like where the internet was in the 1990s.
If you remember,
if you had the URL dog.com,
you were a billionaire.
And then the year 2000, people realized,
wait a second, you can't make money
on these internet companies.
We had the internet crash.
It almost took down the world economy.
And now we actually do have internet companies
like Google and Apple and Microsoft
that are worth hundreds of billions of dollars.
It's going to be a similar story with 3D printing.
Right now we're in the hype phase. There's interesting niche applications. The exciting applications, I believe, will be in
the early 2020s. You'll be able to print out clothing. We need sub-micron resolutions. We'll
be there then. You'll have millions of free, cool designs that are open source. You can download for
free and print out for pennies per pound. You'll also have a proprietary market.
People are already, I'm involved with a company where we're printing out organs, human organs,
and successfully installing them with some limitations in animal models.
Within 10 years, I believe you'll be able to print out human organs.
Okay, Frankenstein.
Well, you know, if you... Did you say that so casually?
If you need a new lung...
Can you print a brain?
Is that in the future?
You probably don't want to do that
because I think it makes more sense
to emulate thinking non-biologically
because that's really what we're working on.
But, you know, if you need a new kidney,
a new lung, you'll be very happy
to be able to print it out on a 3D printer.
Ultimately, we'll be able to print out most of the physical things we need as we go through the 2020s, and it will revolutionize manufacturing.
Gary, you wrote an article for The New Yorker on growing brain parts.
That's equivalent, perhaps, to 3D printing a body part.
Where did you land on that?
I mean, it's something that's going to happen here.
I agree with Ray.
The only differences I have with Ray are usually about dates.
I mean, he says these things with certitude.
It's going to be in the 2020s.
And we don't know because there's all kinds of complications about infection and how to get the parts right.
Actually, every clip we've played for you, you say, yeah, I agree with him there.
But we began with you, like, totally saying he had, you know.
No, I disagreed on the neuroscience.
I disagree on the structure of the mind, and I disagree on the dates.
I do agree we're going to be able to print brain parts and other body parts.
I mean, that's already happening.
It's not to, you know, production level stuff that you can ship.
Yeah, just go to the local hardware store.
Soon it'll be in vending machines.
It will.
Look at that.
Lost an ear somehow.
Mike Tyson bit off my ear.
Let me replace it.
He has to put in a drop of blood so they can, like, customize it to you or something like that.
But, yeah, I think that will happen someday.
Not as soon as Ray thinks, but sure.
I mean, the real question is not when is this going to happen.
What's the world going to be like after all these things do happen?
You know, I think maybe it takes 50 years for things he says take 20 years, but the world is going to be different.
So what are some of the unintended consequences of these advancements? Because this is the thing that kills me about all of these technological forays that we make is that we go
there and then we go, oh, wait a minute. I didn't realize that this was going to happen. So, you
know, you did what? Right. Exactly. You know, My favorite line, I forgot where I got this, it was
the last words
ever spoken by
humans ever
is
let's try the experiment the other way.
That's it.
I like that.
We don't want those to be our last words.
That was the end of the world.
Let's wire it the other way and see what happens. We don't want oops to be our last word. That was the end of the world. Let's wire it the other way and see
what happens. We don't want oops to be our last word. I think that's right. And so, I mean, I
think a lot of the things that are on the table and discussion now are science fiction. They're
not going to happen, but we don't know which ones are science fiction, which ones we need to watch
out for. And so I do think it's worth investing real money. I got some data here. What was it?
Printed body parts brought in nearly half a billion dollars in 2014, up 30% from the previous year, according to an article in Nature News in April 2015.
It's already a thing.
Yeah.
I mean, not printing brain parts.
There's an old joke about brain transplants are the one thing where you want to be the donor rather than the recipient.
And that's a little bit more complicated, right?
And let me tell you something.
I'll tell you where you can go get your money for that is the NFL.
Okay. They will gladly fund that research they are the nfl is you know just play nfl with head bucking and then when you get out you just get a new brain right there you go it's more like people
want to figure out how to repair your brain after a concussion i mean that's that is in fact a big
industry and people are making progress on it well and and repairing the brain right why
get a new brain if you can just go in there and repair it yeah yeah it's just a few neurosynapses
that you want to keep your ideas your personality your memories and so forth i got that in the chip
don't you know get with the program here right we already got that on my usb chip that's that's
where ray is confused about reality impossible future so i know from nasa NASA, I hosted a panel with some NASA representatives where they're
printing machine parts that you might not be able to get at the local hardware store when you're on
Mars, right? On a long trip and you got to make a bolt or a fan blade or something. You just print
it. And those, of course, are just sort of dumb instruments. They don't have the complexity of an organ, a human organ.
But I can imagine...
That's what Damon should have said in The Martian.
I'm going to 3D print the shit out of this.
So in the future, maybe it's not about sending a medical doctor
to help your body repair its own wounds.
It's about sending...
Hewlett-Packard.
Hewlett-P. Hewlett Packard
is very big in 3D printing.
A machine, and you just
replace stuff that breaks.
And kind of the way they did in Star Wars,
Luke's hand, they just, you know,
you pick him up in the next scene,
test in the fingers, and he's
good to go. Right. I think that
will happen. Not
even a thing. So are you,
so you don't fear the future of 3d printing. That's not a scary plight for you. Um, the version
of it that does scare me a bit is 3d bioprinting. Um, which is also, I mean, we were talking about
a piece of it now. Um, people are going to be able to 3d print, you know, viruses and things
like that. And I worry about what worry about what we call them unpleasant actors,
to put it diplomatically,
trying to do those kinds of things.
I asked Ray, you know,
as our conversation drew to a close
and all of this stuff is on the table,
I had to ask him,
what are the best and the worst possible outcomes?
Let's see where he goes.
The best, and I think one that we are actually following,
although people don't necessarily agree with this,
is that life is getting better.
We will, I believe, by mid-late 2020s, eliminate poverty.
To get to 2030s, we'll eliminate disease.
We're going to dramatically expand human life expectancy.
People say, oh, we're going to run out human life expectancy people say we're going to run out of
energy and other resources we have 10 000 times more sunlight than we need for free
that to meet all of our energy needs and there's lots of other resources we have lots of water but
it's you know it's dirty or salinated but if you have inexpensive energy we know how to to clean up
water vertical agriculture where we grow food in
vertical buildings controlled by artificial intelligence will produce very inexpensive
high quality food hydroponic plants which is already really happening we have inexpensive
high calorie food that's compared to 50 years ago absolutely but basically it's moving towards
you know greater prosperity so what's the downside?
The downside is, I mean, there's lots of movies about the downside.
I don't need you to tell me about what the movies do. You know, the AI is trying to destroy humanity,
and it's the humans against the artificial intelligence of the robots.
It's two groups of humans fighting for control of the AI.
One thing that gives me comfort is we don't have one or two AIs in the world.
We have two or three billion AIs.
I mean, a smartphone is an artificial intelligence.
It's, in fact, very intelligent and it accesses the cloud and makes it even more intelligent.
And it's in billions of hands.
So, Gary, what are your reflective thoughts on this
as we wind down here? I think the thing that he left out is income inequality. So I think that
it's certainly the case that these new technologies, other things being equal, are going to
make a more productive world. There's going to be more resources tapped more efficiently. But
what he leaves out is that the data is very concentrated in a few players, and so are some of the best AI techniques.
Patents and things like that are locked up in a small number of companies.
That's an early problem that possibly could be resolved later because so many people have smartphones now, and even poor people have smartphones, right?
It could be, but I think there's a whole issue about employment we didn't talk about today.
That's a homeless guy.
Call me.
Cell phones are going to become essentially free.
Yeah, exactly.
But jobs are going to become scarce, and that's an issue that Ray didn't touch on.
And then the other issue is what is the real probability of some risky scenario where machines turn us into paperclips?
I think it's very small, but it's probably not zero, and it behooves us as a species to make sure it comes as close to zero as we can possibly make it.
Okay, that's a rosy outlook.
I like it. Chuck, you got any? We're good to go.
Listen, you guys
had me at Sex Robot.
We were good.
Let's see if we can hook you up. This is all headed, as far as
I'm concerned. In the right direction. It will go
to Sex Robot, and everything
will go from there. It will branch out.
We will never see you again.
Chuck! Anyone saw Chuck? to sex robot and everything from will go from there it will branch out we will never see you again chuck well they might saw chuck oh my god i wish that wasn't true well i i went in as his biggest skeptic because i you know don't tell me that the world is going to
be you know everyone always says the world's going to be different because of something they thought up. And I, I do a lot of the reading of history. So does he,
of course. And in a, when I finally confronted him just to see how, how sort of sane he was and how
straight he was. And, and, and when he makes a prediction, it's based on a calculation. It's
not just some dreamy stuff he thought up in an armchair, you know, in his lazy boy armchair.
And so I gained a level of respect that I didn't know was possible within me to grant to him.
And so whether or not he's right, I mean, futurists don't have to always be right.
If you write one out of five times, that's an awesome track record.
And I think he's right a little more than one out of five times.
Unless it's a coin and you're flipping and, you know, one out of two would be better.
Yeah, that's true.
But if you have many, many, many predictions and you see how they play out.
There's actually interesting stuff on Wikipedia that talks about his track record.
People are interested in the data.
Okay, excellent.
Excellent.
But I'm going to lean his way more than I was before.
And I'm glad somebody's at least thinking about it because I certainly wasn't before I chatted with him.
You've been listening to StarTalk Radio.
Gary, thanks for being on the show.
The future of the brain.
I'm going to pick one of these up
and find out where you guys are coming from.
And Chuck, this might be the last show we have of you
because you'll be sex botting for all the future time.
So that's our show.
As always, I've been your personal astrophysicist.
And as always, I bid you to keep looking up.
This has been StarTalk.