Making Sense with Sam Harris - #255 — The Future of Intelligence

Episode Date: July 9, 2021

Sam Harris speaks with Jeff Hawkins about the nature of intelligence. They discuss how the neocortex creates models of the world, the role of prediction in sensory-motor experience, cortical columns, ...reference frames, thought as movement in conceptual space, the future of artificial intelligence, AI risk, the “alignment problem,” the distinction between reason and emotion, the “illusory truth effect,” bad outcomes vs existential risk, and other topics. If the Making Sense podcast logo in your player is BLACK, you can SUBSCRIBE to gain access to all full-length episodes at samharris.org/subscribe.   Learning how to train your mind is the single greatest investment you can make in life. That’s why Sam Harris created the Waking Up app. From rational mindfulness practice to lessons on some of life’s most important topics, join Sam as he demystifies the practice of meditation and explores the theory behind it.

Transcript
Discussion (0)
Starting point is 00:00:00 Thank you. of the Making Sense podcast, you'll need to subscribe at samharris.org. There you'll find our private RSS feed to add to your favorite podcatcher, along with other subscriber-only content. We don't run ads on the podcast, and therefore it's made possible entirely through the support of our subscribers. So if you enjoy what we're doing here, please consider becoming Today I'm speaking with Jeff Hawkins. Jeff is the co-founder of Numenta, a neuroscience research company, and also the founder of the Redwood Neuroscience Institute. And before that, he was one of the founders of the field of handheld computing, starting Palm and Handspring. He's also a member of the National Academy of Engineering, and he's the author of two books. The first is On Intelligence,
Starting point is 00:01:14 and the second and most recent is A Thousand Brains, A New Theory of Intelligence. And Jeff and I talk about intelligence from a few different sides here. We start with the brain. We talk about how the cortex creates models of the world, the role of prediction in experience. We discuss the idea that thought is analogous to movement in conceptual space. But for the bulk of the conversation, we have a debate about the future of artificial intelligence, and in particular, the alignment problem and the prospect that AI could pose some kind of existential risk to us. As you'll hear, Jeff and I have very different takes on that problem. Our intuitions divide fairly sharply, and as a consequence, we have a very spirited exchange. Anyway, it was a lot of fun. I hope
Starting point is 00:02:16 you enjoy it. And now I bring you Jeff Hawkins. I am here with Jeff Hawkins. Jeff, thanks for joining me. Thanks for having me, Sam. It's a pleasure. I think we met probably just once, but I feel like we met about 15 years ago at one of those Beyond Belief conferences at the Salk Institute. Does that ring a bell? You know, I was at one of the Beyond Belief conferences, and I don't recall meeting you there, but it's totally possible. And I just... Yeah, it's possible we didn't meet, but I remember, I think we had an exchange where, you know, one of us was in the audience and the other was, I mean, so we had an exchange over 50 feet or whatever. Yeah. Oh, that makes sense. Yeah. I was in the audience and I was speaking up.
Starting point is 00:03:04 Yeah. Okay. And I was probably on stage defending some cockamamie conviction. Well, anyway, nice to almost meet you once again. And you have a new book, which we'll cover part of, by no means exhausting its topics of interest, but the new book is A Thousand Brains, and it's a work of neuroscience and also a discussion about the frontiers of AI and where all this is heading. But maybe we should start with the brain part of it and start with the really novel and circuitous and entrepreneurial route you've taken to get into neuroscience. This is the non-standard course to becoming a neuroscientist. Give us your brief biography here. How did you get into these topics? Well, I fell in love with brains when I just got out of college. So I studied electrical engineering in college. And right after I started my first job at Intel, I read an article by Francis Crick about brains and how we don't understand they work. And I just became enamored.
Starting point is 00:04:15 I said, oh, my God, we should understand this. This is me. I am my brain. No one seems to know how this thing is working. And I just couldn't accept that. And so I decided to dedicate my life to figuring out what's going on when I'm thinking and who we are, basically, as a species. And it was a difficult path. So I quit my job. I essentially applied to become a graduate student
Starting point is 00:04:41 first at MIT in AI, but then I settled at Berkeley in neuroscience. And I said, okay, we're going to spend my life figuring out how the neocortex works. And I found out very quickly that that was a very, not difficult thing to do scientifically, but difficult to do from the practical aspects of science, that you couldn't get funding for that. It was considered too ambitious. It was theoretical work and people didn't fund theoretical work. So after a couple of years as a graduate student at Berkeley, I set a different path. I said, okay, I'm going to go back to work in industry for a few years to mature, to figure out how to make institutional change because I was up against an institutional problem, not just a scientific problem. And that turned into a series
Starting point is 00:05:25 of successful businesses that I was involved with and started, including Palm and Handspring. These are some of the early handheld computing companies. And we were having a tremendous amount of success with that. But it was never my mission to stay in the handheld computing industry. I wanted to get back to neuroscience. And everybody who worked for me knew this. In fact, I told the investors, I'm only going to do this for four years, and they said, what? I said, yeah, that's it.
Starting point is 00:05:52 But it turned out to be a lot longer than that because of all the success we had. But eventually, I just extracted myself from it, and I said, I'm going to go, and I have so many years left in my life. So after having all that success in the mobile computing space, I started a neuroscience institute.
Starting point is 00:06:07 This is at the recommendation of some neuroscience friends of mine. So they helped me do that. And I ran that for three years, and now I've been running sort of a private lab just doing pure neuroscience for the last 17 years. That's Numenta, right? That's Numenta, yeah. And we've made some really significant progress in our goals, and the book documents some of the recent really significant discoveries we've made.
Starting point is 00:06:33 So am I right in thinking that you made enough money at Palm and Handspring that you could self-fund your first neuroscience institute? Or is that not the case? Did you have to go raise money? Well, it was a bit of both. Certainly, I was a major contributor. I wasn't the only one, but I didn't want the funding to be the driver of what we did and how we spent all our time. So at the institute, we had collaborations with both Berkeley and Stanford. We didn't get funds from them, but we did work with them on various things. And then we had, but that was mostly funded by myself. Numenta is still, I'm a major contributor to it, but there are other people who've invested
Starting point is 00:07:15 in Numenta. And we have one outside venture capitalist and several people, but I'm still a major contributor to it. I don't, I just view that as a sort of a necessary thing to get onto the science and not have to worry about it. Because when I was at Berkeley, what I was told over and over again, and I really came to understand this. In fact, I went and eventually, after that, when I was running the Redwood Neuroscience Institute, I went to Washington to talk about, to the National Science Foundation and the National Institute of Health and also to DARPA, who were the funders of
Starting point is 00:07:49 neuroscience. And everyone thought what we were doing, which is sort of big theory, large-scale theories of cortical function, that this was like the most important problem to work on. But everyone said they can't fund it for various reasons. And so over the years, I've come to appreciate that it's very difficult to be a scientist doing what we do with traditional funding sources. But we don't work outside of science. We partner with labs, and we go to conferences, and we publish papers. We do all the regular stuff. Right, right. Yeah, it's amazing how much comes down to funding or lack of funding and the incentives that would dictate whether something
Starting point is 00:08:26 gets funded in the first place. It's really, it's by no means a perfect system. This is a, it's a kind of an intellectual market failure. Yeah, it is fascinating. And we could have a whole conversation about that sometimes, perhaps, because I ask myself, why is it so hard? Why the people can't fund this? And there's reasons for it. And it's a complex, strange thing when people were telling me, this is the most important thing anyone could be working on. And we think your approaches are great, but we can't fund that. And why is that? But I just accepted the way it was. I said, okay, this is the world I'm living in. I'm going to get one chance here. If I can't do this through working my way as a graduate student to getting a position in
Starting point is 00:09:09 university, how am I going to do it? And I said, okay, it's not what I thought, but this is what's going to be. Nice. Well, let's jump into the neuroscience side of it. Generally speaking, we're going to be talking about intelligence and how it's accomplished in physical systems. So let's start with a definition, however loose. What is intelligence in your view? So I didn't know and didn't have any pre-ideas about what this would be. It was a mystery to me.
Starting point is 00:09:42 But we've learned what a good portion of your brain is doing. And so we started the neocortex, which is about 70% of the volume of a human brain. And I now know what that does. And so I'm going to take that as my definition for intelligence here. What's going on in your neocortex is it's learning a model of the world, an internal recreation of all the things in the world that you know of. And how it does that's the key in what we've discovered. But it's this internal model. And intelligence requires having an internal model of the world in your head. And it allows you to recognize where you are. It allows you to act on things. It allows you
Starting point is 00:10:21 to plan and think about the future. So if I'm going to say, what happens when I do this? The model tells you that. So to me, intelligence is just about having a model in your head and using that for planning and action. It's not about doing anything particular. It's about understanding the world. Yeah, that's interesting. I think most people would, that's kind of an internal definition of intelligence, but I think most people would reach for an external one or a, you know, kind of a functional one that has to take in the environment. I mean, something about being able to flexibly meet your goals under a range of conditions, you know, more flexibly than rigidly. I guess there's rigid forms of intelligence, but when we're talking about anything like general intelligence, we're talking about something that is not merely
Starting point is 00:11:12 hardwired and reflexive, but flexible. Well, yes, but if you have an internal model of the world, you had to learn it, at least from a human point of view. There's some things we have built in when we're born, but the vast majority of what you and I know, Sam, is learned. We didn't know what a computer was when you're born. You don't know what a coffee cup is. You don't know what a building is. You don't know what doors are. You don't know what computer codes are.
Starting point is 00:11:36 None of this stuff. Everything that – almost everything we interact with in the world today, even language. We don't know any particular language when we're born. We don't know mathematics. So we had to learn all these things. So if you want to say there might be an internal model that wasn't learned, well, that's pretty trivial. But I'm talking about models that are learned, and you have to interact with the world to learn it. You can't learn it without being present in the world, without having an embodiment, without moving about, touching and seeing and hearing things. So a large part of what people think about, like you brought up, is, okay, you know, we
Starting point is 00:12:07 are able to solve a goal. But that's what a model lets you to do. That is not what intelligence itself is. Intelligence is having this ability to solve any goal, right? Because if your model covers that part of the world, you can figure out how to manipulate that part of the world and achieve what you want. So I'll give you a little further analogy. It's a little bit like computers. When we talk about like a universal Turing machine or what a computer is, it's not defined by what the computer is applied to do. It's like a computer isn't something that solves a particular problem.
Starting point is 00:12:38 A computer is something that works on a set of principles. And that's how I think about intelligence. It's a modeling system that works on a set of principles. And that's how I think about intelligence. It's a modeling system that works on a set of principles. Those principles can exist in a mouse and a dog and a cat and a human and probably birds, but don't focus on what those animals are doing. Yeah, I think it's important to point out that a model need not be a conscious model. In fact, most of our models are not conscious and might not even be, in principle, available to consciousness. Although I think at the boundary, something that you'd say is happening entirely in the dark does have a kind of, or can have a kind of liminal conscious
Starting point is 00:13:20 aspect. So I mean, to take, you know, the coffee cup example, this leads us into a more granular discussion of what it means to have a model of anything at the level of the cortex. But if I reach for my coffee cup and grasp it, the ordinary experience of doing that is something I'm conscious of. I'm not conscious of all of the prediction that is built into my accomplishing that and experiencing what I experience when I touch a coffee cup. And yet it's prediction that is required, having some ongoing expectation of what's going to happen there when each finger touches the surface of the cup, that allows for me to detect any error there or to be surprised by something truly anomalous. So if I reach for a coffee cup
Starting point is 00:14:11 and it turns out that's a, you know, it's a hologram of a coffee cup and my hand passes right through it, the element of surprise there seems predicated on some ongoing prediction processing to which the results of my behavior is being compared. So maybe you can talk about what you mean by having a model at the level of the cortex and how prediction is built into that. Yeah. Well, my first book, which I published like 14 years ago called Law and Intelligence, was just about that topic. It was about how it is the brain is making all these predictions all the time and all your sensory modalities, and you're not aware of it. And so that's sort of the foundation. And you can't make a prediction without a model. I mean, a model, to make a prediction, you had to have some expectation, the expectation, whether you're not aware of it or not, but you have an expectation.
Starting point is 00:15:04 And that has to be driven from some internal representation of the world that says, hey, you're about to touch this thing. I know what it is. It's supposed to feel this way. And even if you're not aware that you're doing that. One of the key discoveries we made, and this was maybe about eight years ago, we had to get to the bottom, like how do neurons make predictions? What is the physical manifestation of a prediction in the brain? And most of these predictions, as you point out, are not conscious. You're not aware of them. They're just happening. And if something is wrong, then your attention is drawn to it. So if you felt the coffee cup and there was a little
Starting point is 00:15:39 burr on the side or a crack, and you didn't know that was expected, that you'd say, oh, there's a crack. What was the brain doing when it was making that prediction? And we have a theory about this, and I wrote about it in the book a bit. And it's a beautiful, I think it's a beautiful theory, but it's basically most of the predictions that are going on in your brain, most of them, not all of them, but most of them, happen inside individual neurons. It is internal to the individual neurons. Now, not a single neuron can predict something, but an ensemble of neurons do this. But it's an internal state. And we wrote a paper that came out in 2016, which is called, Why Do Neurons Have So Many Synapses? And what we posit in that paper, and I'm pretty sure this is correct, is that neurons have these thousands of synapses.
Starting point is 00:16:35 Most of those synapses are being used for prediction. And when a neuron recognizes a pattern and says, okay, I'm supposed to be active soon. I should be becoming active soon. If everything is according to our model here, I should be becoming active soon. And it goes into this internal state. The neuron itself is saying, okay, I'm expecting to become active. And you can't detect that consciously. It's internal to the, it's essentially just a depolarization or change of the voltage of the neuron. And so neuron. But we showed how the network of these neurons, what'll happen is if your prediction is correct, then a small subset of the neurons become active. But if the prediction is incorrect, a whole bunch of neurons become active at the same
Starting point is 00:17:17 time. And then that draws your attention to the problem. So it's a fascinating problem, but most of the predictions going on in your brain are not accessible outside of individual neurons. So there's no way you could be conscious about it. Hmm. I guess most people are familiar with the general anatomy of a neuron where you have this spindly-looking thing where there's a cell body and there's a long process, the axon leading away, which carries the action potential if that neuron fires to the synapse and communicates neurotransmitters to other neurons. But on the other side of, in the standard case, on the other side of the cell body, there's this really, often really profuse arborization of dendrites, which is kind of
Starting point is 00:18:07 the mad tangle of processes which receive information from other neurons to which this neuron is connected. And it's the integration of information on that side, but before that neuron fires, the change, the probability of its firing, that's the place you are locating the full set of predictive changes or the full set of changes that constitute prediction in the case of a system of neurons. Yeah. Essentially, for many years, people looked at the connections on the dendrites, on that bushy part called synapses. And when they activated a synapse, most of the synapses were so far from the cell body that they didn't really have much of an effect.
Starting point is 00:18:55 They didn't seem like they could make anything happen. But there are thousands and thousands of them out there, but they don't seem powerful enough to make anything occur. And what was discovered basically over the last 20 years, that there's a second type of spike. So you mentioned the one that goes down the axon, that's the action potential, but there are spikes that travel along the dendrites. And so basically what happens is the individual sections of the dendrite, like little branches of this tree, each one of them can recognize patterns on their own. They can recognize hundreds of separate patterns on these different branches. And they can cause this spike to travel along the dendrite.
Starting point is 00:19:37 And that lowers the, changes the voltage of the cell body a little bit. And that is what we call the predictive state. The cell is a little bit. And that is what we call the predictive state. The cell is like prime. It says, oh, if I fire, I'm ready to fire. And it's not actually a probability change. It's the timing. And so a cell that's in this predictive state that says, I think I should be firing now or very shortly, if it does generate the regular spike, the action potential, it does it a little bit sooner than it would have otherwise.
Starting point is 00:20:07 And it's timing that is the key to making the whole circuit work. We're getting pretty down in the weeds here about neuroscience. I don't know if all your readers or your listeners will appreciate that. Yeah, no, I think it's useful, though. More weeds here. One of the novel things about your argument is that it was inspired by some much earlier theorizing. You mark your debt to Vernon Mountcastle. But the idea is that there's a common algorithm operating more or less everywhere at the level of the cortex. That is, it's more or less that the cortex is doing essentially the
Starting point is 00:20:46 same thing, whether it's producing language or vision or any other sensory channel or motor behavior. So talk about the general principle that you spend a lot of time on in the book of just the organization of the neocortex into cortical columns and the implications this has for how we view what the brain is doing in terms of sensory and motor learning and all of its consequences. This is, Vernon Mountcastle made this proposal back in the 70s. And it's just a dramatic idea. And it's an incredible idea. And so incredible that some people just refuse to believe it. But other people really think it's a tremendous discovery.
Starting point is 00:21:34 But what he noticed was if you look at the neocortex, if you could take one out of your head or out of a human's head, it's like a sheet. It's about two and a half millimeters thick. It is about the size of a large dinner napkin or 1,500 square centimeters. And if you could fold it, lay it flat. And the different parts of it, like they do different things. There's parts that do vision, there's parts that do language, and parts that do hearing and so on.
Starting point is 00:21:59 But if you cut into it and you look at the structure in any one of these areas, it's very complicated. There are dozens of different cell types, but they're very prototypically connected, and they're arranged in certain patterns and layers and different types of things. So it's a very complex structure, but it's almost the same everywhere. It's not the same everywhere, but almost the same everywhere. and so this is not just true in a human neocortex but if you look at a rat's neocortex or a dog's neocortex or a cat or a monkey the same basic structure is there and what vernon malcos said is that all the parts of the neocortex are actually we think of them as doing things different things but they're actually all
Starting point is 00:22:42 doing some fundamental algorithm which is the same So hearing and touch and vision are really the same thing. He says, if you took part of the cortex and you hook it up to your eyes, you'll get vision. If you hook it up to your ears, you'll get hearing. If you hook it up to other parts of the neural cortex, you'll get language. And so he spent many years giving the evidence for this. He proposed further that this algorithm was contained in what's called a column. And so if you would take a small area of this neocortex, remember it's like two and a half millimeters thick, you take a very sort of skinny little one millimeter column out of it, that that is the processing element. And so our human neocortex, we have about 150,000 of these columns. Other animals have more or less.
Starting point is 00:23:30 People should picture something resembling a grain of rice in terms of scale here. Yeah, yeah. I sometimes say take a piece of skinny spaghetti, like angel hair pasta or something like that, and cut it into two little two and a half millimeter lengths and stack them side by side. Now, the funny thing about columns
Starting point is 00:23:45 is you can't see them. They're not visual things. You can't look under a microscope, you won't see it. But he pointed out why they're there. It has to do with how they're connected. So all the cells in one of these little millimeter pieces of rice or spaghetti, if you will, are all processing the same thing. And the next piece of rice over processing something different and the next piece of rice over processing something different. And so he didn't know what was going on in the cortical column. He articulated the architecture. He talked about the evidence that this exists. He said, here's the evidence why these things are all doing the same thing. But he didn't know what it was. And it's kind of hard to imagine what it is that this algorithm could be doing. But that
Starting point is 00:24:31 was essentially the core of our research. That's what we've been focused on for close to 20 years. So it's also hard to imagine the microanatomy here, because in each one of these little columns, there's something like 150,000 neurons on average. And if you could just unravel all of the connections there, you know, the tiny filaments of nerve endings, what you would have there is on the order of kilometers in length, you know, all wound up into that tiny structure. you know, all wound up into that tiny structure. So it's a strange juxtaposition of simplicity and complexity, but there's certainly a mad tangle of processes in there. Yeah. This is why brains are so hard to study. You know, if you look at another organ in the body, whether it's the heart or the liver or something like that, and you take a little
Starting point is 00:25:21 section of it, it's pretty uniform. You know what I'm saying? But here, if you take a teeny, teeny piece of the cortex, it's got this incredible complexity in it, which is not random. It's very specific. And so, yeah, it's hard to wrap your heads around how complex it is. But we need it to be complex because what we do as humans is extremely complex. And we shouldn't be fooled that we're just a bunch of neurons that are doing some mass action. No, there's a very complex processing going on in your brain. It's not just a blob of neurons that are pulsating. Very detailed mechanisms that are undergoing it. And we figured out what some of those are.
Starting point is 00:26:04 mechanisms that are undergoing it. And we figured out what some of those are. So describe to me what you mean by this phrase, a reference frame. What does that mean at the level of the cortex and cortical columns? Yeah. So we're jumping to the endpoint, because that's not where we started. We were trying to figure out how cortical columns work. that's not where we started. We were trying to figure out how cortical columns work. And what we realized is that they're little modeling engines. Each one of these cortical columns is able to build a model of its input. And that model is what we would call a sensory motor model. That is, it's getting input. Let's assume it's getting input from your finger, right? A tip of your finger. One of the columns is getting input from the tip of your finger. And as your finger, right? A tip of your finger, one of the columns is getting input from the tip of your finger. And as your finger moves and touches something, the input changes. But it's not just efficient to how the input changes. For you to build a model of the object
Starting point is 00:26:54 you're touching, and I use the coffee cup example quite a bit because that's how we used it. If you move your finger over the coffee cup and you're not even looking at the coffee cup, you could learn a model of the coffee cup. You could feel just with one finger, you could feel like, oh, this is what its shape is. But to do that, your brain, that cortical column, your brain as a whole, but that cortical column individually has to know something about where your finger is relative to the cup. It's not just a changing pattern that's coming in. It has to know how your finger's moving and where your finger is as it touches it. So the idea of a reference frame is a way of noting a location. You have to have a location signal.
Starting point is 00:27:30 You have to have some knowledge about where things are in the world relative to other things. In this case, where is your finger relative to the object you're trying to touch, the coffee cup? And we realized that for your brain to make a prediction of what you're going to feel when you touch the edge of the cup, and again, a prediction of what you're going to feel when you touch the edge of the cup, and again, you mentioned earlier, you're not conscious of this. You'd reach the cup and you just, but your brain's predicting what all your fingers are going to feel.
Starting point is 00:27:53 It needs to know where the finger is going to be. And it has to know what the object is. It's a cup. It needs to know where it's going to be. And that requires a reference frame. A reference frame is just a way of noting a location. It's saying, relative to this cup, your finger is over here, not over there, not on the handle, up at the top, whatever it is. And this is a deduced property. We can say for certainty that this has to exist. If your finger is going to make a prediction when it reaches and touches the coffee cup, it needs to know where the finger is. That location has to be relative to the cup. So we can just say for certainty that there need to be reference frames in the brain. And this is not a controversial idea.
Starting point is 00:28:31 Perhaps it's novel is that we realized that these reference frames exist in every cortical column. And it's the structure of knowledge. It applies to not just what your finger feels on a coffee cup and what you see when you look at it, but also how you arrange all your knowledge in the world is stored in these reference frames. And so we're jumping ahead here many steps, but when we think and when we posit, when we try to reason in our head, even my language right now is where the neurons are walking through locations in reference frames, recalling the information stored there. And that's what comes into your head, recalling the information stored there. And that's what comes into your head, or that's what you say. So it becomes the core reference,
Starting point is 00:29:14 the reference frame becomes the core structure for the entire, everything you do. It's knowledge about the world is in these reference frames. Yeah, you make a strong claim about the primacy of motion, right? Because everyone knows that there's part of the cortex devoted to motor action, we refer to it as the motor cortex, and distinguish it from sensory cortex in that way. But it's also true that other regions of the cortex, and perhaps every region of the cortex, does have some connection to lower structures that can affect motion, right? So it's not that it's just motor cortex that's in the motion game. And by analogy or by direct implication, you think of thought as itself being a kind of movement in conceptual space, right? So there's a mapping of the sensory world that can really
Starting point is 00:30:05 only be accomplished by acting on it, you know, and therefore moving, right? So the only way to map the cup, you know, is to touch it with your fingers in the end. There is an analogous kind of motion in conceptual space and, you know, even abstract you know, abstract ideas like, I think some of the examples you give in the book are like, you know, democracy, right? You know, or money or what, how we understand these things. So, let's go back to the first thing you said there. The idea that there's motor cortex and sensory cortex is sort of no longer considered right. As you mentioned, we, the neurons that in these cortical columns, there are certain neurons that are the motor output neurons.
Starting point is 00:30:47 These are in a particular layer five, as they're called. And so in the motor cortex, they were really big and they project to the spinal cord and say, oh, that's how you move your fingers. But if you look at the neurons, the columns in the visual cortex, the parts that get input from the eyes, they have the same layer five cells. And these cells project to a part of the brain called the superior colliculus, which is what controls eye motion. So this goes against the original idea that, oh, there's sensory cortex and motor cortex. No one believes that. Well, I don't know nobody,
Starting point is 00:31:18 but very few people believe that anymore. As far as we know, every part of the cortex has a motor output. And so every part of the cortex is getting some sort of input and it has some motor output. And so the basic algorithm of the cortex is a sensory motor system. It's not divided. It's not like we have sensory areas and motor areas. As far as we know, ever it's been seen, there's these motor cells everywhere. So we can put that aside. So we can put that aside.
Starting point is 00:31:52 Now, I can very clearly walk you through and in some sense prove from logic that when you're learning what a coffee cup feels like, and I could even do this for vision, that you have to have this idea of a reference frame, that you have to know where your finger is relative to the cup, and that's how you build a model of it. And so we can build out this cortical column that explains how it does that. How do the parts of your cortex representing your fingers are able to learn the structure of a coffee cup? Now, Mountcastle, go back to him, he said, look, it's the same algorithm everywhere. And he says, it looks the same everywhere. So it's the same algorithm everywhere. So that's just sort of say, hmm, well, if I'm thinking about something that doesn't seem like a sensory motor system, like I'm not touching something or looking, I'm just thinking about something.
Starting point is 00:32:30 If Mountcastle was right, then the same basic algorithm would be applying there. So that was one constraint. Like, well, that, you know, and the evidence is that Mountcastle's right. I mean, the physical evidence suggests he's right. It just becomes a little bit odd to think like, well, how is language like this? And how is mathematics like, you know, touching a coffee cup? But then we realize that it's just reference frames are a way of storing everything. And the way we move through a reference frame, it's like, how do you move from one location? How do the neurons activate one location after another location after another location? We do that to this idea of movement. So I'm moving, if I want to access the locations on a coffee cup, I move my finger. But the same concept could apply to mathematics or to politics, but you're not actually physically
Starting point is 00:33:16 moving something, but you're still walking through a structure. A good bridge example is if I say to you, imagine know, imagine your house and I ask you to walk, you know, tell me about your house. What you'll do is you'll mentally imagine walking through your house. It won't be random. You just won't have random thoughts come to your head. But you will mentally imagine walking through your house. And as you walk through your house, you'll recall what is supposed to be seen in different
Starting point is 00:33:40 directions. You can say, oh, I'll walk in the front door and I'll look to the right. What do I see? I'll look to the left. What do I see? This is sort of an example you could relate it to something physically you could move to, but that's pretty much what's going on when you're thinking about anything. If you're thinking about your podcast and how you get more subscribers, you have a model of that in your head and you are trying it out, thinking about
Starting point is 00:34:01 different aspects by literally invoking these different locations and reference frames. And so that's sort of the core of all knowledge. Yeah, it's interesting. I guess back to Mountcastle for a second. One piece of evidence in favor of this view of a common cortical algorithm is the fact that adjacent areas of cortex can be appropriated by various functions. If you lose your vision, say, classical visual cortex can be appropriated by other senses, and there's this plasticity that can ignore some of the previous boundaries
Starting point is 00:34:37 between separate senses in the cortex. Yeah, that's right. There's this tremendous plasticity, and you can also recover from various sorts of trauma and so on. I mean, there's right. There's this tremendous plasticity, and you can also recover from various sorts of trauma and so on. I mean, there's some rewiring that has to occur, but it does show that whatever the circuitry in the visual cortex was, quote, if you were a sighted person, what it would do. If you're not a sighted person, well, it'll just do something else. And so that is a very very strong
Starting point is 00:35:05 argument for that there's a famous scientist um bakarita who did an experiment where he i'm trying to remember the animal he used um maybe you recall it but anyway um it'll come to me a ferret i think it was a ferret we took the they took a before the animal's born he took the optic nerve and ran it over to one part of the a different part of the neocortex and took the auditory nerve and ran it to a different part of the neocortex, you know, basically rewired the animal. I'm not sure we do these experiments today, but, and, you know, and the argument was that the animals, you know, still saw and still heard and so on, maybe not as well as an unaltered one, but the evidence was that, yeah, that really works. Yeah, so what is genetically determined and what is learned here? I mean, it seems that the genetics at minimum are determining what is hooked up to what initially, right? You know,
Starting point is 00:35:57 borrowing. Yeah, roughly, roughly, that's right. I think, you know, like where do the eyes, the optic nerve from the eyes, where do they project? And where do the regions that get the input from the eyes, where do they project? And so this rough sort of overall architecture is specified. And as we just talked through trauma and other reasons, sometimes that architecture can get rewired. I think also the basic algorithm that goes on in each of these cortical columns, the circuitry inside the neocortex, is pretty well determined by genetics.
Starting point is 00:36:29 In fact, one of Myocast's arguments was that the human neocortex got large, and we have a very large one relative to our body size, just because all evolution had to do was discover just make more copies of these columns. You don't have to do anything new, just make more copies. And that's something easy for genes to specify. So human brains got large quickly in evolutionary time by that just replicate more of it type of thing. Okay, so let's go beyond the human now and talk about artificial intelligence. And before we talk about the risks or the imagined risks, tell me what you think the path looks like going forward. I mean, what are we doing now,
Starting point is 00:37:14 and what do you think we need to do to have our dreams of true artificial general intelligence realized? Well, today's AI, as powerful as it is and successful as it is, I think most senior AI practitioners will admit, and many of them have, that they don't really think they're intelligent. They're really wonderful pattern classifiers and they can do all kinds of clever things, but there are very few practitioners who would say, hey, this AI system that's recognizing faces is really intelligent. And there's sort of a lack of understanding what intelligence is and how to go forward and how do you make a system that could solve general problems, could do more than one thing, right? And so in the second part of my
Starting point is 00:38:01 book, I lay out what I believe are the requirements to do that. And my approach has always been, for 40 years, has been like, well, I think we need to first figure out what brains do and how they do them. And then we'll know how to build intelligent machines because we just don't seem able to intuit what an intelligent machine is. So I think the way I look at this problem, if we want to make, you know, what's the recipe for making an intelligent machine, is you have to say, what are the principles by which the brain works that we need to replicate and which principles don't we need to replicate? And so I made a list of these in the book, but if you can think of a very high level, they have to have some sort of embodiment. They have to have the ability to move their sensors somehow in the world.
Starting point is 00:38:46 You know, you can't really learn how to use tools and how to, you know, run factories and how to do things unless you can move in the world. And it requires these reference frames I was talking about because movement requires reference frames. That's not a controversial statement. It's just a fact. You're going to have to know where things are in the world. And then the final, there's a set of things, but one of the other big ones, which we haven't talked about yet, and which is where the
Starting point is 00:39:15 title of the book comes from, A Thousand Brains, is that the way to think about our near cortex, it has 150,000 of these columns. We have essentially 150,000 separate modeling systems going on in our brain, and they work together by voting. And so that concept of a distributed intelligence system is important. We're not just one thing. It feels like we're one thing, but we're really 150,000 of these things. And we're only conscious of being one thing, but that's not really what's happening under the covers. So those are some of the key ideas. I would just stick to very, very high ideas.
Starting point is 00:39:50 It has to have an embodiment, it has to be able to move its sensors, it has to be able to organize information in reference frames, and it has to be distributed. And that's how we can do multiple sensors and sensory integration, things like that. Hmm.
Starting point is 00:40:03 can do multiple sensors and sensory integration, things like that. I guess I question the criteria of embodiment and movement, right? I mean, I understand that practically speaking, that's how a useful intelligence can get trained up in our world to do things physically in our world. But it seems like you could have a perfectly intelligent system, i.e. a mind that is turned loose on simulated worlds and are capable of solving problems that don't require effectors of any kind. I mean, chess is obviously a very low-level analogy, but just imagine a thousand things like chess that represent real theory building or cognition in a box. Yeah, I think you're right. And so, when I use the word movement or embodiment,
Starting point is 00:41:00 and I'm careful to define it in the book because it doesn't have to be physical. And I'm careful to define it in the book because it doesn't have to be physical. It, you know, example I gave, you can imagine an intelligent agent that lives in the internet and a movement is following links, right? It's not a physical thing, but there's still this conceptual mathematical idea of what it means to move. And so you're changing the location of some representation. And that could be virtual. It could be, you know,
Starting point is 00:41:29 it doesn't have to have a physical embodiment. But in the end, you can't learn about the world just by looking at a set of pictures. That's not going to happen. You can learn to classify pictures. So some AI systems will have to be physically embodied like a robot, I guess, if you want. Many will not be. Many will be virtual. But they all have this internal process which I could point to the thing that says, here's where the reference frame is, here's where your
Starting point is 00:41:57 current location is, here's how it's moving to a new location based on some movement vector. You know, like a verb, a word, you can think of that as like an action. And so you can have an action that's not physical, but it's still an action, and it moves to a new location in this internal representation. Right, right. Okay, well, let's talk about risk,
Starting point is 00:42:16 because this is the place where I think you and I have very different intuitions. You are, as far as I can tell from your book, you seem very sanguine about AI risk. And really, you seem to think that the only real risk, the serious risk of things going very badly for us is that bad people will do bad things with much more powerful tools. So the heuristic here would be, you know, don't give your super intelligent AI to the next Hitler, because that would be bad. But other than that, the generic
Starting point is 00:42:52 problem of self-replication, which you talk about briefly, and you point out we face that on other fronts, like with, you know, with the pandemic we've been dealing with. I mean, so natural viruses and bacteria or computer viruses, I mean, anything that can self-replicate can be dangerous. But that aside, you seem quite confident that AI will not get away from us, that there won't be an intelligence explosion, and we don't have to worry too much about the so-called alignment problem. And at one point, you even question whether it makes sense to expect that we'll produce
Starting point is 00:43:29 something that can be appropriately called superhuman intelligence. So perhaps you can explain the basis for your optimism here. So I think what most people, and perhaps yourself, have fears about is they use humans as an example of how things can go wrong. And so we think about the alignment problem, or we think about motivations of an AI system. Well, okay, does the AI system have motivations or not? Does it have a desire to do anything? Now, as a human, an animal, we all have desires, right? But if you take apart what parts of the human brain are doing, different parts, there's some parts that are just building this model of the world. And this is the core of our
Starting point is 00:44:17 intelligence. This is what it means to be intelligent. That part itself is benign. It has no motivations on its own. It doesn't desire to do anything. I use an example of a map. A map is a model of the world. And a map can be a very powerful tool for some to do good or to do bad. But on its own, the map doesn't do anything. So if you think about the neocortex, on its own, it sits on top of the rest of your brain. And the rest of your brain is really what makes us motivated. It gets us, you know, we have our good sides and our bad sides, you know, our desire to maintain our life and have sex and aggression and all this stuff. The neocortex is just sitting there. It's like a map. It says,
Starting point is 00:45:01 you know, I understand the world and you can use me as you want. So when we build intelligent machines, we have the option, and I think almost the imperative, not to build the old parts of the brain, too. Why do that? We just have this thing which is inherently smart, but on its own doesn't really want to do anything. And so some of the risks that come about from people's fears about the alignment problem specifically is that the intelligent agent will decide on its own or decide for some reason to do things that are in its best interest and not in our best interest,
Starting point is 00:45:38 or maybe it'll listen to us but then not listen to us or something like this. I just don't see how that can physically happen. And for people, most people don't understand this separation. They just assume that this intelligence is wrapped up in all the things that make us human. The intelligence explosion problem is a separate issue. I'm not sure which one of those you're more worried about. Yeah, well, let's deal with the alignment issue first. I mean, I do think that's more critical, but let me see if I can capture what troubles me about this picture you've painted here. It seems that you're, to my mind, you're being strangely anthropomorphic on one side, but not anthropomorphic enough on the other. I mean, so like, you know, you think that to understand intelligence and actually truly implement it in machines, we really have to be
Starting point is 00:46:34 focused on ourselves first, and we have to understand how the human brain works and then emulate those principles pretty directly in machines. That strikes me as possibly true, but possibly not true. And if I had to bet, I think I would probably bet against it. Although even here, you seem to be not taking full account of what the human brain is doing. I mean, we can't partition reason and emotion as clearly as we thought we could hundreds of years ago. And in fact, certain emotions, certain drives are built into our being able to reason effectively. I think that's, you know, I'll take exception to that. I know this is an opinion that you had, Lisa Barrett, on your program recently.
Starting point is 00:47:24 Yeah, Antonio Damasio is the person who's banged on about this the most. Yeah, I know. And I just disagree. You can separate these two. And I can say this because I understand actually what's going on in the neurocortex. And I can see what's... I have a very good sense of what these actual neurons are actually doing when it's modeling the world and so on. And you do not, it does not require this emotional component. A human does. Now, you say, you know, on one hand, I don't argue we should replicate the brain. I say we should replicate the structures of the neocortex, which is not replicating the brain. It's just one part of the brain. not replicating the brain. It's just one part of the brain. And so I'm specifically saying, you know, I don't really care too much about how the spinal cord works or how, you know, the brainstem does this or that. It's interesting. Maybe I know a little bit about it, but that's not important. The cortex sits on top of another structure and the cortex does its own thing and they interact. Of course they interact. And our emotions affect what we learn and what we
Starting point is 00:48:23 don't learn. But it doesn't have to be that way in a system, another system that we build. That's the way humans are structured. Yeah, okay. So I would agree with that, except the boundary between what is an emotion or a drive or a motivation or a goal and what is a value-neutral mapping of reality, I think that boundary is perhaps harder to specify than you think it is, and that certain of these things are connected, right? Which is to... I mean, here's an example. This is probably not a perfect analogy, but this gets at some of the surprising features of cognition that may await us. So we think
Starting point is 00:49:05 intuitively that understanding a proposition is cognitively quite distinct from believing it, right? So like I can give you a statement that you can believe or disbelieve or be uncertain about, and I can say, you know, there's 2 plus 2 equals 4, 2 plus 2 equals 5, and that can give you some gigantic number and say this number is prime. And presumably, in the first condition, you'll say, yes, I believe that. In the second, you'll say, no, that's false. And in the third, you won't know whether or not it's prime or not. So those are distinct states that we can intuitively differentiate. But there's also evidence to suggest that merely
Starting point is 00:49:46 comprehending a statement, if I give you a statement and you parse it successfully, the parsing itself contains an actual default acceptance of it as true, and rejecting it as false is a separate operation added to that. I mean, there's not a ton of evidence for this, but there's certainly some behavioral evidence. So if I put you in a paradigm where we gave you statements that were true and false, and all you had to do was judge them true and false, and they were all matched for complexity, so, you know, 2 plus 2 equals 4 is no more or less complex than 2 plus 2 equals 5, but it'll take you longer, systematically longer, to judge very simple statements to be false than to judge them to be true, suggesting
Starting point is 00:50:32 that you're doing a further operation. Now, we can remain agnostic as to whether or not that's actually true, but if true, it's counterintuitive that merely understanding something entails some credence, epistemic credence given to it by default, and that to reject it as false represents a subsequent act. But that's the kind of thing that already we're on territory that is not coldly rational. Some of the all-too-apish appetites have kind of crept into cognition here in ways that we didn't really budget for. And so the question is just how much of that is avoidable in building a new type of mind? So I haven't heard of that. But to me, none of these things are surprising in any way. If you start thinking about the brain as basically trying to build models, it's constantly trying to build models. In fact, as you walk around your life day to day, moment to moment, and you see
Starting point is 00:51:38 things, you're building the model. The model is being constructed. Even like, where are things in the refrigerator right now? Your brain will update. You open the refrigerator, oh, the milk's on the left today, whatever. And so if someone gives you a proposition like two plus two equals five, I don't know what the evidence that you believe it and then falsify it, but I certainly imagine you can imagine it trying to see if it's right. It'd be like me saying to you, hey, Sam, the milk was on the right in your refrigerator. And you'd have to think about it for a second. You'd say, well, let me think. No, last time I saw it was on the left. No, that's wrong. But you would walk through the process of trying to imagine it and trying to see,
Starting point is 00:52:12 does that fit my model? And yes or no. And it's not surprising to me that you would have to process it the way as if it was true. It's just a matter of saying, can you imagine this? Go imagine it. Do you think it's right? It's not like I believe that now I've falsified it. It's more like- Well, actually, I'll just give you one other datum here because it's just intellectually interesting and socially all too consequential. This effect goes by several names, I think, but one is the illusory truth effect, which is even in the act of disconfirming something to be false, some specious rumor or conspiracy theory, merely having to invoke it, I mean, have people entertain the concept again, even in the context of debunking it,
Starting point is 00:52:57 ramifies a belief in it in many, many people. It's just, it becomes harder to discredit things because you have to talk about them in the first place. Yeah. I mean, so look, we're talking about language here, right? And in language, so much of what we humans know is via language. And we have no idea if it's true when someone says something to you, right? How do you know? And so you have to, so I mean, I gave an example, like I've never been to the city of Havana. Well, I believe it's there. I believe it's true, but I don't know.
Starting point is 00:53:29 I've never been there. I've never actually touched it or smelled it or saw it. So maybe it's false. So I just, I mean, this is one of the issues we have. I have a whole chapter on false beliefs because so much of our knowledge of the world is built up on language and the default assumption under language that if someone says something, it's true. It's like it's a pattern in the world. You're going to accept it.
Starting point is 00:53:51 If I touch a coffee cup, I accept that that's what it feels like. And if I look at something, I accept that's what it looks like. Well, someone says something, my initial acceptance is, okay, that's what it is. So, you know, and then instead of the fact, well, someone says something that's false, of course, well, that's a problem because just by the fact that I've experienced it, it's now part of my world model. And if that's what you're referring to, I can see this is really a problem of language we face. And this is the root cause of almost all of our false beliefs, is that someone just says something enough times, and that's good
Starting point is 00:54:25 enough. And you have to seek out contrary evidence for it. Yeah, sometimes it's good enough even when you're the one saying it. You just overhear the voice of your own mind saying it. And no, I know. It's been proven that everyone is susceptible to that kind of distortion of our beliefs, especially our memories. Just remembering something over and over again changes it. Yeah. Okay, so let's get back to AI risk here, because here's where I think you and I have very different
Starting point is 00:54:55 intuitions. And the intuition that many of us have, the people who have informed my views here, people like Stuart Russell, who you probably know at Berkeley, and Nick Bostrom and Eliezer Yudkowsky and just lots of people in this spot worrying about the same thing to one or another degree. a second chance to create a truly autonomous superintelligence, right? It seems that in principle, this is the kind of thing you have to get right on the first try, right? And having to get anything right on the first try just seems extraordinarily dangerous, because we rarely, if ever, do that when doing something complicated. And another way of putting this is that it seems like in the space of all possible super-intelligent minds, there are more ways to build one that isn't perfectly aligned with our long-term well-being than there are ways to build one that is perfectly
Starting point is 00:56:01 aligned with our long-term well-being. And from my point of view, your optimism and the optimism of many other people who take your side of this debate is based on, is not really taking the prospect of intelligence seriously enough and the autonomy that is intrinsic to it. If we actually built a true general intelligence, what that means is that we would suddenly find ourselves in relationship to something that we actually can't perfectly understand. It's like it will be
Starting point is 00:56:41 analogous to a strange person walking into the room. You know, you're in relationship, and if this person can think a thousand times or a million times faster than you can, and has goals that are less than perfectly aligned with your own, that's going to be a problem eventually. that's going to be a problem eventually. We can't find ourselves in a state of perpetual negotiation with systems that are more competent and powerful and intelligent than we are. I think there's two mistakes in your argument.
Starting point is 00:57:15 The first one is you say my intuition and your intuition. I think most of the people who have this fear have an intuition about what might happen. I don't have an intuition. I'm basically my apartment. along with other subscriber-only content, including bonus episodes and AMAs and the conversations I've been having on the Waking Up app. The Making Sense podcast is ad-free and relies entirely on listener support.
Starting point is 00:57:52 And you can subscribe now at SamHarris.org.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.