Making Sense with Sam Harris - #151 — Will We Destroy the Future?

Episode Date: March 18, 2019

Sam Harris speaks with Nick Bostrom about the problem of existential risk. They discuss public goods, moral illusions, the asymmetry between happiness and suffering, utilitarianism, "the vulnerable wo...rld hypothesis," the history of nuclear deterrence, the possible need for "turnkey totalitarianism," whether we're living in a computer simulation, the Doomsday Argument, the implications of extraterrestrial life, and other topics. If the Making Sense podcast logo in your player is BLACK, you can SUBSCRIBE to gain access to all full-length episodes at samharris.org/subscribe.

Transcript
Discussion (0)
Starting point is 00:00:00 Thank you. of the Making Sense Podcast, you'll need to subscribe at SamHarris.org. There you'll find our private RSS feed to add to your favorite podcatcher, along with other subscriber-only content. We don't run ads on the podcast, and therefore it's made possible entirely through the support of our subscribers. So if you enjoy what we're doing here, please consider becoming one. Welcome to the Making Sense Podcast. This is Sam Harris. Today I am speaking with Nick Bostrom. Nick is someone I've been hoping to get on the podcast for quite some time. He is a Swedish-born philosopher with a background in theoretical physics and computational neuroscience and logic and AI and many other interesting intersecting topics. But officially, he's a professor of
Starting point is 00:01:16 philosophy at Oxford University, where he leads the Future of Humanity Institute. And this organization is a research center which is focused largely on the problem of existential risk. And today we get deep into his views on existential risk by focusing on three of his papers, which I'll describe in our conversation. We talk about what he calls the vulnerable world hypothesis, which gets us into many interesting tangents with respect to the history of nuclear deterrence and the possible need for what he calls turnkey totalitarianism. We talk about whether we're living in a computer simulation. He's the father of the now famous simulation argument. We talk about the doomsday
Starting point is 00:02:05 argument, which is not his, but it's one of these philosophical thought experiments that have convinced many people that we might be living close to the end of human history. We talk about the implications of there being extraterrestrial life out there in the galaxy, and many other topics. But all of it is focused on the question of whether humanity is close to the end of its career here, or near the very beginning. And I hope you'll agree that the difference between those two scenarios is one of the more significant ones we can find. Anyway, I really enjoyed talking to Nick. I find his work fascinating and very consequential, and that's a good combination.
Starting point is 00:02:51 And now I bring you Nick Bostrom. I am here with Nick Bostrom. Nick, thanks for coming on the podcast. Yeah, thanks for having me. So you are fast becoming, or not too fast, it's been years now that I've been aware of your work, but you are becoming one of the most provocative philosophers I can think of. And really, it's wonderful to read your work. And I want to introduce you, but perhaps to begin with, how do you view your work and what you focus on at this point? How do you summarize your interests as a philosopher?
Starting point is 00:03:32 Well, that's always been a challenge for me. Broadly speaking, I'm interested in big picture questions for humanity and figuring out which direction is up and which is down. That is, out of all the things you can be pushing on or pulling on in the world, which ones would actually tend to make things better in expectation? Yeah, and then various kind of sub-questions that come out from that ultimate quest to figure out which direction we should be heading in. Yeah, so when I think about your work, I see a concern that unifies much of it, certainly, with existential risk. And I don't know if this
Starting point is 00:04:07 is a phrase that you have popularized or if it's just derivative of your work, but how do you think of existential risk and why is it so hard for most people to care about? It's amazing to me that this is such an esoteric concern and you really have brought it into prominence. Yeah, I introduced the concept in a paper I wrote back in the early 2000s. The concept being that of a risk either to the survival of Earth originating in talent life or a risk that could permanently and drastically reduce our potential for desirable future developments. So in other words, something that could permanently destroy the future. I mean, even that phrase, I mean, you really have a talent for coming up with phrases
Starting point is 00:04:56 that are arresting and, you know, it's such a simple one. It permanently destroy the future. You know, there are probably more people working in my local McDonald's than are thinking about the prospect of permanently destroying the future. How long have you been focused on this particular problem? And again, why is it, there's something bewildering about trying to export this concern to the rest of culture, even to very, very smart people who claim to be worried about things like climate change. Why is existential risk still such an esoteric concern? Well, it's become less so over the last few years.
Starting point is 00:05:38 There is now actually a community of folk around the rationalist community, the EA community, you know, various academic centers. is an effective altruism yeah and about not not just these but but kind of radiating out from these a number of individuals that that are quite interested in this so i think the comparison to the mcdonald's restaurant would no longer be true now maybe it was true several mcdonald's years ago why it is well i guess you could ask that or you could ask why it's no longer the case. I mean, I don't know that the default should be, if we're looking at academia, that questions receive attention in proportion to their importance. I think that's just kind of a poor model of what
Starting point is 00:06:18 to expect from academic research. So one can ask why it's changed. I mean, on one level, you're asking people to care about the unborn if the time horizon is beyond their lives and the lives of their children, which seems on its face probably harder than caring about distant strangers who are currently alive. And we know we're already pretty bad at that. Is that a major variable here? Sure. It's an extreme case of a public good, right? So generally, in a simple model of the market economy, public goods tend to be undersupplied because the creator of them captures only a small fraction of the benefits. The global public goods are normally seen as the extreme of this. If all of humanity
Starting point is 00:07:05 benefits from some activity or is harmed by some activity, as in maybe the case of global warming or something like that, then the incentives facing the individual producer are just very dissociated from their overall consequences. But with existential risk, it's even more extreme, actually, because it's a transgenerational good in the sense that all future generations are also impacted by our decisions concerning what we do about existential risks. And they are obviously not in a position in any direct way to influence our decisions. They can't reward us if we do things that are good for them. So if one thinks of human beings as selfish, one would expect the good of existential risk reduction to be undersupplied.
Starting point is 00:07:47 Like it could imagine if somehow people could go back in time that future generations would be willing to like spend huge amounts of money to compensate us for our efforts to reduce X risk. But since you can't, that transaction is not possible, then there is this undersupply. So that could be one way of explaining why there's relatively little. And there was something about what you said about it's harder to care.
Starting point is 00:08:16 It's a little strange that caring should be something that requires effort. If one does care about it, it doesn't seem like it should be a straining thing to do. And if one doesn't care, then it's not clear why one should be, what motive one would have for trying to strain to start caring. It's a framing problem in many cases, so that there's a certain set of facts, let's say the reality of human suffering in some distant country that you have never visited, you have no connections there, this information can just be transmitted to you about the reality of the suffering. And it transmitted one way, you find that you don't care, and transmitted another way,
Starting point is 00:08:58 the reality of it and the analogy to your own life or the lives of your children can be made more salient. And so we know that, we know, you know, through the lives of your children can be made more salient. And so we know that we know, you know, through the work of someone like Paul Slovic, we know there are moral illusions here where people can be shown to care more about the fate of one little girl who's delivered to them in the form of a personalized story. And they'll care less about the fate of that same little girl plus her brother, and they'll care less still if you tell them about the little girl, her brother, and the 500,000 other children who are also suffering from a famine. And you just get this diminishment of altruistic impulse
Starting point is 00:09:39 and the amount of money they're willing to give to charity and all the rest, and it goes in the wrong direction. As you scale the problem, people care less. So we know we have some moral bugs in our psychology. Yeah. So the original paper about existential risk made the point that from a certain type of ethical theory, it looks like existential risk reduction is a very important goal. If you have a broadly aggregative consequentialist philosophy, say if you're a utilitarian, and if you work the numbers out, the number of possible future generations, the number of individuals in each of those that can live very happy lives, then you multiply that together. And then it looks like even a very small chance in the probability that we will eventually achieve this would have a very high expected value.
Starting point is 00:10:30 In fact, a higher expected value than any other impacts that we might have in more direct ways on the world here and now. So that reducing existential risk by, you know, one thousandth of one percentage point would be, from this utilitarian perspective, worth more than eliminating world hunger or curing cancer. Now, that, of course, says nothing about the question of whether this kind of utilitarian perspective is correct or is agreeable to us. But it just notes that that does seem to be an implication. I'm definitely a consequentialist of a certain kind, so we don't need to argue that point. But one thing that's interesting here, and this may be playing into it, is that there seems to be a clear asymmetry between how we value suffering and its mitigation, and how we value the mere preemption of well-being or flourishing or positive states.
Starting point is 00:11:29 So that suffering is worse than pleasure or happiness is good. You know, I think if you told most people, here are two scenarios for how you can spend the rest of your day. You can spend it as you were planning to, living within the normal bounds of human experience, or I can give you one hour of the worst possible misery followed by another hour of the deepest possible happiness. Would you like to sample the two extremes on the phenomenological continuum? I think most people would balk at this
Starting point is 00:12:03 because we think, we have a sense that suffering is, on some level, you know, in the limit, is worse than any pleasure or happiness could be. And so we look at the prospect of, let's say, you know, curing cancer and mitigating the suffering from that as being more important ethically than simply not allowing the door to close on future states of creativity and insight and beauty. Yeah, I think one might want to decompose different ways in which that intuition could be produced. So one might just be that for us humans, as we are currently constituted, it's a lot easier to create pain than to create a corresponding degree of pleasure. We might just evolutionarily be such that we have a kind of deeper bottom
Starting point is 00:12:52 than we have a height. It might be possible, if you think about it in biological term, in a short period of time to introduce more damage to damage reproductive fitness more than you could possibly gain within the same amount of time.
Starting point is 00:13:07 So if we are thinking about these vast futures, you want to probably factor that out in that you could re-engineer, say, human hedonic systems or the hedonic systems of whatever inhabitants would exist in this future so that they would have a much larger capacity for upside. Right. And it's not obvious that there would have a much larger capacity for upside. Right. And it's not obvious that there would be an asymmetry there.
Starting point is 00:13:28 Now, you might nevertheless think that even in some sense, equal amounts of pleasure and pain, and it's a little unclear exactly what the metric is here, that there would nevertheless be some more basically ethical reason why one should place a higher priority on removing the negative. A lot of people have intuitions about equality, say, in economic context, where helping the worst off is more important than further promoting the welfare of the best off. Maybe that's the source of some of those intuitions. Actually, there's one other variable here, I think, which is that there is no victim or, you know, beneficiary of the consequence of closing the door to the future. So if you ask
Starting point is 00:14:14 someone, well, what would be wrong with the prospect of everyone dying painlessly in their sleep tonight, and there are no future generations, there's no one to be bereaved by that outcome. There's no one suffering the pain of the loss or the pain of the deaths even. So people are kind of at a loss for the place where the moral injury would land. Yeah, so that is a distinction within utilitarian frameworks, between total utilitarians who think you basically count up all the good and subtract all the bad. And then other views that try to take a more so-called person-affecting perspective, where
Starting point is 00:14:57 what matters is what kind of happens to people, but coming into existence is not necessarily a benefit. And now I would say some kinds of existential catastrophe would have a continuing population of people who would be experiencing the bad. If you might say the world getting locked into some totalitarian, like really dystopian totalitarian regime, maybe there would be people living for a very long time, but just having much less good lives than could have existed.
Starting point is 00:15:31 So in some scenarios of existential catastrophe, there would still be inhabitants there. Yeah, no, I think it's pretty clear that destroying the future could be pretty unpleasant for people who are along for the ride. Now, I'd like just to harken back a few minutes ago, like on the general premise here. So I don't see it so much as a premise, this utilitarian view. I mean, in fact,
Starting point is 00:15:55 I wouldn't really describe myself as a utilitarian, but more just pointing out the consequence. There are various views about how we should reason about ethics, and there might be other things we care about as well, aside from ethics. And rather than directly trying to answer what do we have most reason to do all things considered, you might break it down and say, well, given this particular ethical theory, what do we have most reason to do given this other value or this other goal we might have? And then at the end of the day, you might want to add all of that up again. But insofar as we are trying to reason about our ethical obligations, I have a kind of normative uncertainty over different moral frameworks. And so the way I would try to go about making decisions from a moral point of view would be to think, I have this moral parliament model. It's a kind of metaphor,
Starting point is 00:16:45 but where you try to factor in the viewpoints of a number of different ethical theories, kind of in proportion to the degree to which you assign them probability. It's kind of interesting. When I'm out and about in the world, I usually have to make the case for utilitarianism, or at least you should consider this perspective. People are scope insensitive. the case for utilitarianism, or at least you should consider this perspective. Like people are scope insensitive. You should look at the numbers. If this thing has millions of people and this one only has hundreds of people being affected, clearly that, and yet when I'm back here at the headquarters, as it were, I usually am
Starting point is 00:17:18 the one who has to kind of advocate against utilitarian perspective because so many of my friends are so deeply dyed in the wool utilitarians yeah and so narrowly focused on x-risk mitigation but i feel that i'm always the the odd one out well you know i would love to get into a conversation with you about meta-ethics some other time because i you know i i think your views about the limits of consequentialism would be fascinating to explore. But I have so much I want to talk to you about with respect to X-Risk and a few of your papers that I think, let's just table that for another time. In fact, I don't even think we're going to be able to cover
Starting point is 00:17:57 your book Superintelligence. I mean, maybe if we have a little time at the end, we'll touch it. But I should just want to say that this book was incredibly influential on many of us in arguing the case for there being a potential existential risk with respect to the development of artificial intelligence and artificial general intelligence in particular. And so it's, you know, this is something, the reason why I wouldn't cover this with you for the entirety of this conversation is I've had several conversations on my podcast that have been deeply informed by your view. I mean, I've had Stuart Russell on, I've had Eliezer Yudkowsky on, and basically every
Starting point is 00:18:37 time I talk about AI, I consider what I say to be, you know, fairly derivative of your book, and I often remind people of that. So my audience will be familiar with your views on AI, even if they're not familiar with you. So if we have time, we'll get back to that. But what I really want to talk about are a few of your papers. The first is the vulnerable world hypothesis. Maybe I'll just name the papers here that I hope will cover. Vulnerable world, the second is, are you living in a computer simulation? And the third is, where are they? Which is your analysis of the Fermi problem, asking where is the rest of the intelligent life in the galaxy?
Starting point is 00:19:20 Let's start with the vulnerable world hypothesis. What do you mean by that phrase? Well, the hypothesis is, roughly speaking, that there is some level of technological development at which the world gets destroyed by default, as it were. So then what does it mean to get destroyed by default? I define something I call the semi-anarchic default condition, which is a condition in which there are a wide range of different actors with a wide range of different human recognizable motives.
Starting point is 00:19:52 But then, more importantly, two conditions hold. One is that there is no very reliable way of resolving global coordination problems. And the other is that we don't have a very extremely reliable way of preventing individuals from committing actions that are extremely strongly disapproved of by a great majority of other people. Maybe it's better to come at it through a metaphor. Yeah, the urn. The urn metaphor. So you can kind of think of the history of So you can kind of think of the history of technological discovery as the process of pulling balls out of a giant urn,
Starting point is 00:20:32 the urn of creativity. And we reach in and we get an idea out, and then we reach in and get another out. And we've extracted throughout history a great many of these balls. And the net effect of this has been hugely beneficial, I would say. This is why we now sit in our air-conditioned offices and struggle not to eat too much, rather than to try to get enough to eat in large parts of the world. But what if in this ball, in this area, there is a black ball in there somewhere?
Starting point is 00:21:05 Is there some possible technology that could be such that whichever civilization discovers it, invariably gets destroyed? Just to add a little color here, Nick. So in your paper, you refer to this as the urn of inventions. And we have been, as you say, pulling balls out as quickly as we can get our hands on them. And on some level, the scientific ethos is really just a matter of pulling balls out as fast as you can and making sure that everybody knows about them. We have this norm of transparency in science, and we have pulled out, thus far, only white or gray balls, and the white balls are the ones, or the technologies, or the memes, or the norms, or the social institutions that just have good
Starting point is 00:21:52 consequences, and the gray ones are norms and memes and institutions and, in most cases, technology that has mixed results or that can be used for good or for ill. And nuclear energy is a classic case where we can power our cities with it, but we also produce fantastic amounts of pollution that's difficult to deal with. And in the worst case, we build weapons. So I just want to give a little more context to this analogy. Yeah, and I guess most technologies are in some sense double-edged, but maybe the positive predominate. I think there might be some technologies that are mainly negative if you think of, I don't know, nerve gases or other tools.
Starting point is 00:22:35 But what we haven't so far done is extract a black ball, one that is so harmful that it destroys the civilization that discovers it and um what if there is such a black ball in the urn though i mean we can ask about how likely that is to be the case we can also look at what what is our current strategy with respect to this possibility and it seems to me that currently our strategy with respect to the possibility that the urn might contain a black ball is simply to hope that it doesn't. So we keep extracting balls as fast as we can. We have become quite good at that, but we have no ability to put balls back into the urn. We cannot uninvent our
Starting point is 00:23:17 inventions. So the first part of this paper tries to identify what are the types of ways in which the world could be vulnerable, the types of ways in which there could be some possible blackball technology that we might invent. And the first and most obvious type of way the world could be vulnerable is if there is some technology that greatly empowers individuals to cause sufficiently large quantities of destruction. Motivate this with a, or illustrate it by means of a historical counterfactual. We, in the last century, discovered how to split the atom and release the energy that
Starting point is 00:24:00 is contained within some of the energy that's contained within the nucleus. And it turned out that this is quite difficult to do. You need special materials. You need plutonium or highly enriched uranium. So really only states can do this kind of stuff to produce nuclear weapons. But what if it had turned out that there had been an easier way to release the energy of the atom? What if you could have made a nuclear bomb by baking sand in the microwave oven or something like that?
Starting point is 00:24:30 So then that might well have been the end of human civilization in that it's hard to see how you could have cities, let us say, if anybody who wanted to could destroy millions of people. So maybe we were just lucky. Now we know, of course, that it is physically impossible to create an atomic detonation by baking sand in the microwave, but before you actually did the relevant nuclear physics, how could you possibly have known how it would turn out? Let's just spell out that, because I want to conserve everyone's intuitions as we go on this harrowing ride to your terminus here, because the punchline of this paper is fairly startling when you get to what the remedies are.
Starting point is 00:25:13 So why is it that civilization could not endure the prospect of what you call easy nukes? the prospect of what you call easy nukes. If it were that easy to create a Hiroshima-level blast or beyond, why is it just a foregone conclusion that that would mean the end of cities and perhaps the end of most things we recognize? I think foregone conclusion is maybe a little too strong. It depends a little bit on the exact parameters we plug in. And the intuition is that in a large enough population of people, like amongst every population with millions of people, there will always be a few people who, for whatever reason, would like to kill a million people or more if they could, whether they are just crazy or evil, or they have some
Starting point is 00:26:06 weird ideological doctrine, or they're trying to extort other people or threaten other people. Just humans are very diverse, and in a large enough set of people, for practically any desire, you can specify there will be somebody in there that has that. So if each of those destructively inclined people would be able to cause a sufficient amount of destruction, then everything would get destroyed. Now, if one imagines this actually playing out in history, then to tell whether all of civilization really would get destroyed or some horrible catastrophe short of that would happen instead, would depend on various things. Like just what kind of nuclear weapon, would it be like a small kind of Hiroshima type
Starting point is 00:26:51 of thing or a thermonuclear bomb? How easy would it be? Could literally anybody do it in five minutes? Or would it take some engineer working for half a year? And so depending on exactly what values you pick for those and some other variables, you might get scenarios ranging from very bad to kind of existential catastrophe. But the point is just to illustrate that there historically have been these technological transitions where we have been lucky in that destructive capability we discovered were hard to to wield you know and maybe a plausible way in which this kind of very highly destructive
Starting point is 00:27:33 capability could become easy to wield in the future would be through developments in biotechnology that maybe makes it easy to create designer viruses and so forth right that doesn't don't require high amounts of energy or special difficult materials and so forth and there you might have an even stronger case like so with a nuclear weapon like one nuclear weapon can only destroy one city right where the viruses and stuff potentially can spread so yeah and we should remind people that we're in an environment now where people talk with some degree of flippancy about the prospect of every household one day having something like a desktop printer that can print DNA sequences, right? That everyone becomes their own bespoke molecular biologist, and you can just print your
Starting point is 00:28:26 own medicine at home, or your own genetic intervention at home, and this stuff really is, you know, the recipe under those conditions, the recipe to weaponize the 1918 flu could just be sent to you like a PDF. It's not beyond the bounds of plausible sci-fi that we could be in a condition where it really would be within the power of one nihilistic or otherwise ideological person to destroy the lives of millions and even billions in the wrong case. Yeah, or send us a PDF or you could just download it from the internet. So the full genomes of the number of highly virulent organisms are in the public domain and just ready to download. So yeah, so I mean, we could talk more about that. I think that I would rather see a future where DNA synthesis was a service provided by a few places in the world where it would be able,
Starting point is 00:29:22 if the need arose, to exert some control, some screening, rather than something that every lab needs to have its own separate little machine. Yeah, so these are examples of type one vulnerability, where the problem really arises from individuals becoming too empowered in their ability to create massive amounts of harm. Now, so there are other ways in which the world could be vulnerable that are slightly more subtle, but I think also worth bearing in mind. So these have to do more about the way that technological developments
Starting point is 00:29:58 could change the incentives that different actors face. We can again return to the nuclear history case for an illustration of this. And actually, this is maybe the closest to a black ball we've gotten so far with thermonuclear weapons and the big arms race during the Cold War led to something like 70,000 warheads being on here, trigger alert. So it looks like when we can see some of the archives of this history that have recently opened up, that there were a number of close calls. The world actually came quite close to the brink on several occasions, and we might have been quite lucky to get through.
Starting point is 00:30:38 It might not have been that we were in such a stable situation. It rather might have been that this was a kind of slightly blackballish technology and we just had enough luck to get through. But you could imagine it could have been worse. You could imagine properties of this technology that would have created stronger incentives, say for a first strike, so that you would have crisis instability. If it had been easier, let us say in a first strike to take out all the adversary's nuclear weapons, then it might not have taken a lot in a crisis situation to just have enough fear that you would have to strike first for fear that the adversary otherwise would
Starting point is 00:31:19 do the same to you. Yeah. Remind people that in the aftermath of the Cuban Missile Crisis, Yeah. Remind people that in the aftermath of the Cuban Missile Crisis, the people who were closest to the action felt that the odds of an exchange had been something like a coin toss, something like 30 to 50 percent. And what you're envisioning is a situation where, which you describe as safe first strike, which is, you know, there's just no reasonable fear that you're not going to be able to annihilate your enemy, provided you strike first. That would be a far less stable situation. And it's also forgotten that the status quo of mutually assured destruction was actually a step towards stability. Before the Russians had, or the Soviets had, their own arsenals, there was a greater game theoretic concern that we would be more tempted to use ours because nuclear deterrence wasn't a thing yet.
Starting point is 00:32:14 Yeah, so some degree of stabilizing influence, although of course maybe at the expense of the outcome, maybe more so. If you'd like to continue listening to this conversation, you'll need to subscribe at You're welcome. The Making Sense podcast is ad-free and relies entirely on listener support. And you can subscribe now at SamHarris.org.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.