Making Sense with Sam Harris - #151 — Will We Destroy the Future?
Episode Date: March 18, 2019Sam Harris speaks with Nick Bostrom about the problem of existential risk. They discuss public goods, moral illusions, the asymmetry between happiness and suffering, utilitarianism, "the vulnerable wo...rld hypothesis," the history of nuclear deterrence, the possible need for "turnkey totalitarianism," whether we're living in a computer simulation, the Doomsday Argument, the implications of extraterrestrial life, and other topics. If the Making Sense podcast logo in your player is BLACK, you can SUBSCRIBE to gain access to all full-length episodes at samharris.org/subscribe.
 Transcript
 Discussion  (0)
    
                                         Thank you. of the Making Sense Podcast, you'll need to subscribe at SamHarris.org. There you'll find our private RSS feed to add to your favorite podcatcher, along with other subscriber-only
                                         
                                         content. We don't run ads on the podcast, and therefore it's made possible entirely
                                         
                                         through the support of our subscribers. So if you enjoy what we're doing here,
                                         
                                         please consider becoming one.
                                         
                                         Welcome to the Making Sense Podcast. This is Sam Harris.
                                         
                                         Today I am speaking with Nick Bostrom.
                                         
                                         Nick is someone I've been hoping to get on the podcast for quite some time.
                                         
                                         He is a Swedish-born philosopher with a background in theoretical physics and computational neuroscience and logic and AI and many other interesting intersecting topics. But officially, he's a professor of
                                         
    
                                         philosophy at Oxford University, where he leads the Future of Humanity Institute. And this
                                         
                                         organization is a research center which is focused largely
                                         
                                         on the problem of existential risk. And today we get deep into his views on existential risk
                                         
                                         by focusing on three of his papers, which I'll describe in our conversation. We talk about what
                                         
                                         he calls the vulnerable world hypothesis, which gets us into many
                                         
                                         interesting tangents with respect to the history of nuclear deterrence and the possible need for
                                         
                                         what he calls turnkey totalitarianism. We talk about whether we're living in a computer simulation.
                                         
                                         He's the father of the now famous simulation argument. We talk about the doomsday
                                         
    
                                         argument, which is not his, but it's one of these philosophical thought experiments that
                                         
                                         have convinced many people that we might be living close to the end of human history.
                                         
                                         We talk about the implications of there being extraterrestrial life out there in the galaxy,
                                         
                                         and many other topics. But all of it is focused on the question of whether
                                         
                                         humanity is close to the end of its career here, or near the very beginning. And I hope you'll
                                         
                                         agree that the difference between those two scenarios is one of the more significant ones
                                         
                                         we can find. Anyway, I really enjoyed talking to Nick.
                                         
                                         I find his work fascinating and very consequential, and that's a good combination.
                                         
    
                                         And now I bring you Nick Bostrom.
                                         
                                         I am here with Nick Bostrom. Nick, thanks for coming on the podcast.
                                         
                                         Yeah, thanks for having me.
                                         
                                         So you are fast becoming, or not too fast, it's been years now that I've been aware of your work,
                                         
                                         but you are becoming one of the most provocative philosophers I can think of. And really,
                                         
                                         it's wonderful to read your work. And I want to introduce you, but perhaps to begin with, how do you view your work and
                                         
                                         what you focus on at this point?
                                         
                                         How do you summarize your interests as a philosopher?
                                         
    
                                         Well, that's always been a challenge for me.
                                         
                                         Broadly speaking, I'm interested in big picture questions for humanity and figuring out which
                                         
                                         direction is up and which is down.
                                         
                                         That is, out of all the things you can be pushing on or
                                         
                                         pulling on in the world, which ones would actually tend to make things better in expectation? Yeah,
                                         
                                         and then various kind of sub-questions that come out from that ultimate quest to figure out which
                                         
                                         direction we should be heading in. Yeah, so when I think about your work, I see a concern
                                         
                                         that unifies much of it, certainly, with existential risk. And I don't know if this
                                         
    
                                         is a phrase that you have popularized or if it's just derivative of your work, but how do you think
                                         
                                         of existential risk and why is it so hard for most people to care about? It's amazing to me
                                         
                                         that this is such an esoteric concern and you really have brought it into prominence.
                                         
                                         Yeah, I introduced the concept in a paper I wrote back in the early 2000s.
                                         
                                         The concept being that of a risk either to the survival of Earth originating in talent
                                         
                                         life or a risk that could permanently and drastically reduce our potential for desirable future developments.
                                         
                                         So in other words, something that could permanently destroy the future.
                                         
                                         I mean, even that phrase, I mean, you really have a talent for coming up with phrases
                                         
    
                                         that are arresting and, you know, it's such a simple one. It permanently destroy the future.
                                         
                                         You know, there are probably more people working in my local McDonald's than are thinking about the prospect of permanently
                                         
                                         destroying the future. How long have you been focused on this particular problem? And again,
                                         
                                         why is it, there's something bewildering about trying to export this concern to the rest of culture,
                                         
                                         even to very, very smart people
                                         
                                         who claim to be worried about things like climate change.
                                         
                                         Why is existential risk still such an esoteric concern?
                                         
                                         Well, it's become less so over the last few years.
                                         
    
                                         There is now actually a community of folk
                                         
                                         around the rationalist community, the EA community,
                                         
                                         you know, various academic centers. is an effective altruism yeah and about not not just these but but kind of radiating out from
                                         
                                         these a number of individuals that that are quite interested in this so i think the comparison to
                                         
                                         the mcdonald's restaurant would no longer be true now maybe it was true several mcdonald's years ago
                                         
                                         why it is well i guess you could ask that or you could ask why it's no longer the case. I mean,
                                         
                                         I don't know that the default should be, if we're looking at academia, that questions receive
                                         
                                         attention in proportion to their importance. I think that's just kind of a poor model of what
                                         
    
                                         to expect from academic research. So one can ask why it's changed.
                                         
                                         I mean, on one level, you're asking people to
                                         
                                         care about the unborn if the time horizon is beyond their lives and the lives of their children,
                                         
                                         which seems on its face probably harder than caring about distant strangers who are currently
                                         
                                         alive. And we know we're already pretty bad at that. Is that a major variable here?
                                         
                                         Sure. It's an extreme case of a public good, right? So generally, in a simple model of the
                                         
                                         market economy, public goods tend to be undersupplied because the creator of them
                                         
                                         captures only a small fraction of the benefits. The global public goods are normally seen as the extreme of this. If all of humanity
                                         
    
                                         benefits from some activity or is harmed by some activity, as in maybe the case of global warming
                                         
                                         or something like that, then the incentives facing the individual producer are just very
                                         
                                         dissociated from their overall consequences. But with existential risk, it's even more extreme,
                                         
                                         actually, because it's a transgenerational good in the sense that all future generations are also impacted by our decisions concerning what we do about existential
                                         
                                         risks. And they are obviously not in a position in any direct way to influence our decisions.
                                         
                                         They can't reward us if we do things that are good for them. So if one thinks of human beings
                                         
                                         as selfish, one would expect the good of existential risk reduction
                                         
                                         to be undersupplied.
                                         
    
                                         Like it could imagine if somehow people could go back in time
                                         
                                         that future generations would be willing to like
                                         
                                         spend huge amounts of money to compensate us
                                         
                                         for our efforts to reduce X risk.
                                         
                                         But since you can't, that transaction is not possible,
                                         
                                         then there is this undersupply.
                                         
                                         So that could be one way of explaining why there's relatively little.
                                         
                                         And there was something about what you said about it's harder to care.
                                         
    
                                         It's a little strange that caring should be something that requires effort.
                                         
                                         If one does care about it, it doesn't seem like it should be a straining thing
                                         
                                         to do. And if one doesn't care, then it's not clear why one should be, what motive one would
                                         
                                         have for trying to strain to start caring. It's a framing problem in many cases, so that there's
                                         
                                         a certain set of facts, let's say the reality of human suffering in some distant country that you
                                         
                                         have never visited, you have no connections
                                         
                                         there, this information can just be transmitted to you about the reality of the suffering.
                                         
                                         And it transmitted one way, you find that you don't care, and transmitted another way,
                                         
    
                                         the reality of it and the analogy to your own life or the lives of your children can be made
                                         
                                         more salient. And so we know that, we know, you know, through the lives of your children can be made more salient.
                                         
                                         And so we know that we know, you know, through the work of someone like Paul Slovic,
                                         
                                         we know there are moral illusions here where people can be shown to care more about the fate
                                         
                                         of one little girl who's delivered to them in the form of a personalized story. And they'll care
                                         
                                         less about the fate of that same little girl plus her brother, and they'll care less still if you tell them about the little girl, her brother,
                                         
                                         and the 500,000 other children who are also suffering from a famine.
                                         
                                         And you just get this diminishment of altruistic impulse
                                         
    
                                         and the amount of money they're willing to give to charity and all the rest,
                                         
                                         and it goes in the wrong direction. As you scale the problem, people care less. So we know we have some moral bugs
                                         
                                         in our psychology. Yeah. So the original paper about existential risk made the point that from
                                         
                                         a certain type of ethical theory, it looks like existential risk reduction is a very important goal.
                                         
                                         If you have a broadly aggregative consequentialist philosophy, say if you're a utilitarian,
                                         
                                         and if you work the numbers out, the number of possible future generations,
                                         
                                         the number of individuals in each of those that can live very happy lives,
                                         
                                         then you multiply that together. And then it looks like even a very small chance in the probability that we will eventually achieve this would have a very high expected value.
                                         
    
                                         In fact, a higher expected value than any other impacts that we might have in more direct ways on the world here and now.
                                         
                                         So that reducing existential risk by, you know, one thousandth of one percentage point would be, from this utilitarian perspective, worth more than eliminating world hunger or curing cancer.
                                         
                                         Now, that, of course, says nothing about the question of whether this kind of utilitarian
                                         
                                         perspective is correct or is agreeable to us. But it just notes that that does seem to be an
                                         
                                         implication. I'm definitely a consequentialist of a certain kind,
                                         
                                         so we don't need to argue that point. But one thing that's interesting here, and this may be
                                         
                                         playing into it, is that there seems to be a clear asymmetry between how we value suffering
                                         
                                         and its mitigation, and how we value the mere preemption of well-being or flourishing or positive states.
                                         
    
                                         So that suffering is worse than pleasure or happiness is good. You know, I think if you
                                         
                                         told most people, here are two scenarios for how you can spend the rest of your day. You can spend
                                         
                                         it as you were planning to, living within the normal bounds of human experience,
                                         
                                         or I can give you one hour of the worst possible misery
                                         
                                         followed by another hour of the deepest possible happiness.
                                         
                                         Would you like to sample the two extremes
                                         
                                         on the phenomenological continuum?
                                         
                                         I think most people would balk at this
                                         
    
                                         because we think, we have a sense that suffering is, on some level, you know, in the limit, is worse than any pleasure or
                                         
                                         happiness could be. And so we look at the prospect of, let's say, you know, curing cancer and
                                         
                                         mitigating the suffering from that as being more important ethically than simply not allowing the door to close on future states of creativity
                                         
                                         and insight and beauty. Yeah, I think one might want to decompose different ways in which that
                                         
                                         intuition could be produced. So one might just be that for us humans, as we are currently
                                         
                                         constituted, it's a lot easier to create pain than to create a corresponding degree of pleasure.
                                         
                                         We might just evolutionarily be such
                                         
                                         that we have a kind of deeper bottom
                                         
    
                                         than we have a height.
                                         
                                         It might be possible,
                                         
                                         if you think about it in biological term,
                                         
                                         in a short period of time
                                         
                                         to introduce more damage
                                         
                                         to damage reproductive fitness
                                         
                                         more than you could possibly gain
                                         
                                         within the same amount of time.
                                         
    
                                         So if we are thinking about these vast futures,
                                         
                                         you want to probably factor that out
                                         
                                         in that you could re-engineer, say, human hedonic systems
                                         
                                         or the hedonic systems of whatever inhabitants
                                         
                                         would exist in this future
                                         
                                         so that they would have a much larger capacity for upside.
                                         
                                         Right. And it's not obvious that there would have a much larger capacity for upside. Right.
                                         
                                         And it's not obvious that there would be an asymmetry there.
                                         
    
                                         Now, you might nevertheless think that even in some sense, equal amounts of pleasure and
                                         
                                         pain, and it's a little unclear exactly what the metric is here, that there would nevertheless
                                         
                                         be some more basically ethical reason why one should place a higher priority on removing
                                         
                                         the negative. A lot of
                                         
                                         people have intuitions about equality, say, in economic context, where helping the worst off
                                         
                                         is more important than further promoting the welfare of the best off. Maybe that's the source
                                         
                                         of some of those intuitions. Actually, there's one other variable here, I think, which is that there is no victim
                                         
                                         or, you know, beneficiary of the consequence of closing the door to the future. So if you ask
                                         
    
                                         someone, well, what would be wrong with the prospect of everyone dying painlessly in their
                                         
                                         sleep tonight, and there are no future generations, there's no one
                                         
                                         to be bereaved by that outcome.
                                         
                                         There's no one suffering the pain of the loss or the pain of the deaths even.
                                         
                                         So people are kind of at a loss for the place where the moral injury would land.
                                         
                                         Yeah, so that is a distinction within utilitarian frameworks, between total utilitarians who
                                         
                                         think you basically count up all the good and subtract all the bad.
                                         
                                         And then other views that try to take a more so-called person-affecting perspective, where
                                         
    
                                         what matters is what kind of happens to people, but coming into existence is not necessarily a benefit.
                                         
                                         And now I would say some kinds of existential catastrophe would have a continuing population
                                         
                                         of people who would be experiencing the bad.
                                         
                                         If you might say the world getting locked into some totalitarian, like really dystopian
                                         
                                         totalitarian regime, maybe there would be people living for a
                                         
                                         very long time, but just having much less
                                         
                                         good lives than could
                                         
                                         have existed.
                                         
    
                                         So in some scenarios of existential
                                         
                                         catastrophe, there would
                                         
                                         still be inhabitants there.
                                         
                                         Yeah, no, I think
                                         
                                         it's pretty clear that destroying the future could be
                                         
                                         pretty unpleasant for people who
                                         
                                         are along for the ride.
                                         
                                         Now, I'd like just to harken back a few minutes ago, like on the general premise here. So I don't see it so much as a premise, this utilitarian view. I mean, in fact,
                                         
    
                                         I wouldn't really describe myself as a utilitarian, but more just pointing out the
                                         
                                         consequence. There are various views about how we should reason about ethics, and there might be
                                         
                                         other things we care about as well, aside from ethics. And rather than directly trying to answer
                                         
                                         what do we have most reason to do all things considered, you might break it down and say,
                                         
                                         well, given this particular ethical theory, what do we have most reason to do given this other
                                         
                                         value or this other goal we might have? And then at the end of the day, you might want to add all of that up again. But insofar as we are trying to reason about our ethical obligations, I have a kind of normative uncertainty over different
                                         
                                         moral frameworks. And so the way I would try to go about making decisions from a moral point of
                                         
                                         view would be to think, I have this moral parliament model. It's a kind of metaphor,
                                         
    
                                         but where you try to factor in the viewpoints of a number of different ethical theories,
                                         
                                         kind of in proportion to the degree to which you assign them probability.
                                         
                                         It's kind of interesting. When I'm out and about in the world, I usually have to make the case for
                                         
                                         utilitarianism, or at least you should consider this perspective. People are scope insensitive.
                                         
                                         the case for utilitarianism, or at least you should consider this perspective. Like people are scope insensitive.
                                         
                                         You should look at the numbers.
                                         
                                         If this thing has millions of people and this one only has hundreds of people being affected,
                                         
                                         clearly that, and yet when I'm back here at the headquarters, as it were, I usually am
                                         
    
                                         the one who has to kind of advocate against utilitarian perspective because so many of
                                         
                                         my friends are so deeply dyed in the wool
                                         
                                         utilitarians yeah and so narrowly focused on x-risk mitigation but i feel that i'm always the
                                         
                                         the odd one out well you know i would love to get into a conversation with you about meta-ethics
                                         
                                         some other time because i you know i i think your views about the limits of consequentialism
                                         
                                         would be fascinating to explore. But I have so
                                         
                                         much I want to talk to you about with respect to X-Risk and a few of your papers that I think,
                                         
                                         let's just table that for another time. In fact, I don't even think we're going to be able to cover
                                         
    
                                         your book Superintelligence. I mean, maybe if we have a little time at the end, we'll touch it. But
                                         
                                         I should just want to say that this book was incredibly influential on many of us in arguing the case
                                         
                                         for there being a potential existential risk with respect to the development of artificial
                                         
                                         intelligence and artificial general intelligence in particular.
                                         
                                         And so it's, you know, this is something, the reason why I wouldn't cover this with
                                         
                                         you for the entirety of this conversation is I've had several conversations on my podcast that
                                         
                                         have been deeply informed by your view.
                                         
                                         I mean, I've had Stuart Russell on, I've had Eliezer Yudkowsky on, and basically every
                                         
    
                                         time I talk about AI, I consider what I say to be, you know, fairly derivative of your
                                         
                                         book, and I often remind people of that. So my audience
                                         
                                         will be familiar with your views on AI, even if they're not familiar with you. So if we have time,
                                         
                                         we'll get back to that. But what I really want to talk about are a few of your papers. The first is
                                         
                                         the vulnerable world hypothesis. Maybe I'll just name the papers here that I hope will cover. Vulnerable world, the second is, are you living in a computer simulation?
                                         
                                         And the third is, where are they?
                                         
                                         Which is your analysis of the Fermi problem, asking where is the rest of the intelligent
                                         
                                         life in the galaxy?
                                         
    
                                         Let's start with the vulnerable world hypothesis.
                                         
                                         What do you mean by that phrase?
                                         
                                         Well, the hypothesis is, roughly speaking, that there is some level of technological development
                                         
                                         at which the world gets destroyed by default, as it were. So then what does it mean to get
                                         
                                         destroyed by default? I define something I call the semi-anarchic default condition,
                                         
                                         which is a condition in
                                         
                                         which there are a wide range of different actors with a wide range of different human
                                         
                                         recognizable motives.
                                         
    
                                         But then, more importantly, two conditions hold.
                                         
                                         One is that there is no very reliable way of resolving global coordination problems.
                                         
                                         And the other is that we don't have a very extremely reliable way of preventing
                                         
                                         individuals from committing actions that are extremely strongly disapproved of by a great
                                         
                                         majority of other people. Maybe it's better to come at it through a metaphor.
                                         
                                         Yeah, the urn.
                                         
                                         The urn metaphor. So you can kind of think of the history of
                                         
                                         So you can kind of think of the history of technological discovery as the process of pulling balls out of a giant urn,
                                         
    
                                         the urn of creativity.
                                         
                                         And we reach in and we get an idea out,
                                         
                                         and then we reach in and get another out.
                                         
                                         And we've extracted throughout history a great many of these balls.
                                         
                                         And the net effect of this has been hugely beneficial, I would say.
                                         
                                         This is why we now sit in our air-conditioned offices and struggle not to eat too much,
                                         
                                         rather than to try to get enough to eat in large parts of the world. But what if in this ball,
                                         
                                         in this area, there is a black ball in there somewhere?
                                         
    
                                         Is there some possible technology that could be such that whichever civilization discovers it, invariably gets destroyed?
                                         
                                         Just to add a little color here, Nick.
                                         
                                         So in your paper, you refer to this as the urn of inventions.
                                         
                                         And we have been, as you say, pulling balls out as quickly as we can get our hands on them.
                                         
                                         And on some level, the scientific ethos is really just a matter of pulling balls out as fast as you
                                         
                                         can and making sure that everybody knows about them. We have this norm of transparency in science,
                                         
                                         and we have pulled out, thus far, only white or gray balls, and the white balls are the ones,
                                         
                                         or the technologies, or the memes, or the norms, or the social institutions that just have good
                                         
    
                                         consequences, and the gray ones are norms and memes and institutions and, in most cases,
                                         
                                         technology that has mixed results or that can be used for good or for ill. And nuclear energy is
                                         
                                         a classic case where we can power our cities with it, but we also produce fantastic amounts of
                                         
                                         pollution that's difficult to deal with. And in the worst case, we build weapons. So I just want
                                         
                                         to give a little more context to this analogy. Yeah, and I guess most technologies are in some sense double-edged,
                                         
                                         but maybe the positive predominate.
                                         
                                         I think there might be some technologies that are mainly negative
                                         
                                         if you think of, I don't know, nerve gases or other tools.
                                         
    
                                         But what we haven't so far done is extract a black ball,
                                         
                                         one that is so harmful that it destroys the civilization that discovers it and um what if
                                         
                                         there is such a black ball in the urn though i mean we can ask about how likely that is to be
                                         
                                         the case we can also look at what what is our current strategy with respect to this possibility
                                         
                                         and it seems to me that currently our strategy with respect to the possibility that the urn
                                         
                                         might contain a black
                                         
                                         ball is simply to hope that it doesn't. So we keep extracting balls as fast as we can. We have become
                                         
                                         quite good at that, but we have no ability to put balls back into the urn. We cannot uninvent our
                                         
    
                                         inventions. So the first part of this paper tries to identify what are the types of ways in which the world
                                         
                                         could be vulnerable, the types of ways in which there could be some possible blackball
                                         
                                         technology that we might invent.
                                         
                                         And the first and most obvious type of way the world could be vulnerable is if there
                                         
                                         is some technology that greatly empowers individuals to cause sufficiently large quantities of
                                         
                                         destruction.
                                         
                                         Motivate this with a, or illustrate it by means of a historical counterfactual.
                                         
                                         We, in the last century, discovered how to split the atom and release the energy that
                                         
    
                                         is contained within some of the energy that's contained within the nucleus.
                                         
                                         And it turned out that this is quite difficult to do.
                                         
                                         You need special materials.
                                         
                                         You need plutonium or highly enriched uranium.
                                         
                                         So really only states can do this kind of stuff to produce nuclear weapons.
                                         
                                         But what if it had turned out that there had been an easier way to release the energy of
                                         
                                         the atom?
                                         
                                         What if you could have made a nuclear bomb by baking sand in the microwave oven or something like that?
                                         
    
                                         So then that might well have been the end of human civilization in that
                                         
                                         it's hard to see how you could have cities, let us say, if anybody who wanted to could
                                         
                                         destroy millions of people. So maybe we were just lucky. Now we know, of course,
                                         
                                         that it is physically impossible to create an atomic detonation by baking sand in the microwave,
                                         
                                         but before you actually did the relevant nuclear physics, how could you possibly have known how
                                         
                                         it would turn out? Let's just spell out that, because I want to conserve everyone's intuitions
                                         
                                         as we go on this harrowing ride to your terminus here,
                                         
                                         because the punchline of this paper is fairly startling when you get to what the remedies are.
                                         
    
                                         So why is it that civilization could not endure the prospect of what you call easy nukes?
                                         
                                         the prospect of what you call easy nukes. If it were that easy to create a Hiroshima-level blast or beyond, why is it just a foregone conclusion that that would mean the end of
                                         
                                         cities and perhaps the end of most things we recognize?
                                         
                                         I think foregone conclusion is maybe a little too strong. It depends a little bit on the
                                         
                                         exact parameters
                                         
                                         we plug in. And the intuition is that in a large enough population of people, like amongst every
                                         
                                         population with millions of people, there will always be a few people who, for whatever reason,
                                         
                                         would like to kill a million people or more if they could, whether they are just crazy or evil, or they have some
                                         
    
                                         weird ideological doctrine, or they're trying to extort other people or threaten other people.
                                         
                                         Just humans are very diverse, and in a large enough set of people, for practically any
                                         
                                         desire, you can specify there will be somebody in there that has that.
                                         
                                         So if each of those destructively inclined people would be able to cause a sufficient amount of destruction, then
                                         
                                         everything would get destroyed. Now, if one imagines this actually playing out in history,
                                         
                                         then to tell whether all of civilization really would get destroyed or some horrible catastrophe
                                         
                                         short of that would happen instead, would depend on various things.
                                         
                                         Like just what kind of nuclear weapon, would it be like a small kind of Hiroshima type
                                         
    
                                         of thing or a thermonuclear bomb?
                                         
                                         How easy would it be?
                                         
                                         Could literally anybody do it in five minutes?
                                         
                                         Or would it take some engineer working for half a year?
                                         
                                         And so depending on exactly what values you pick for those and some other variables, you might
                                         
                                         get scenarios ranging from very bad to kind of existential catastrophe. But the point is just
                                         
                                         to illustrate that there historically have been these technological transitions where we have been lucky in that destructive capability we discovered were
                                         
                                         hard to to wield you know and maybe a plausible way in which this kind of very highly destructive
                                         
    
                                         capability could become easy to wield in the future would be through developments in
                                         
                                         biotechnology that maybe makes it easy to create designer viruses and so forth right that doesn't don't require high
                                         
                                         amounts of energy or special difficult materials and so forth and there you might have an even
                                         
                                         stronger case like so with a nuclear weapon like one nuclear weapon can only destroy one city right
                                         
                                         where the viruses and stuff potentially can spread so yeah and we should remind people that we're in an environment now where people
                                         
                                         talk with some degree of flippancy about the prospect of every household one day having
                                         
                                         something like a desktop printer that can print DNA sequences, right? That everyone becomes their
                                         
                                         own bespoke molecular biologist, and you can just print your
                                         
    
                                         own medicine at home, or your own genetic intervention at home, and this stuff really is,
                                         
                                         you know, the recipe under those conditions, the recipe to weaponize the 1918 flu could just be
                                         
                                         sent to you like a PDF. It's not beyond the bounds of plausible sci-fi that we could be in a
                                         
                                         condition where it really would be within the power of one nihilistic or otherwise ideological
                                         
                                         person to destroy the lives of millions and even billions in the wrong case.
                                         
                                         Yeah, or send us a PDF or you could just download it from the internet. So the full genomes of the number of highly virulent organisms are in the public domain and just ready to download.
                                         
                                         So yeah, so I mean, we could talk more about that. I think that I would rather see a future
                                         
                                         where DNA synthesis was a service provided by a few places in the world where it would be able,
                                         
    
                                         if the need arose, to exert some control,
                                         
                                         some screening, rather than something that every lab needs to have its own separate little
                                         
                                         machine.
                                         
                                         Yeah, so these are examples of type one vulnerability, where the problem really arises from
                                         
                                         individuals becoming too empowered in their ability to create massive amounts of harm.
                                         
                                         Now, so there are other ways in which the world could be vulnerable
                                         
                                         that are slightly more subtle, but I think also worth bearing in mind.
                                         
                                         So these have to do more about the way that technological developments
                                         
    
                                         could change the incentives that different actors face.
                                         
                                         We can again return to the nuclear history case for an illustration of this.
                                         
                                         And actually, this is maybe the closest to a black ball we've gotten so far with thermonuclear
                                         
                                         weapons and the big arms race during the Cold War led to something like 70,000 warheads being on
                                         
                                         here, trigger alert. So it looks like when we can see some of the archives of this history that have recently
                                         
                                         opened up, that there were a number of close calls.
                                         
                                         The world actually came quite close to the brink on several occasions, and we might have
                                         
                                         been quite lucky to get through.
                                         
    
                                         It might not have been that we were in such a stable situation.
                                         
                                         It rather might have been that this was a kind of slightly blackballish
                                         
                                         technology and we just had enough luck to get through. But you could imagine it could have
                                         
                                         been worse. You could imagine properties of this technology that would have created stronger
                                         
                                         incentives, say for a first strike, so that you would have crisis instability. If it had been
                                         
                                         easier, let us say in a first strike to take out all the
                                         
                                         adversary's nuclear weapons, then it might not have taken a lot in a crisis situation to just
                                         
                                         have enough fear that you would have to strike first for fear that the adversary otherwise would
                                         
    
                                         do the same to you. Yeah. Remind people that in the aftermath of the Cuban Missile Crisis,
                                         
                                         Yeah. Remind people that in the aftermath of the Cuban Missile Crisis, the people who were closest to the action felt that the odds of an exchange had been something like a coin toss, something like 30 to 50 percent.
                                         
                                         And what you're envisioning is a situation where, which you describe as safe first strike, which is, you know, there's just no reasonable fear that you're not going to be able to annihilate your enemy,
                                         
                                         provided you strike first.
                                         
                                         That would be a far less stable situation.
                                         
                                         And it's also forgotten that the status quo of mutually assured destruction was actually a step towards stability.
                                         
                                         Before the Russians had, or the Soviets had, their own arsenals, there was a greater game theoretic
                                         
                                         concern that we would be more tempted to use ours because nuclear deterrence wasn't a thing yet.
                                         
    
                                         Yeah, so some degree of stabilizing influence, although of course maybe at the expense of the
                                         
                                         outcome, maybe more so.
                                         
                                         If you'd like to continue listening to this conversation, you'll need to subscribe at You're welcome. The Making Sense podcast is ad-free and relies entirely on listener support. And you can subscribe now at SamHarris.org.
                                         
