Making Sense with Sam Harris - #292 — How Much Does the Future Matter?

Episode Date: August 14, 2022

Sam Harris speaks with William MacAskill about his new book, What We Owe the Future. They discuss the philosophy of effective altruism (EA), longtermism, existential risk, criticism of EA, problems w...ith expected-value reasoning, doing good vs feeling good, why it's hard to care about future people, how the future gives meaning to the present, why this moment in history is unusual, the pace of economic and technological growth, bad political incentives, value lock-in, the well-being of conscious creatures as the foundation of ethics, the risk of unaligned AI, how bad we are at predicting technological change, and other topics. If the Making Sense podcast logo in your player is BLACK, you can SUBSCRIBE to gain access to all full-length episodes at samharris.org/subscribe.   Learning how to train your mind is the single greatest investment you can make in life. That’s why Sam Harris created the Waking Up app. From rational mindfulness practice to lessons on some of life’s most important topics, join Sam as he demystifies the practice of meditation and explores the theory behind it.

Transcript
Discussion (0)
Starting point is 00:00:00 Welcome to the Making Sense Podcast. This is Sam Harris. Well, it's been a big week for news. I think I'm going to wait for a few more shoes to drop before touching any of it. Earlier in the week, Trump's home at Mar-a-Lago. Is that really his home? home at Mar-a-Lago. Is that really his home? That golf club that looks like a dictator's palace in what he might call a shithole country? It was raided by, quote, raided by the FBI in search of misappropriated documents that may or may not contain secrets about nuclear weapons or nuclear policy or something nuclear. Again, I think I'll wait for more information to surface before I comment on this, but needless to say, it would not surprise me if Trump had done something terrifyingly inept and self-serving and even nefarious. So let's just
Starting point is 00:01:28 wait and see what happens there. I share many people's concern that the execution of this search warrant was inflammatory enough that if the facts in the end don't seem to merit it, it will be yet another thing that bolsters Trump's political stature in the personality cult that has subsumed the Republican Party, and that would be a bad thing. Once again, this is an asymmetrical war of ideas, and any misstep that seems to suggest that our institutions have become politicized, in this case the FBI, run, we should note, by a Trump appointee, Christopher Wray, not the fittest object for allegations of anti-Trump partisanship, really, really. But any apparent bias here makes any further discussion of facts more or less impossible.
Starting point is 00:02:37 So anyway, let's wait for the dust to settle. Let's see what was actually in those boxes, and maybe there'll be more to say. And then yesterday, the author Salman Rushdie was attacked on stage by a man wielding a knife under conditions where he obviously did not have enough security. And it sounds like he is on life support now. So that obviously is terrible. And I share everyone's horror on his account. I said as much on Twitter, not so many hours ago, where I wrote, like many of you I'm thinking about Salman Rushdie now. The threat he's lived under for so long, which was so horrifically realized today, was the product not merely of the hatred and zeal of religious fanatics, but of the cowardice and confusion of secularists. Everyone in arts and
Starting point is 00:03:30 letters should have stood shoulder to shoulder with Salman in 1989, thereby distributing the risk. And the fact that so few did is a moral scandal that still casts its shadow over the present. its shadow over the present. And now I now notice in my at-mentions that I am being attacked by many people for being soft on Islam and for using the phrase religious fanatics out of an apparent unwillingness to name Islam and its theology as the culpable party here. Islam and its theology as the culpable party here. All I can say is I'm surprised and somewhat amused by this. I mean, these are people who clearly don't know my history. I have been accused of Islamophobia even more than I've been accused of having Trump derangement syndrome. It might surprise some of you to learn. Anyway, my conscience on that score is quite clear, and I invite those of you who doubt my bona fides as a critic of Islam to do some googling.
Starting point is 00:04:33 I even wrote an op-ed with Salman in defense of our mutual friend Ayaan Hirsi Ali, I think in 2007. That's on my blog somewhere. I believe the title is Ayaan Hirsi Ali Abandoned to Fanatics. It was published in the Los Angeles Times, but I think it's behind a paywall there. Anyway, I don't actually know Salman, despite the fact that we wrote that op-ed together. That was accomplished by email.
Starting point is 00:05:03 We've never met in person, which surprises me. We've always had several good friends in common, Ayan being one, Hitch being another. Salman was one of Hitch's best friends. I was merely a friend. Anyway, thinking about him, I was merely a friend. Anyway, thinking about him, and I would just say, by way of spelling out further the point I was making on Twitter, my use of the phrase religious fanatics was not at all an instance of my mincing words
Starting point is 00:05:38 with respect to Islam, although I can see how someone might have thought that if they were unfamiliar with my work, I was just assuming enough familiarity with what I have said and written on this topic that it simply never occurred to me that I could be accused of pandering to Islamic extremists here or to the confused liberal secularists
Starting point is 00:06:01 who would support such pandering. I just assumed my reputation preceded me here. Anyway, the real point I was making here was that the problem has always been larger than mere Islamic extremism and its intolerance to free speech. This extremism has been given an immense amount of oxygen by confused and cowardly and, in some cases, well-intentioned people who worry that any criticism of Islam will give energy to racism and xenophobia. Salman was attacked by such people from the moment the Ayatollah pronounced his fatwa. And the list of people who rushed to heap shame upon their heads by criticizing a novelist forced into hiding out of deference for the sensitivities of religious lunatics,
Starting point is 00:07:06 is an awful list of names. And so it is with more recent atrocities, like the Charlie Hebdo massacre. The people who explicitly or implicitly took the side of the jihadist murderers in criticizing the butchered cartoonists for their alleged insensitivity, these are people who are still in good standing in many cases, and I'm not aware that any have apologized for having got those
Starting point is 00:07:32 moments of moral and political emergency so disastrously wrong. I'm doing my best not to name names here, but anyone interested can see who took the side of religious barbarians there. And there have been so many moments like that since the Fatwa on Salman was first pronounced. And many of you who are living by podcast and sub-stack newsletter might take a look, because some of your modern free speech heroes were people who thought that just maybe those cartoonists had it coming to them. Anyway, this is an awful thing. I hope Salman recovers, and whatever happens there, if there's more to say, I'll do a podcast on it. I'll do a podcast on it.
Starting point is 00:08:24 Okay. Today I'm speaking with Will McCaskill. Will is an associate professor in philosophy at the University of Oxford and a TED speaker, a past Forbes 30 Under 30 social entrepreneur. He also co-founded the Center for Effective Altruism, which has raised over a billion dollars for charities. And he is the author, most recently, of a wonderful book titled What We Owe the Future. And that is in part the topic of today's discussion. I'll just read you the blurb I wrote for that book, because it'll give you a sense of how
Starting point is 00:09:01 important I think Will's work is. No living philosopher has had a greater impact upon my ethics than Will McCaskill. In What We Owe the Future, McCaskill has transformed my thinking once again by patiently dismantling the lazy intuitions that rendered me morally blind to the interests of future generations. This is an altogether thrilling and necessary book. And that pretty much says it. The book is also blurbed by several friends,
Starting point is 00:09:31 Paul Bloom, Tim Urban, Stephen Fry. As you'll hear, Will has been getting a lot of press recently. There was a New Yorker profile that came out a couple of days ago. Also the cover of Time magazine was devoted to him and the effective altruism movement. And all of this is great to see. So Will and I get into the book. We talk about effective altruism in general and its emphasis on long-termism and existential risk. We talk about criticisms of the philosophy and some of the problems with expected value reasoning. We talk about the difference between doing good and feeling good, why it's so hard to care about
Starting point is 00:10:12 future people, how the future gives meaning to the present, why this moment in history is unusual, which relates to the pace of economic and technological growth. We discussed bad political incentives, value lock-in, the well-being of conscious creatures as a foundation for ethics, the risk of unaligned artificial intelligence, how we're bad at predicting technological change, and many other topics.
Starting point is 00:10:40 No paywall for this one, as I think it's so important. As always, if you want to support what I'm doing here, you can subscribe to the podcast at samharris.org. And now I bring you Will McCaskill. I am back once again with the great Will McCaskill. Will, thanks for joining me again. Thanks so much. It's great to be back on. So you are having a bit of a wild ride. You have a new book, and that book is What We Owe the Future. And I got a very early look at that, and it's really wonderful.
Starting point is 00:11:18 I don't think we'll cover all of it here, but we'll definitely get into it. don't think we'll cover all of it here, but we'll definitely get into it. But before we do that, I want to talk about effective altruism more generally, and in particular, talk about some of the pushback I've noticed of late against it. And some of this echoes reservations I've had with the movement, and some of which we've discussed, and some of it strikes me as, frankly, crazy. So yeah, let's just get into that. But before we do, how are you doing? How's life? I mean, things are very good overall. I definitely feel a little bit shell-shocked by the wave of attention I'm currently through. And, you know, that's, I don't know, I'm alternating between excitement and anxiety. But overall, it's very good indeed. And I'm happy these ideas are getting
Starting point is 00:12:11 more their time. Am I right in thinking that I just saw EA and a long interview with you on the cover of Time magazine? That's exactly right. So that was the news of today that came out this morning. And I think it's a beautiful piece by Naina Bajikal. And yeah, I'd encourage anyone to read it if they want to get a sense of what the EA movement is like from the inside. bend more of it here. There's a lot we've already said about this, but I think we have to assume that many people in the audience have not heard the many hours you and I have gone round and round on these issues already. So at the top here, can you define a few terms? I think you should define effective altruism, certainly what you mean or you think we should mean by that phrase, and also long-termism and existential risk, because those will be coming up a lot here. Sure. So effective altruism is a philosophy and a community that's about trying to ask the question,
Starting point is 00:13:19 how can we do as much good as possible with our time and with our money, and then taking action to put the results of that or answers to that question into practice to try to make the world better. Long-termism is a philosophical perspective that says at least one of the key priorities of our time is to make the long-term future of humanity go well. That means just taking seriously just the sheer scale of the future, how high the stakes could be in shaping it. The fact that there really might be events that occur in our lifetimes that could have an enormous impact for good or ill in our own lifetimes and taking seriously that we can help steer those events in a more positive direction,
Starting point is 00:14:04 helping set humanity on a better path to improve the lives of our grandkids and their grandkids and their grandkids and so on. Then the final term, existential risks, are a category of issues that I and others have increasingly focused on or started thinking and worrying about. And these are risks that could be enormously impactful, not just in the present, but in fact for the whole future. So some of these risks are risks that threaten extinction, such as the use of our next generation of bioweapons in World War III with viruses of unprecedented lethality and infectiousness that could really result in the collapse of civilization or the extinction of everyone alive today in the worst
Starting point is 00:14:52 case. Or very rapid advances in artificial intelligence where that could result in, as you've argued, the AIs being just far more capable than human beings, such that the AI systems merely see humans as an obstacle to their goals. And humans are disempowered in light of that, or in fact, again, just made extinct or eliminated by such systems. Or existential risks can lead to permanent loss of value or future value, not just via extinction. So a perpetual dystopia that was, you know, perhaps it's a little better than extinction, but not much better, with a lock-in of fascist values.
Starting point is 00:15:42 That would count as an existential catastrophe as well. And increasingly, I've been getting to know the leading scientists who think that the scale of such existential catastrophe, it's actually not small. It's not like, in any way like this, one in a million chance of an asteroid strike in the next century, but much more like the risk of dying in a car crash in one's lifetime, where these risks are certainly in the next century, but much more like the risk of dying in a car crash in one's lifetime, where these risks are certainly in the multiple percentage points, maybe even tens of percentage points, just in our lifetime alone. So in the definition of effective altruism, I think one could be forgiven for wondering what is novel about simply wanting to do as much good as one can,
Starting point is 00:16:29 right? I mean, how is that a new contribution to the way we think about doing good in the world? Hasn't everyone who has attempted to do good wanted to do as much good as they could get their hands around? What do you view as being novel in this movement around that? their hands around. What do you view as being novel in this movement around that? So I'm certainly not claiming that effective altruism is some radically new idea. I think you can see its intellectual roots going back decades. I actually argue you could see it going back two and a half thousand years to the first consequentialist model philosophers, the Mohists. the first consequentialist model philosophers, the Mohists. But I actually think that this mindset of trying to do as much good as possible is at least somewhat unusual, where in general, when people try to do good, and this was certainly true for me before I started thinking about this,
Starting point is 00:17:20 they are looking at what causes have been made salient to them, where in my case, that was extreme poverty. I remember as a teenager reading about how many people died of AIDS each year, just thinking, oh my God, this is an atrocity. How can we not be taking more action on this? But what I wasn't doing at the time was sitting back, kind of taking a step back and thinking of all the things, all the many, many problems in the world, what are the things, what's the thing that I should be most focusing on if I want to have the biggest impact I can? What's the best evidence? What are the best, you know, what are the best evidence saying about this? What are the best arguments saying about
Starting point is 00:17:59 this? And really thinking that through. And so I actually think that mode of reasoning is somewhat unusual, and that's what's distinctive about effective altruism. Yeah, so let's deal with some of this pushback before moving on. I'm sure there's been much more that I haven't noticed, but in the last few days I noticed a Wall Street Journal article that really did not have much substance to it. It seemed to have been written by someone who just doesn't like paying taxes and didn't seem to like Sam Bankman Freed's politics. I didn't really detect any deep arguments against effective altruism there. I guess his one salient point was that creating economic value, i.e. doing business, is generally the best way to help people
Starting point is 00:18:47 at scale and reduce unnecessary suffering. And I guess I think we could easily concede that insofar as that's true, and wherever it happens to be true, it would be very easy for us to say, well then, sure, let business solve our problems. And so there's, you know, and I don't think you would resist that conclusion wherever it's true. I mean, one could certainly argue that someone like Elon Musk, by, you know, building desirable electric cars, has done, if not more good, a different species of good than many forms of environmental activism have to move us away from a climate change catastrophe. So if that's true, let's bracket that, and it's hard to know what exactly is true there, but I certainly have an intuition that there's a place for business,
Starting point is 00:19:41 and even a majority place for business, affect our climate outcome and that it might be greater than charity. He seemed to think that was a real, you know, some kind of coup de gras against effective altruism and even philanthropy in general. But I think that blow just passes right by us, doesn't it? Yeah. So the thing that I'll say that's the grain of truth in that criticism is that in some circumstances, markets can be extremely good ways of allocating resources. So take the problem of people having access to the internet and to computing technology in the United States, at least for people who are middle class or above, then, you know, businesses are dealing with that, like, at least relatively well. There's great incentives to give this product that
Starting point is 00:20:38 are benefiting people, people are willing to pay for it, and there aren't kind of major externalities. However, as has been well known in economics for I think 100 years now, there are very many cases where markets don't produce the ideal social outcome, which they call externalities. So take the case of climate change, which is, you know, an absolute paradigm of this example, where if someone mines coal and burns it, that can be good for them because they're creating energy. But it has this negative effect on third parties that were not kind of privy to that decision. They had no influence over it because it warms the climate and it makes those third parties worse off. And so the standard solution there, and this is endorsed
Starting point is 00:21:22 by economists on the right as well as the left, it's just uncontroversial, is that you want some amount of government involvement there, at least in the form of taxation. So putting a tax on carbon. There's other things you can do as well, but that would be one way, in order to guide market behavior to better outcomes. However, I would go further than this as well and say, you know, there are not only cases of market failure, there are also cases of democratic failure as well, where if you consider the extreme poor who are living in a country with poor political institutions, perhaps even a dictatorship, they do not have democratic say over the policies that rich countries like the UK and the US are enacting. And so they will not be given adequate representation,
Starting point is 00:22:12 neither in the market, nor in political decision making. Or if we look at non-human animals, they have no say at all. Or if we look at future people, those who will be born not now, but in the generations to come, they have no say over, they have no participation in the market. So they have no influence there. They have no say in political decision-making either. And so it's no surprise, I think, that their interests are systematically neglected. And we have no guarantee at all that kind of business as usual for the world would lead to the best outcomes possible for them. Yeah, yeah. Okay, so then there was this article I noticed in this online journal, maybe it's also a print journal, called Current Affairs, which I hesitate to mention because
Starting point is 00:22:59 the editor over there strikes me as someone who is truly mentally unwell, given some of the insane things he's written about me and the length at which he's written them. But this article, not written by him, written by somebody else, raised some points that I think are valid, as well as many that weren't. And the valid ones are around the issue of how we think about expected value and, you know, kind of a probability calculus over various possibilities in the future. And this is really integral to your new book and let's talk about the concept of prioritizing, to whatever degree, the well-being of people who don't yet exist. When people say, well, we should prioritize the future to some significant degree, it's very easy for them to do this by reference to their children, degree, it's very easy for them to do this by reference to their children, you know, young people who, you know, currently exist and have no control over anything. That's easy to see. But when you're talking about the hypothetical billions and trillions of a future humanity
Starting point is 00:24:17 that may thrive or not even exist, depending on how we play our cards at this moment, I think people's moral intuitions begin to fall apart. And the most extreme case of the problem here is, if you try to run any kind of utility function based on a straightforward calculation of probabilities, if you say, well, if I could just reduce the risk that we will be wiped out by some catastrophic event, let's say a global pandemic, if I could reduce that by even one in a million, given the fact that you have trillions upon trillions of potential lives on that side of the balance, well, then that's more important than anything I could do now, including saving the lives of identifiable people, even thousands of identifiable people.
Starting point is 00:25:11 And then that's where people feel like something has simply gone wrong here in how we're thinking about human well-being. So perhaps you can address that. Sure. I'm so glad you brought it up because this is a flaming that often gets mentioned. And I think it's really unfortunate. And I want to push against it in the strongest possible terms, where if it were the case that the argument for concern and serious concern about how we might impact future generations rested on these vanishingly small probabilities of enormous amounts of value, then I wouldn't be writing this book. I would be doing other things.
Starting point is 00:25:50 And I don't think it does. And the reason this framing came up is because in academic articles, Nick Bostrom, who I believe you've had on, he sketched out what are the implications of a very particular moral view, uh, expectational total utilitarianism. And he said, well, if you're taking this view seriously, then that's the conclusion you get. What he wasn't saying is that we should endorse that conclusion. In fact, he, and essentially everyone I know, at least in practice outside of the ivory tower, take that as a reductio rather than an implication. So Nick himself is not a consequentialist even, and wrote other papers on this idea of Pascal's mugging. So it's similar like Pascal's wager,
Starting point is 00:26:39 but just without infinite stakes, where these tiny probabilities of huge amounts of value, look, it just, it doesn't seem like you just do the math and multiply them. Can you actually describe the Pascal's mugging example? Sure, I'm happy to. And so the idea is just, you know, you've had a nice time at the pub, you walk outside, a man comes up to you and says, give me your wallet. You say no. And he says, okay, fair. If you give me the contents of your wallet, let's say that's $100, I tomorrow will come back and I will give you any finite amount of well-being. So just whatever amount of well-being you ask for, I will give you it. And you say, no, I don't,
Starting point is 00:27:28 I just think I'm certain that that's not going to happen. And the man says, come on, you're not certain, are you? Look at my eyes. I've got this pale complexion. I look a bit, you know, it's possible that I'm this, you know, traveler from the future or this alien with like amazing powers, or I'm some sort of god it's possible maybe it's a one in a billion billion billion billion that I will be able to do this but it's at least possible and you take that one in a billion billion billion billion multiply it by a sufficiently large amount of well-being for you that in what decision theorists call expected value which is where you take the probability of an
Starting point is 00:28:05 outcome and the value of that outcome and multiply them together to get the expected value, that expected value is greater than that 100 pounds for you. And then maybe you say, oh yeah, that's okay. I guess that seems correct. You hand over the 100 pounds. And now what Bostrom thinks, and what almost everyone who works in this thinks, is that that's absurd. That if you've got a model, if you've got a decision theory, it's actually not about morality, it's about your theory of rationality, or your decision theory. But if you've got a decision theory that says you should give the wallet to the mugger, then something has gone badly wrong.
Starting point is 00:28:46 And one way that I like to illustrate this, and this applies to the, and this problem gets called the fanaticism problem. And it's a really interesting area of decision theory because I actually think it's the strongest argument against a theory that is otherwise very good, which is expected value theory. But seeing one way of this going wrong is if this argument is used for giving your money to the mugger or for, you know, looking for future amounts of value, even if it were just one in a billion chance that you could change it, well, why are you going after only finite amounts of value? There's a long religious tradition that says you can get infinite amounts of value via heaven. And so really the upshot, if you're taking expected value theory seriously, is to become a missionary to try and ensure that you get into as many
Starting point is 00:29:37 heavens as possible, according to as many different forms of religious belief as possible, and try and convince others to the same, to do the same. And that's where expected value theory, if you're really going with the math and doing it, takes you. And I know of zero people who have done that. Everyone says, look, okay, something, we don't know what, I think this is an incredibly hard area within decision theory. We don't know what the right answer is, but it's not that. And the argument for concern about the long-term future is the fact that these risks are more like the risks of dying in a car crash, and much less like the one-in-a-billion-billion Pascal's mugger kind of risk. Well, admittedly, this is a problem I have not thought much about, but I have a couple
Starting point is 00:30:26 of intuitions here kindling as I listen to you describe it. One is that, obviously, expected value becomes eminently rational within the domain of normal probabilities. I mean, it works in poker, it works in any kind of decision analysis where you're just trying to figure out what you should do under uncertainty, and you're dealing with probabilities of, you know, one-third or 10%. That is how we organize our intuitions about the future in the most systematic way. And I guess there are even corner conditions within the domain of normal probabilities where you basically, this is not a repeated game. You've got one shot to take this risk. And it may be that, you know, risking your life, whatever you put on the other side of
Starting point is 00:31:15 the scale doesn't make sense, except we don't live that way. We risk our lives every time we get into a car or get on a plane or ride a bike or do it. We've all made these trade-offs, whether we've actually done the decision analysis or not and consciously assigned probabilities or not. We just have this vague sense that certain types of risks are acceptable. And I guess my other intuition is that this is somewhat analogous to the problems disgorged by population ethics, where when we talk about things like the repugnant conclusion, and you're talking about very large numbers of very small things, or very small utility, you seem to get these grotesque outcomes ethically. Where do you think this is breaking
Starting point is 00:32:00 down? Or why do you think it is breaking down? So the standard response within decision theory would be to have what's, so it all gets technical quickly. And I apologize if I don't do a good job of explaining the ideas. But the standard response is to have what's called a bounded value function. And it's easiest to see this when it comes to money, let's say. So if the mugger in Pascal's Mugging was saying, okay, I will take £100 off you today, and I will come back tomorrow with any finite amount of money you like that you can spend on yourself.
Starting point is 00:32:37 Let's just assume you're completely self-interested to keep it simple. I could say, look, my probability of you coming back is so low that there is no amount of money that you could give me such that the probability times that amount of money would add up to more expected well-being for me than £100 now. So let's say I think it's one in a billion billion that the mugger will come back. Okay, sure. The mugger could give me, you know, a thousand billion billion dollars. And that would mean more, I would have greater expected financial return. However, I wouldn't have greater expected well-being because the more money I have,
Starting point is 00:33:25 the less it impacts my well-being. And there's some amount, there's only so much that money can buy for me. At some point, it starts making this like vanishingly small difference such that there's a level of well-being such that I can't get like a thousand times greater well-being with any amount more money. And so that's very clear when it comes to financial returns. But you might think the same thing happens with well-being. So supposing I have the best possible life, just like peaks of bliss and enjoyment and amazing relationships and creative insight over and over. And that's for some like very, very long lived life. Perhaps you just imagine that. Now you might think, oh, it's just not
Starting point is 00:34:13 possible to have a life that's a thousand times better than that life we've just imagined. And in fact, I think that's how people do in fact reason. So even if you'd say to someone, look, you can have a 50-50 chance, you either die now or you get to live for as long as you like and choose when you pass away. I think most people don't take that bet. And that suggests that we actually think that perhaps the kind of,
Starting point is 00:34:41 I mean, there's extra issues with kind of, you know, very long lives, but we can reformulate the thought experiment. Either way, I think people just do think there's just an upper bound to how good a life could be. And as soon as you've accepted that, then in the self-interest case, you avoid, you can very readily avoid this kind of Pascal's mugging or Pascal's wager situation. And perhaps you can make the same move when it comes to moral value as well, where you're ultimately saying there's a limit to how good the world or the universe can be. And as soon as you've got that, then you avoid Pascal's mugging, you avoid Pascal's wager.
Starting point is 00:35:22 However, you do get into other problems. and so that's why it's so difficult. Yeah, I mean, the thing that I think is especially novel about what you've been doing and about EA in general, which I always tend to emphasize in this conversation, I think I've brought this up with you many times before, but its novelty has not faded for me. And here's the non-novel part about which someone could be somewhat cynical. There's just this basic recognition that some charities are much more effective than others, and we should want to know which are which and prioritize our resources accordingly. I mean, that's the effective part of effective altruism. That's great, and that's not what has captivated me about this. I think the thing that is most interesting to me is by prioritizing effectiveness in
Starting point is 00:36:11 an unsentimental way, what you've done for me is you've uncoupled the ethics of what we're doing from positive psychology. Really just divorcing the question of how I can do the most good from the question of what I find to be most rewarding about doing good in the world. I'm always tempted to figure out how to marry them more and more, because I think that would be the most motivating and, by definition, the most rewarding way to live. But I'm truly agnostic as to how married they might be. And when I survey my own experience, I notice they're really not married at all. For instance, over at Waking Up, we launched this campaign, which we've called 100 Days of Giving. And so for 100 days, we're giving
Starting point is 00:37:02 $10,000 a day to various charities, and we're having subscribers over at Waking Up decide which charities among the list that you and friends have helped me vet. So every single day, we're giving $10,000 away to some great cause, and we're about 30 days into this campaign. cause. And we're about 30 days into this campaign. And honestly, I think for most of those 30 days, I have not thought about it at all. Right. I mean, literally, like it has not even been a 10 second memory on any given day. And, you know, I consider that on some level a feature rather than a bug. But it is striking to me that in terms of my own
Starting point is 00:37:47 personal well-being and my own sense of connection to the well-being of others, I might actually be more rewarded. In fact, I think I'm almost guaranteed to be more rewarded by simply holding the door open for someone walking into a coffee shop later this afternoon and just having a moment of, you know, a smile, right? Like that absolutely trivial moment of good feeling will be more salient to me than the $10,000 we're giving today and tomorrow and the next day and onward, you know, for months. So I'm just wondering what you do with that. I mean, it's a bug and a feature. It's a feature because I really do think
Starting point is 00:38:28 this is how you do the most good. You automate these things, you decide rationally what you want to do, and you put that in place, and then it's no longer vulnerable to the vagaries of your own moral intuitions or you getting captivated by some other bright, shiny object
Starting point is 00:38:43 and forgetting about all the good you intended to do. But I do hold out some hope that we can actually draw our feelings of well-being more and more directly from the actual good we're accomplishing in the world. And I just think there's more to think about on the front of how to accomplish that. Yeah, you're exactly right. And it's an old chestnut, but it's true that our psychology, human psychology, did not evolve to respond to the modern world. It evolved for the ancestral environment, our community, the people that we could affect amounted to the dozens of people, perhaps. Certainly the idea that you could impact someone on the other side of the world or in generations to come, that was not true in the
Starting point is 00:39:37 way it's true today. And so we have a radical misalignment between our the moral sentiments that we have evolved and the moral reality that we face today and so one thing either there's this project that effective altruism as a community has tried to embark on to some extent which is just trying to align those two things a little bit better and it's tough because we're not fundamentally aligned. But the ideal circumstance is where you are part of a community where you can get the reward and support and reassurance for doing the thing that actually helps other people the most, rather than the one where you get that immediate kind of warm glow. And that's not to say you shouldn't do those things too. I certainly do. And I think it's part of living a good, well-rounded life. But it's just a fact that there are people
Starting point is 00:40:31 in very poor countries who are dying of easily preventable diseases. We can take action to help them. There are animals suffering greatly on factory farms. We can take action to alleviate that suffering. There are risks that are maybe going to affect the entire fate of the planet, as well as killing billions of people in the present generation. Ah, like that is not something that my monkey brain is equipped to deal with. And so, you know, we're just, we've got to try as best we can. And that's something, at least for me, where having a community of people who share this idea, who I can talk these ideas through with, and who can, if needed, if I find these things hard, provide support and reassurance, that's at least a partial solution to this problem to get my monkey brain to care more
Starting point is 00:41:23 about the things that really matter. Well, so are there any other problems that you've noticed or that have been raised by EA's critics that you think we should touch on before we jump into the book and this discussion of long-termism and existential risk? Sure. I think the number one thing that is in some sense a problem, but I think is just kind of where the debate should be, is just what follows from long-termism, where the issue, the thoughts that I want to promote and get out into the world are that future people count. There could be enormous numbers of them. And we really can and in fact are impacting how their lives go. And then there's the second question of, well, what do we do about
Starting point is 00:42:12 that? What's the top priorities for the world today? Is that the standard environmentalist answer of mitigate climate change, reduce resource depletion, reduce species loss? Is it the classic existential risk answer of reduce the risk of misaligned AI, reduce the risk of a worst case pandemic? Is it the idea of just promoting better values? Is it something that I think you're sympathetic to and others like Tyler Coe and Apatik Collison are sympathetic to, which is just generally make the world saner and more well-functioning because, you know, predicting the future is so hard. Or is it perhaps about just trying to put the next generation into a situation such that they can make a better and more informed decision and have at least the option to decide
Starting point is 00:43:00 kind of how their future will go? These are extremely hard questions. decide kind of how their future will go. These are extremely hard questions. And, you know, I actively want to see an enormous amount of debate on what follows from that. You know, perhaps a thousand times as much work as has gone into it, I would love to see discussing these issues. And there's an enormous, these are hard, there's an enormous amount to say. And so that's where I think the kind of real meat of debate lies. Okay, so let's talk about the book, because you spoke a couple of minutes ago about how we are misaligned in psychological terms by virtue of imperfect evolution to think about problems at scale, and that's certainly true, but we have made impressive progress with respect to scaling our concern about our fellow human beings
Starting point is 00:43:55 over physical distance. Distance with respect to space, we kind of understand ethically now. I mean, Peter Singer has done a lot of work on this topic, urging us to expand the circle of our moral concern. And I think something like 100% of the people listening to us will understand that you can't be a good person and say that you only care about people who live within a thousand miles of your home, right? I mean, that's just not compatible with life in the 21st century. Although there's some discounting function, no doubt, and I think normal psychology would cause us to care more about family and
Starting point is 00:44:39 friends than about perfect strangers, and certainly perfect strangers in faraway cultures for which we have no affinity, etc. And we can, I think we've discussed this in previous conversations, just what might be normative there, whether we actually would want perfect indifference and dispassion with respect to all those variables, or if there's some hierarchy of concern that's still appropriate. And I think it probably is. But the topic of today's conversation and of your book is just how abysmally we have failed to generate strong intuitions about the moral responsibility we have and the moral opportunity we have to safeguard the future. So perhaps you can, with that set up, you can give me the elevator synopsis of the new book.
Starting point is 00:45:33 Sure. So like I said, this book is about long-termism, about just taking seriously the future that's ahead of us, taking seriously the moral value that future generations have, and then also taking seriously the fact that what we do as a society really will impact the very long term. And I think there are two categories of things that impact the very long term, where there's things that could just end civilization altogether. And this concern really got on society's radar with the advent of nuclear weapons. I think an all-out nuclear war, I think unlikely to kill literally everybody, but it could result in complete, you know, it would result in hundreds of millions, maybe even billions dead, unimaginable tragedy, which I think would make the world worse, not just for those people, but for the extremely long time, and could result in
Starting point is 00:46:30 the collapse of civilization too. At the more extreme level, there's worries about, in particular, engineered pandemics. And this is a risk that we've been worrying about for many years now, This is a risk that we've been worrying about for many years now, where pandemics occur occasionally and are extremely bad. And so society doesn't prepare for them in the way they should. I think the risk is going to get much worse over time, because not only do we have to contend with natural pandemics, we will also have to contend with pandemics from viruses that have been enhanced to have more destructive properties, or even just created de novo, such as in a bioweapons program. pandemics from viruses that have been enhanced to have more destructive properties, or even just created de novo, such as in a bioweapons program. This is a very scary prospect, and current estimates put the risk of all-out catastrophe, 95% dead, something like 1%, essentially. A second kind of way, set of events that could impact the very long term,
Starting point is 00:47:28 way, set of events that could impact the very long term are things that don't impact the length of civilization. That is, it wouldn't kill us off early, but does affect the value of that civilization, kind of civilization's quality of life, as it were. And in What We Are the Future, I focus in particular on values, where we're used to values changing an enormous amount over time. But that's actually something that I think is contingent about our world. And I think there are ways, especially via technological and political developments, that that change could slow. And in the worst case, our future could be one of perpetual totalitarian dictatorship. That would be very scary indeed. That would be a loss of almost all value that we could have and an existential risk itself, even if that civilization were very long lived. And so what can we do? And things vary in terms
Starting point is 00:48:16 of how tractable they are, like how much there is actually to do. But there are many things I think we can do. So on the ensuring we have a future side of things, I actually think that there's very concrete actions we can take to reduce the risk of the next pandemic, such as technology to monitor wastewater for new pathogens, such as certain, there's this new form of lighting called Far UVC that I'm extremely excited about. If we can get the cost down enough,
Starting point is 00:48:44 then we can just be sterilizing rooms in an ongoing way, in a way that's not harmful for human health. Or on the values kind of side of things, well, we can be pushing for people to be more altruistic, more reasonable in the way they think about model matters, perhaps more impartial to taking into seriously the interests of all sentient beings, whether they're human or animal or other, or whether they're now or in the future. And we can be kind of guarding against the lie, like certain, you know, authoritarian tendencies that I think are very common, and we have seen in history over and over again, and that I think are quite scary from a long-term perspective. And then lastly, we can just start reasoning more about this. I said earlier that,
Starting point is 00:49:29 you know, I don't know, I'm not like super confident about what we should be doing, but certainly we can be doing more investigation, building a movement that takes seriously, oh yeah, like there's this enormous potential future ahead of us, and we can do an enormous amount of good to make it better. Yeah, I want to talk about value lock-in and some of the issues there. But before we get there, I think we should linger on what makes thinking about the future so difficult. I think you say somewhere in the book that, both with respect to distance in space and distance in time, I think you say something like we mistake distance for unreality, right? Like the fact that something is far away from us
Starting point is 00:50:14 makes part of our brain suspect that it may not quite exist. And this is, again, I think this has changed fairly categorically with respect to space. You just can't really imagine that the people suffering over in Yemen don't exist. But it's much harder to assimilate this new intuition with respect to time, because in this case, we are talking about people who don't actually exist. There'll be some number of thousands of people, probably 3,000 to 5,000 people or so, born over the course of our having this conversation, globally speaking. I think the math is somewhere around there. And so it's easy to argue, okay, these are people who are coming into this world for which they have no responsibility. And we, the present generation or generations, are entirely responsible for the quality of the world they inherit and the quality of the lives they begin to lead. And I would say we don't, collectively know, collectively, we don't think nearly enough about that responsibility. But when you're talking about people who may or may not be born
Starting point is 00:51:32 in some distant future, we don't think about these people at all. And I would, I suspect that most people wouldn't see any problem with that because it's just, when you think about what a crime it would be for us to foreclose the future in some way, you know, if we all die in our sleep tonight painlessly and there is no human future, what's bad about that? joy and happiness and creativity that exists in potentia on that side of the ledger. But there are no victims deprived of all of that goodness because they don't exist, right? So this is the very essence of a victimless crime. And I think you and I have touched that issue in a previous conversation. But it's just hard to think about the far future, much less prioritize it. And I'm just wondering if there's something we can do to shore up our intuitions here. I guess the first point I would make is that the only reason why these people are hypothetical
Starting point is 00:52:36 is because you, the listener, are holding out some possibility that they may not exist, when you ask yourself, what would cause subsequent future generations of people not to exist, perhaps there's some other version, but the only thing I can imagine is some catastrophe of the sort that you have begun sketching out, right? Some pandemic, some asteroid impact, some nuclear war, some combination of those things that destroy everything for everyone, which is to say some absolutely horrific cataclysm, which we should be highly motivated to avoid, right? So if we successfully avoid the worst possible outcome for ourselves and our children and our grandchildren, the future people we're talking about are guaranteed to exist. It's only time shifted. It's not, in fact, in question. And the
Starting point is 00:53:32 only thing that would put it in question is some horrific outcome that we are responsible for foreseeing and avoiding. And really, there's no one else to do the job. Absolutely. I mean, so, And really, there's no one else to do the job. Absolutely. I mean, so yeah, there's this general issue where why do we care about people on the other side of the planet much more now than we did, let's say, 100 years ago? And it's clear that a big part of it is that now we can see people on the other side of the world. We get information about them in a way that we never used to, where literally there can
Starting point is 00:54:03 be a live stream video of someone in Nigeria or Pakistan or India or any other country around the world. So that, you know, those people's lives become vivid to us in a way that's just not possible for future generations. And that makes this challenge just so deep, in fact, to get people to care. If there were news from the future, you know, the newspaper of tomorrow, I think we wouldn't be in such a bad state. You know, we could report on what the world is like in the year 2100 or 2200. And I think people would care. But that's not the news we get. It's all abstract. And that's really tough. The thing I'll say is, so, you know, you mentioned the idea of
Starting point is 00:54:46 some catastrophe just painlessly kills all of us in our sleep. And, you know, what would that be a tragedy? And I mean, I think yes. And I argue in the book at length that it would be a tragedy, both because I think that we should expect the future to be on balance good, and secondly because I think the loss of lives that are at least sufficiently good, and certainly the loss of all the moral and artistic and technological progress that we might see in the future, that those are great losses. But I want to emphasize the case for long-termism does not rely on that. You could think that everyone dying painlessly in their sleep would not be a bad thing, morally speaking. Nonetheless, a future of perpetual totalitarian dictatorship really would be compared to a wonderful, free, thriving, and liberal future
Starting point is 00:55:38 where different people can pursue their visions of the good. And we've eradicated issues that plague today like injustice and poverty. I just think you really don't need to engage in much moral philosophy to say that if we have a choice between one or the other, then it's better and much better to think about the thriving, flourishing future than this dystopian one. Yeah, no, that's a great point. Forget about the thought experiments. It should be pretty easy to acknowledge that there's a difference between having to live under the thousand-year Reich and, about the significance of the future? I mean, I guess I'm wondering if you've thought at all about how just, because again, this is, it's surprising that I think we think about this almost not at all, and yet our assumptions about
Starting point is 00:56:38 it do a lot of psychological work for us. I mean, like I really, you know, I mean, this is my job to have conversations like this, and I'm surprised at how little I think about the far future, right? In any kind of significant way. And yet, if I imagine how I would feel if I knew that it wouldn't exist, right? I mean, if we just figured out that, oh, actually, the whole story ends in 80 years, right? So we're unlikely to be around to suffer much, or we're going to be out of the game pretty soon. And people haven't had kids, don't need to worry about them. And so the basic picture is just the career of the species is radically foreclosed at some point inside of a century. What would that do to everything? And how would that sap the meaning from what we do? And it's just an interesting question, which I haven't thought
Starting point is 00:57:41 much about. But do you have any intuitions about that? It's a great thought experiment, and it's represented in this film, Children of Men, where the global population becomes sterile. And it's discussed at length in the writings of philosopher Samuel Scheffler in a book called Death and the Afterlife. And he argues that if it were the case that just, yeah, there's no world at all, let's say, past our generation or even our children's generation, it would sap enormous amounts of, in fact, perhaps most meaning from our lives. This is actually a point made by John Stuart Mill in the end of the 19th century, where he briefly had this phase of being concerned by future generations, arguing actually that it means they should be keeping coal in the ground, which I think wasn't quite the right, you know, it's maybe made sense given their views at the time or something,
Starting point is 00:58:32 that their views were incorrect. But he has this beautiful speech to Parliament, and he has this wonderful line, why should we care about posterity? After all, what has posterity ever done for us? And he goes on to argue that actually, like, posterity has done an enormous amount for us because it gives our projects meaning. Where, you know, we build, we plant trees that future generations will sit under.
Starting point is 00:58:56 We build cathedrals that are like a testament to our time and can be recognized by future generations. When we engage in activities like moral progress and political progress and scientific and technological progress, part of how we have meaning in doing those things is because it's like we're part of this grand relay race where we have taken the baton from previous generations, the things they did badly and the things they did well.
Starting point is 00:59:23 And we can hand that baton on to future generations, making the world a little bit better, giving them like, you know, a little bit of a higher platform to stand on. And yeah, I think this is just a very widely shared view. And I think it kind of resonates with me intuitively too, where if I found out that everything was going to end in 50 years, I wouldn't be like, oh, well, I guess I'll just, you know, focus on issues that just concern the near term then. I would, you know, I would be devastated, even if that didn't personally affect anyone I knew or, you know, perhaps it's not 50 years, perhaps it's 200 years.
Starting point is 00:59:59 Nonetheless, I would still, I would be absolutely gutted. Yeah, it's really interesting because it's not something that I've ever been aware of, but it is an integral piece of how I think about what we're doing here, which is when you think about human progress, you think about all the good things we do and all the bad things we can avoid. The idea that that is all going to be brought up short in a few short years, right? I mean, we just think of like, well, what's the point of building increasingly wonderful and powerful and benign technology or creating beautiful art or literature or just making
Starting point is 01:00:39 progress on any front philosophically? on any front, philosophically. I mean, beyond just securing your day-to-day comfort and basic happiness, there's something forward-leaning about all of that. It's a check that, in some sense, never gets cashed or fully cashed, or it just gets continually cashed in the future, if not by our future selves, by others to which we're connected. And the idea that we could ever know that there's just a few more years of this and then it's all over, I mean, we all know that personally with respect to our own mortality. But there is something about the open-endedness of the whole project for others. You know, even if you don't have kids, that changes. Again, I really don't even have my intuitions around this are so unformed, but it really does
Starting point is 01:01:34 feel like it changes something about day to day life. If the rug were truly pulled out from under us in a way that we had to acknowledge, you know, leaving everything else intact, right? I mean, all the buildings are still standing. We've got, you know, there's still a concert Saturday night in the same auditorium and you have tickets, but you knew the world was going to end and you were just going to get a collective dial tone at some point in the near future. You know, I don't know to what degree, but it would change everything to some degree. Yeah. So in What We Are With The Future,
Starting point is 01:02:09 I actually start and end the book on the same note, which is a note of gratitude, where the book is dedicated to my parents, Maya and Robin, and their parents, Tom and Nina and Frank and Daphne, and so on, all the way back, all the history that led us here. And at the very end of the book, in the very last page, there's, in the print version, there's this QR code. And if you scan it, it takes you to a little short story I wrote. I'm not claiming it's any good, but I did it because I think it's
Starting point is 01:02:44 so hard to visualize the future, especially to visualize how good the future could be. And in the end of that short story, one of the characters reflects on everything that led, and it's this depiction of this, I'm not saying it's the best possible future, but at least a good one. And one of the characters just reflects on just everything that led to her life today all of the hardship the ways in the past in which people endured sickness and suffering and rides on the subway packed in with other people who are you know miserable on their commute and just being deeply thankful for all of that and that's certainly how how I feel, where, you know, the past has given us,
Starting point is 01:03:26 past generations have given us this mixed bag. They've given us many ways in which the world is broken, many ways in which we still see enormous suffering. But they've also given us wonderful, good things too. I live in this like liberal democratic country that is exceptionally unusual by historical standards. I don't need to suffer pain when I engage in a surgery. I don't need to risk for my life from diseases like syphilis or tuberculosis. I can see the kind of, just on my phone, I can just see the most wonderful art produced
Starting point is 01:03:59 in the entire world at any time. I can see the world I can go traveling in, like India, Sri Lanka, Ethiopia. Wow, I'm just so grateful for all that's been done. The very least that I can do is try and take that mixed package of goods and goods and bads and try and pass it to the next generation. Try and ensure that there even are future generations that I can pass it to, but also to make it better. And so that's kind of, yeah. Normally I trade pass it to, but also to make it better. Normally, I trade in these logical, rational arguments, but at least in terms of what motivates me on a gut level,
Starting point is 01:04:33 that's a significant part of it. So we have something like 500 million to 1.3 billion good years left on this planet if the behavior of our sun is in any indication. So the challenge here is to not screw it up. And, you know, we can leave aside the possibility of reaching for the stars, which we will almost certainly do if we don't screw things up and, you know, spreading out across the galaxy. But even on this lowly planet, we have just an extraordinary amount of time to make things better and better. And we're just at the point where we can get our hands around the kinds of things that could ruin everything, not of our own making, like, you know, asteroid impacts and naturally spawned pandemics. And there's clear ways in which our time is unusual. I think at one point in the book,
Starting point is 01:05:32 you discuss how for virtually all of the past, we were culturally isolated from one another, and we may one day be again culturally isolated if we expand out across the galaxy, which is to say that there will be a time in the future where human communities grow far enough apart that what they do has really no implication for what anyone else does outside of their community. And that was true of most of human history until recent centuries. I think there's a more fundamental claim in your book, which is that there's good reason to believe that we are at something like a hinge moment in history where the things we do and don't do have an outsized effect over the trajectory of human life in the future. And, you know And there are things that have come up briefly so far
Starting point is 01:06:27 like value lock-in, which are related to that, which we'll talk about. But what do you think about this question of whether or not we're living at a hinge moment and how we can know and how we may or may not be mistaken about that? I think it's clear that we're living at an extremely unusual time that gives a good argument for thinking that we're living at an unusually impactful time too. And I'm not going
Starting point is 01:06:56 to claim we're living at the most unusual time or impactful time in history. Perhaps the next century will be even more influential again. But it's certainly very influential. And the core reason for this is just kind of think about the tree of technology that civilization can kind of go through from the very first stage, there's flint axes and spear throwers and fire and the wheel. And then nowadays, there's maybe it's fusion technology and advanced machine learning and biotechnology that are current developments and at some point there will be future technology that perhaps we've not even imagined yet here's the question just relative to both the past and the future how quickly are we moving through that technology to the
Starting point is 01:07:42 and i think there's very good arguments for thinking that it's much faster. We are moving much faster through that technology tree than for almost all of human history and almost all of the future if we don't go extinct in the near term. And I think that's clear. So economists measure this generally with economic growth, where at least a significant fraction of economic growth comes from technological progress. And the economy is currently growing far, far faster than it has done for almost all of human history, where today the economy doubles every 20 years or so. So it's growing around 2% or so, and compounding that, it gives you a doubling time of something like 25 years or thereabouts. Exactly. Yeah, 2 two three percent when we were hunter-gatherers growth was
Starting point is 01:08:30 very close to zero as farmers growth was more like 0.1 percent so a doubling every a few hundred years or so and so relative to the past we are speeding through technological development much much faster than before, you know, for most of human history before. I mean, honestly, I say before the scientific industrial revolution, really, it's before about the year 1900, because there was a slower kind of ramp up. But then secondly, I think we're speeding through that progress much faster than for most of the future too. And the argument for this is that we'll just assume we keep growing at about two percent per year for another 10 000 years how much growth would there have to be in that case well it turns
Starting point is 01:09:15 out because of the power of compound growth that it would mean that for every atom within reach that is within 10,000 light years, there would be, I think, it's 100 billion trillion current civilizations worth of output. And now, you know, maybe that's possible. I think it's extremely unlikely. this level of progress, of economic and technological progress, we would get very close to basically discovering almost everything that is to discover on the order of a few thousand years. However, as you said, there are hundreds of millions of years to go potentially. And so that means that, as I said, we're moving particularly quickly through the tree of technologies. moving particularly quickly through the tree of technologies. And every technology comes with its distribution of powers and capabilities and benefits and disadvantages. So if we take the ability to harness energy from the atom, well, that gave us clean energy. It gave fire
Starting point is 01:10:20 nuclear power. That's an enormous benefit to society. It also gave us nuclear weapons that could threaten wars more destructive than any we've ever seen in history. And I think future technologies, such as biotechnology that can allow us to create new viruses, or advanced artificial intelligence, that each one of these give us this new distribution of powers. And some of those powers can have extraordinarily long-lasting impacts, such as the power to destroy civilization, or such as the power to lock in a particular narrow ideology for all time. And that's why the world today is at such a crucial moment. Yeah, the principle of compounding has other relevance here when it comes to this question
Starting point is 01:11:15 of just how influential a moment this might be in history. When you just look at how effects of any kind compound over time, it does cast back upon the present kind of an inordinate responsibility to safeguard the future, because we have this enormous potential influence. I mean, just the slightest deflection of trajectory now potentially has massive implications, you know, 20, 30, 40 years from now. And it's obviously hard to do the math on these things. It's hard to predict the full set of consequences from any change we make. But if you just look at what it means to steadily diminish the probability of a full-scale nuclear war year after year, even just by 2% a year, if we could just get that virtuous compounding to happen, roll that same logic to every other problem you care about, it's very easy to see that
Starting point is 01:12:20 getting our act together, however incrementally now on all these fronts, matters enormously in the not-too-distant future. Absolutely. I mean, my colleague Toby Ordon, his wonderful book, The Precipice, just suggests that, well, what's happening now is that our technology is advancing more rapidly than our wisdom. And it's a beautiful beautiful idea and what we're trying to do is accelerate wisdom you know we're trying to just make us a little bit wiser a little bit more cognizant of you know there's most people fall into one of two camps there are the people who are really into technology the kind of stereotype of the silicon valley people and they think
Starting point is 01:13:03 technology is great and we just need to keep advancing it. And then there are people who are not really into technology and the people who are very skeptical of Silicon Valley and worry about what technology could be doing to society. And I think I just really want people to inhabit both frames of mind at once. Technology has benefits and costs. Some technologies are just like anesthetic. I'm like, oh yeah, that's just really good. Let's have more of these like, you know, no downside, unashamedly good technologies. Let's speed them up faster. For technologies that are very dual use, or just extremely powerful, and we don't know how to handle them. Well, let's then be careful and considerate about how we're developing these technologies, have good norms and regulations
Starting point is 01:13:51 in place so that we don't basically just walk backwards into a chasm where we wreak global disaster because we're not even paying attention to the impacts that technologies that are just around the corner will have because of this reason of out of sight and out of mind because we're barely thinking about you know we barely think about like the end of the next year certainly not the end or the end of the next political cycle let alone the decades ahead that could bring about radical changes and how they might impact the centuries ahead. Well, simply by uttering the phrase political cycle, you have referenced what I consider a major problem here, which is we have a politics that is almost by
Starting point is 01:14:40 definition a machine of short-termism, whereas it whereas taking a long-term view is antithetical to the whole project of becoming president and staying president and then raising money for your next campaign. I mean, you're just not, to speak specifically of America at this point, I don't know that it's much better in any other country, you're not incentivized to think about the far future and to figure out how to help all those people whose names you will never know and who will never vote for you, much less give you money.
Starting point is 01:15:13 It's just not in the cards, right? And so one thing that has to be a fairly high priority is to figure out how to engineer a system of political and economic incentives such that we more effortlessly take the future into account. Absolutely. And honestly, it's extremely tough. So I've worked on this topic. I have an article on long-termist political institutions with another philosopher, Tyler John. And there are some things you can do, such as, you know, some countries have an ombudsperson to represent future generations.
Starting point is 01:15:52 I'm particularly attracted to the idea of at least an advisory, at least as an advisory body, a kind of citizens' assembly, so a randomly selected group of people from the populace who are explicitly given the task of representing the interests of future generations and then making recommendations about what the government could do differently in order to make the future go better. I think we can also build in just more information, so more think tanks or quasi-governmental organizations that have the task of really trying to project not just the next few years, but the next few decades, even the next few centuries.
Starting point is 01:16:32 These are some changes we can make. However, the fundamental problem is that future generations are being impacted without representation. they do not get a vote. And the fundamental problem is we cannot give them a vote. There's no way they can represent themselves. And what do we do about that? Honestly, I think it's going to be very hard to make major progress unless there's some significant cultural change. So if we imagine a world where governments are responding to the views and preferences and ideals of their populace, of the electorate, and that electorate cares about the long-term, well, that's a situation where we can align political incentives with the long-term good.
Starting point is 01:17:19 And so that's what I think we really need to get to. That's what I think we really need to get to. It's just amazing how hard it is, even for normal people who have children, to prioritize the next decade. A decade where they have every expectation of being alive and suffering the consequences of our present actions, and they have certainly every expectation that their children will be alive in those decades. When you look at a problem like climate change or anything else that is slowish moving and has various features like, you know, being in part a tragedy of the commons, running counter to any individual's personal economic incentives, etc. It's just, it all falls by the wayside, even when you're not talking about hypothetical people who you will never meet, right? I mean, it's just, there's something so hard about focusing on the big picture. And so, yeah, it does, I mean, when you talk about cultural talk about cultural change, it's hard to see the thing
Starting point is 01:18:27 that would be so sweeping that it would capture more or less everybody all at once. But perhaps that's not even the right goal. I mean, maybe there are only something like 5,000 people whose minds would have to change to fundamentally redirect human energy. I mean, this is an elitist claim and argument, but I suspect that something like that might also be true. Yeah, I think I'm more optimistic on this in general, perhaps, than you are, where, yeah, in the course of doing research for What We Owe the Future, I, you know, dug a lot into the history of various social and moral movements, such as environmentalism, feminism, animal welfare movement, and in particular, the abolition of slavery. And one thing I think it's
Starting point is 01:19:21 easy to underestimate is just how weird these ideas were in the past, at least in particular for the kind of political elite, or at least the people who had power, where you don't have to go that long ago. Let's just focus on the environmentalist movement. I think there's many analogies here where the environment can't represent itself. It has an advantage that you can see the environment, you can relate to it. But 150 years ago, concern for the environment was really not on the table. Maybe you had poets, Williams Wordsworth, starting to cultivate love for nature, but there just really wasn't anything like a movement. Whereas now, it is just a part of common sense morality, at least in many countries around the world, that is, by historical
Starting point is 01:20:06 standards, that is amazingly rapid model change. And so at least if we're able to say, look, we're not going to be able to make this change to convince everyone of this long-term outlook, to convince everyone of having concern for future generations. We're not going to be able to convince them within the next few years or even two decades. But when we look like 100 years out, maybe we can. Maybe actually we really can affect that sort of sweeping moral change where concern for future generations is just on the same level as concern for other people in your community or people of a different ethnicity than you.
Starting point is 01:20:55 I'm holding out hope. I'm holding out hope that those sirens aren't people coming to arrest you for your thought crimes. I'm sorry. This is in New York. Are you recording this podcast while robbing a bank? Look, it's for the greater good, right? The money can go further.
Starting point is 01:21:14 So that's a, you know, don't do that. Just in case my followers take me too literally. Let's go on record. We're against bank robbery. Yeah. At least currently. I'm with that. I won't say anything about value lock-in on that point.
Starting point is 01:21:28 Let's talk about this issue of value lock-in and what you call moments of plasticity and trajectory changes. How do you think about, and you might want to define some of those terms, but what is the issue and opportunity here? Sure. So at the moment, there's a wide diversity of moral views in the world and rapid moral change. So the gay rights movement can build in power and momentum in the 60s and 70s, and then you get legalization of gay marriage a few decades later.
Starting point is 01:22:03 Clear example of positive moral change. And so we're very used to that. We think it's kind of just part of reality. We wouldn't expect that that might change in the future, such that moral change might come to an end. The idea of value lock-in is taken seriously that maybe it will, actually. Maybe moral progress will stall. And so value lock-in is the idea that some particular set of, some particular ideology or moral view, or kind of narrow set of ideologies or moral views, could become globally dominant and persist for an extremely long time. And I'm worried about that. I think that's a way in which the long-term future could go that could just lead to enormous loss of value, could make the future dystopia rather than a positive one.
Starting point is 01:22:54 And, you know, it's easiest to see this by starting off with maybe the most extreme case, and then kind of thinking about ways you might not be quite that extreme, but close. And the most extreme case is just, imagine a different history where it wasn't, you know, the US and UK and liberal countries that won World War II. And instead it was Stalinist USSR,
Starting point is 01:23:21 or let's say the Nazis who won. Now they really aimed ultimately at global domination. They wanted to create a thousand year empire. If that had happened and they had a very narrow and morally abominable ideology, if they had succeeded, would that have been able to persist forever, or at least for an extremely long time? And I think actually the answer is yes, where ideologies in general and social systems can certainly persist for centuries or even thousands of years. So the major world religions are thousands of years old. But technology, I think over time, will give us greater and greater power to control the values of the population that lives in the world
Starting point is 01:24:07 and the values in the future as well. Where at the point in time where we develop the ability to like stop aging such that people's lifespans could last for much, much longer. Well, then a single dictator could, you know, it's not confined to the lifespan of it. The rule is not confined to the lifespan of a single individual, 70 years, but instead would be thousands of years or however long that ruler would last. Even more extreme, I think, is the point of time at which the ruling beings are digital rather than biological. And again, this feels like sci-fi but you know we're talking about i actually think many people i respect greatly think that moment is coming in
Starting point is 01:24:51 in a few decades time but let's just say it's like you know plausible within this century or centuries well consider what the environment is like for a ruler who is a digital being. Well, they don't die. They are in principle immortal. Software can replicate itself, and any piece of machinery would wear out, but the AI systems themselves could persist forever. So one of the main reasons why we get moral change over time, namely that people die and
Starting point is 01:25:25 are replaced by people who have slightly different views, that would be undermined. I think in this kind of scenario, other causes of change would be undermined too, or taken away. So I think the fact that we have competition between different model ideas, because we have this diversity of model ideas. Well, if the Nazis had won World War II, if they had established a world government, you know, again, I'm not claiming this is likely or possible, but imagine the counterfactual history. Well, then we would lose out on that competition of ideas.
Starting point is 01:25:55 There wouldn't be that pressure for model ideas to change. And then finally, it's just potentially a changing environment. Like upsets happen, you know, civilizations fall apart and so on. But I think technology in the future could allow for greater and greater stability. So again, I think AI is particularly worrying here where, you know, again, we're imagining the leaders of the Nazis controlling the world. Okay, well, one way in which dictatorships end is insurrection, whether that's by the army or by the people. If your army is automated, if you have kind of robots policing the streets, then
Starting point is 01:26:31 that dynamic would end, similarly with your police force. And again, all of this seems like sci-fi, but I think we should be thinking about it. We should be taking it seriously, because it really could come sooner than we think. And the technologies we have today would have looked like wild sci-fi just a century ago. Yeah, well, maybe we'll close out with a conversation about AI, because I share your concern here, and I share your interest in elucidating the bad intuitions that make this not a concern for so many other people. So let's get there in a second. But remind me, what is the lock-in paradox that you talk about in the book? Sure. So the lock-in paradox, it's like the liberal paradox of tolerance. But I push in
Starting point is 01:27:20 the book that we don't want to lock in any particular values right now, because probably all of the values we're used to are abominable. You know, I think we're like Plato and Aristotle arguing about the good life, and we've not realized the fact that we all own slaves is, you know, maybe morally problematic. So I think that there's enormous amount of model progress yet to come, and we want to ensure that we guarantee that progress and so even if you know i think motivation of thinking oh well it's you know 21st early 21st century western liberal values that's the best thing and we should just ensure they happen forever i'm like no we need progress the lock-in paradox however is that if we wanted to create a world where we do get a sustained period of reflection and moral progress, that does mean we'll need some restraints. free debate commitment to tolerance of opposing moral views restrictions on ways of gaining power
Starting point is 01:28:29 that aren't via making good arguments and moving people over to your side kind of in the same way that like you know the u.s constitution at least aspirationally is trying to bring about this like liberal society where many different world views can coexist and where you can make moral progress but in order to have that and not fall into a dictatorial the dictatorship you needed to have these restraints on any individual's power and so i think we we may well need something like that for the whole world where we stop any particular ideology or just set of people from gaining too much power such that they can just control the world so that we can just have continued reflection, insight, empathy, and moral progress. statement of this, which is something that allows for the kind of incrementalism you've just sketched out and the error correction that I think is necessary for any progress, including ethical progress, but which locks in the methodology by which one would safeguard the principle of error correction and sensitivity to
Starting point is 01:29:49 the open-endedness of the exploration in ethical space, right? I mean, it's like, you know, I think there's probably, whether you define it as consequentialist or not, I think it's probably less important. I mean, the consequentialism, as we know, has its wrinkles. But I think there's an almost axiomatic claim about the primacy of safeguarding the well-being of conscious creatures at the bottom of, I think, any sane moral enterprise right now. It's not how you think about well-being, how you aspire to measure it or quantify it, how you imagine aggregating these quantities. There are many devils in all those details, but the basic claim that, you know, you have to, at bottom, care about the happiness and suffering of conscious creatures and marry that to a truly open-ended conception of, and, you know, perpetually refined and
Starting point is 01:30:47 refinable conception of what well-being is, all things considered, you know, and that's where all the discussion remains to be done. I think those are the bootstraps in some form that we have to pull ourselves up by. Yeah, so it is a very notable fact that I think basically all plausible moral views that are on the table see well-being as one part of the good. Perhaps they think other things are good too, art, flourishing natural environment, and so on. But at least, you know, happiness, broadly construed, flourishing life, avoidance of suffering, those are good things. And... I would just, well, just not to be too pedantic here, but in my brain, all of that collapses to just a fuller understanding of what well-being is. I don't think any, you know, reasonable
Starting point is 01:31:42 definition of well-being would exclude the things you just mentioned. And the cash value of the things you just mentioned has to be at some level how they impact the actual or potential experiences of conscious creatures. This is an argument I've made somewhere. If you're going to tell me you've got something locked in a box that is really important, it's really valuable, it's important that we consider its fate in everything we do, but this thing in the box cannot, will not affect the experience of any conscious creature now or in the future, right? Well, then I just think that's a a contradiction i mean we what we what we mean by value is something that could be actually or potentially valued by someone
Starting point is 01:32:31 somewhere great so i think we differ a little bit on this where i mean i also think like my best guess is that value is just a property of conscious experiences. And, you know, art and knowledge and the natural environment are all good things insofar as they impact the well-being of conscious experiences. But that's not something I would want to bake in. I'm not so certain in it. You know, there are other views where the satisfaction of carefully considered preferences is of positive value, even if that doesn't go via the change in any conscious experience. And other people do have the view that... But so again, that's the phrase I'm actually not familiar with. When I hear you say
Starting point is 01:33:15 satisfaction of anything, including preferences, aren't we talking about conscious experience, actual or potential? So on this view, it's no. So suppose that Aristotle wanted to be famous, and he wasn't very famous during his time. I'm actually not sure. Let's say that he was only much less famous than he is now. Does the fact that we are talking about Aristotle now increasing his fame, does that increase his well-being?
Starting point is 01:33:44 And it's obviously not impacting his consciousness. I'm prepared to answer that question. I'm sure you're prepared to say no. And honestly, and my best guess view is completely agreeing with you that anything we do to talk about Aristotle now, we could say he smells and we hate him. It doesn't change his well-being at all. However, that's not true. Other people disagree, and I think other smart people disagree. And so I certainly wouldn't want to have, wouldn't want to say, look, we figured that out. I would want to say, look, no, we've got plenty of time to debate this. I think I'm going to win the debate. Maybe the other person thinks they're going to win the
Starting point is 01:34:18 debate, but we've got all the time in the world to really try and figure this out. And I would want to be open to the possibility that I'm badly wrong. I guess we are kind of badly wrong on this. Well, that's interesting. I would love to debate that. So maybe as a sidebar, you and I can figure out who and when and where. Absolutely. Always happy to talk about, yeah. Okay, well, so let's talk about AI because it is, in some sense, the apotheosis of many of these concerns.
Starting point is 01:34:51 And I agree that should anything like a malevolent AGI be built or be put in the service of malevolent people, the prospect of value lock-in and Orwellian totalitarianism, it just goes way, way up. And that's worth worrying about, whatever you think about the importance of future generations. But there are several stumbling blocks on the path to taking these concerns about AI seriously. The first is, and this really is the first knee-jerk reaction of people in the field who don't take it seriously, it's the claim that we're nowhere near doing this, that our language models are still pretty dumb, even though they can produce some interesting texts at this point. There's certainly no concern that there could be anything like an
Starting point is 01:35:46 intelligence explosion in any of the machine learning algorithms we're currently running. AlphaZero, for all of its amazing work, is still not going to get away from us, and it's hard to envision what could allow it to. I know that there are people on our side of this debate who have given a kind of probability distribution over the future, giving it a 10% chance it might happen in 20 years and a 50% chance it'll happen within 50 years, etc. But in my mind, there's no need to even pretend to know when this is going to happen. The only thing you need to acknowledge is that not much needs to be true to make it guaranteed to happen. There are really two assumptions one needs to make. One assumption is we will continue to make progress in building intelligent machines at whatever rate, unless something terrible happens.
Starting point is 01:36:47 This brings us back to the collapse of civilization. The only thing that would cause us to stop making progress in building intelligent machines, given how valuable intelligence is, is something truly awful that renders some generation of future humans just unable to improve hardware and software. So there's that. I mean, so barring catastrophe, we're going to continue to make progress. And then the only other assumption you need is that intelligence is substrate independent, right? That it doesn't require the wetware of a biological brain. It can be instantiated in silico.
Starting point is 01:37:31 And we already know that's true, given the piecemeal AI we've already built. And we already know that you can build a calculator that's better than any human brain calculator. And you can do that in silicon. And there's just no reason to think that the input-output properties of a complex information processing system are magically transformed the moment you make that system out of meat. And so, again, given any rate of progress, and given the assumption of substrate independence, eventually, and again, the time horizon can be left totally unspecified, you know, five years
Starting point is 01:38:12 or 500 years, eventually we will be in the presence of intelligent machines that are far more intelligent and competent and powerful than we are. And the only other point of confusion that I've detected here again and again and again, which can be easily left to one side, is the question of consciousness, right? Will these machines be sentient? And that really is a red herring from my point of view. I mean, it's not a red herring ethically. It's incredibly important ethically because, youically because if they are conscious, well, then we can have a conversation about whether they're ethically suddenly more important than we are or at least are equals and whether we're creating machines that can suffer, et cetera.
Starting point is 01:38:58 If we're creating simulated worlds that are essentially hell realms for sentient algorithms, I mean, that would be a terrible thing to do. So yes, it's all very interesting to consider, but assuming that we can't currently know, and we may one day never know, even in the presence of machines that say they're conscious and that pass the Turing test with flying colors, we may still be uncertain as to whether or not they're conscious. Whether or not they're conscious is a question that is totally orthogonal to the question of what it will be like to be in the presence of machines that are that intelligent, right? I mean, I think you can dissociate, or very likely you can dissociate consciousness and intelligence. I mean, yes, it may be the case that the lights of consciousness magically come on
Starting point is 01:39:46 once you get a sufficient degree of complexity and intelligence on board. That may just be a fact of our world. And we certainly have reason to expect that consciousness arrives far sooner than that when you look around at animals who we deem conscious, who are far less intelligent than we are. So I do think they're just separable questions. But the question of intelligence is really straightforward. We're just talking about, you know, conscious or not, we're talking about machines that can engage in
Starting point is 01:40:16 goal-oriented behavior and learning in ways that are increasingly powerful and ultimately achieve a power that far exceeds our own, given that it seems virtually inevitable given these two measly assumptions, again, any progress and substrate independence for intelligence, we have to acknowledge that we or our children or our grandchildren or some generation of human beings stand a very good chance of finding themselves in relationship to intelligences that are far more powerful than they are. And that's an amazing situation to contemplate. And I'll just remind people of Stuart Russell's analogy here, which somehow, at least in my experience, changes people's intuitions. On his account, it's like we're in the following
Starting point is 01:41:11 situation. We just got a message from some distant point in the galaxy, which reads, people of Earth, we will arrive on your planet in 50 or 75 or 100 years, get ready, right? You know, you don't know when this is going to happen, but you know it's going to happen. And that statement of looming relationship delivered in that form would be quite a bit more arresting than the prospect of continued progress in AI. And, you know, I would argue that it shouldn't be. I mean, they're essentially the same case. Yeah, I just think that framing is excellent. I mean, another framing is if you're chimpanzees
Starting point is 01:41:56 and you see there's other species, homo sapiens, that are perhaps less powerful than you to begin with, but they seem to be getting rapidly increasing in their power, you'd at least be paying attention to that. And yeah, you remind me a little of the second book of The Free Body Problem, where the premise, I hope it's okay to say, is aliens will come.
Starting point is 01:42:19 And I think it's in a thousand years' time. And they have far greater power than civilization on Earth does. But the entire world just unites to take seriously that threat and work against it. And so essentially, I completely agree with you that the precise timelines are just not very important compared to just noting if, well, not if, but basically when it happens that we have AI systems that can do everything that a human being can do except better. That will be one of the most important moments in all of history, plausibly. And, well, we don't know exactly when it'll come,
Starting point is 01:42:59 but uncertainty counts both ways. It could be hundreds of years, and it could be soon, for all we know. But certainly, relative to the tiny amount of time we're thinking about it, well, we should be spending more. But I'll actually just comment on the timelines as well, where there's a recent survey, just came out last week, in fact, of just as many leading machine learning researchers as possible. I'm not sure of the sample size, but if it's like the last survey, it'll be in the hundreds. And they asked, well, at what point in time will you expect,
Starting point is 01:43:37 you know, do you expect a 50-50 chance of developing human-level machine intelligence? That is AI systems that are as good or better than human beings at all tasks, essentially, that humans can do. And they say 37 years. So that's in my lifetime. I will be entering my retirement
Starting point is 01:43:55 for the 50-50 chance of advanced AI. When I talk to people who are really trying to dig into this, they often have shorter timelines than that. They often think that it's 50-50 within more like 20 years and substantial probability on shorter timelines. I tend to be on the more skeptical end of that. But basically, within any of this range, concern about AI, it's not merely like, oh, we should be thinking about something that happens in 200 years because maybe that'll impact 2000 years. It's like, no, we should be thinking about something
Starting point is 01:44:30 that has a chance of catastrophe. We're on the same survey. The probability that those machine learning researchers, that's people who are building these things, who you think would be incentivized to say what they're doing is completely safe the probability of an extremely bad outcome as bad as human extinction or worse uh that have the typical probability was five percent of that so put those numbers together and you have 2.5 chance of dying or being disempowered by ai according to machine learning researchers, not cherry-pecked for being particularly concerned about this, you have a 2.5% chance of dying or being disempowered
Starting point is 01:45:13 by artificial intelligence in your lifetime. That's more than your chance of dying in a car crash. And we think a lot about dying in car crashes and take actions to avoid them. There's like a huge industry, a huge regulatory system so that people do not die in car crashes and take actions to avoid them. There's like a huge industry, a huge regulatory system so that people do not die in car crashes. What's the amount of attention on risks from AI? Well, it's growing.
Starting point is 01:45:32 Thankfully, it's growing and people are taking it really seriously. And I expect that to continue into the future. But it's still very small. It still means that smart, morally motivated people can make a really transformative difference in the field by getting involved, working on the technical side of things to ensure that AI systems are safe and honest, and to also work on the governance side of things to ensure that we as a society are developing this in a responsible way, perhaps digging into, you know, AI is not a monolith. There are different sorts of AI technology, some of which are more risky than others, and ensuring that the safety enhancing parts of AI come earlier, the more dangerous parts come later. This really is something that people can be contributing now. I think they are.
Starting point is 01:46:19 I think it's hard. I'm not claiming that this is like, it's clear cut what to do. There are questions over like how much progress we can make, but we can at least try. Yeah. And I think the crucial piece there, again, this is courtesy of Stuart Russell, who is my vote for, you know, one of the wisest people to listen to on this topic, just to make the safety conversation integral to the conversation about developing the technology in the first place. I mean, he asks us to imagine a world where, how perverse it would be if engineers who were designing bridges thought about bridge safety only as an afterthought and under great duress, right? So there's the question about building bridges, all of the resources put toward that end, but then only as an afterthought do engineers have conversations about what they call the
Starting point is 01:47:17 not falling down problem with respect to bridges, right? I mean, no, it's completely insane to conceive of the project of building bridges as being separable from the project of building bridges that don't fall down. Likewise, it should be completely insane to think about the prospect of building AGI without grappling with the problem of alignment and related matters. So in defense of more near-term concern here, many people have mined the history here and found these great anecdotes that reveal how bad we are at predicting technological progress. I forget the details here, but I think it was the New York Times or some esteemed U.S. paper that confidently published an editorial. I think it was just something like two months before the Wright brothers conquered flight, confidently predicting that human beings will never fly or it will take millions of years to engineer, because it took evolution millions of years to engineer the bird's wing. It would likewise take us millions of years to engineer the bird's wing. It would
Starting point is 01:48:25 likewise take us millions of years to fly. And then two months later, we've got people flying. And I think there's one anecdote that you have in your book that I hadn't heard, but the famous physiologist J.B.S. Haldane in 1927 predicted that... I was going to mention this one, yeah. I'll let you do it. Yeah, yeah. Go ahead. Yeah, go ahead. What was that reference? Oh, yeah, well, so, yeah. One of the great scientists of his time, evolution biologist,
Starting point is 01:48:54 and had some remarkable essays on, you know, the future and predicting future events and how large civilization could get. And he considered the question, will there ever be a return trip to the moon? And he said, yeah probably in eight million years and that was yeah did you say 1927 yeah yeah so that shows you like people like to make fun of people in the past making these bold uh scientific you know predictions about technology such as the early computer pioneers at the, I think it was the
Starting point is 01:49:26 Dartmouth Conference, saying, oh yeah, we'll build human-level artificial intelligence in about six months or so. And people mock that and say, oh, that's so absurd. But what they forget is that there are just as many people on the other side making absurd claims about how long technology will take. In fact, technology, look, things are hard to predict. Technology can come much later than you might expect. It can also come much sooner. And in the face of an uncertain future, we have to be preparing for the near-term scenarios as well as the long-term scenarios. Yeah, and the thing we have to get straight in the meantime is that the incentives are currently not appropriate for wisdom and circumspection here. I mean, it's just, we have something like an arms race condition,
Starting point is 01:50:13 both with respect to private efforts and, you know, rival governments, one must imagine, right? And we're not, you know, we haven't fused our moral horizons with the Chinese and the Israelis and anyone else who might be doing this work and gotten everyone to agree just how careful we need to be in the final yards of this adventure as we approach the end zone. And it does seem patently obvious that an arms race is not optimal for taking safety concerns seriously. It's not what we want. And I mean, the key issue is just how fast do things go? So supposing the development of artificial general intelligence, suppose that's 100 years away or 200 years away. Then it seems more likely to me that safety around AI will look like safety around other technologies. So you mentioned flight. In the early days of flight, a lot of planes crashed, a lot of people died. learned primarily by trial and error what systems make flight safer. And now very few people die in plane crashes compared to other modes of death,
Starting point is 01:51:31 especially in more developed countries. But we could learn by trial and error because a single plane crash would be a tragedy for the people involved and for their families, but it wouldn't destroy the whole world. And if AI development goes slowly, then we could, I, then I would expect we will learn by that same kind of process of trial and error and gradual improvement. But there are arguments for thinking that maybe it will go fast. In fact, leading economic models, if you assume that there is a point at which you get full substitutability between human labor and artificial labor, like as AI, in particular, the process of developing more advanced AI systems, then you get very fast growth indeed.
Starting point is 01:52:13 And if things are moving that quickly, well, this slow incremental process of trial and error that humans, that we as a society normally use to mitigate the risks of new technology, that process is substantially disempowered. And I think that increases the risks greatly. And again, it's something that I really think, you know, that prospect of very rapid growth is at least something that should be very rapid progress should at least be on the table, where again this survey that just came out if you ask leading machine learning researchers though i'll caveat i don't think they're really the the people that one should think of as experts here but they're
Starting point is 01:52:53 still kind of relevant people to consider they think it's about even that that will happen like about 50 50 chance that there will be something akin to what is known as the intelligence explosion that doesn't necessarily mean the kind of classic it's two days but where we move from that there will be something akin to what is known as the intelligence explosion. That doesn't necessarily mean the kind of classic, it's two days where we move from no advanced AI at all to AI that's more powerful than all the rest of the world combined. Perhaps it occurs over the course of years, or even like a small number of decades. But nonetheless, if things are moving a lot faster than we're used to, I think that increases the risks greatly. Well, it seems like things are already moving much faster than we're used to, I think that increases the risks greatly. Well, it seems like things are already moving much faster than we're used to. It's hard to see where we get off this ride because the pace of change is noticeably
Starting point is 01:53:35 different. I'm not a full Kurtzweilian here, but it does seem like his claim about accelerating technological change is fairly palpable now. I mean, I guess there are sectors where virtually nothing is happening or improving. I mean, that is also true. But when you're talking about information technology and what it does to politics and culture. The recent years have been kind of a blizzard of change, and it's hard to know where the brake is to pull. Yeah, and people should just take a look at the recent developments in machine learning, where they're not very famous yet, but they will be soon,
Starting point is 01:54:27 and it's truly remarkable what's happening. I mean, obviously it's still far away from human-level artificial intelligence. I'm not claiming it's not. But when forecasts have been done, again, of people working in the area, of what sort of progress we expect to make, things are moving faster,
Starting point is 01:54:44 even than the kind of forecast of people who thought it was expect to make, things are moving faster, even than the kind of the forecast of people who thought it was going to be moving fast. So in particular, a recent, they're called language models. It's where they just get trained on enormous amounts of, enormous corpuses of human text. But one got trained to solve, do math proofs, essentially, at about college level, kind of easy college level. And the innovation behind this model
Starting point is 01:55:09 was very, very small indeed. They actually just cleaned up the data so that it really could read formula online, whereas that was a problem beforehand. And it just blew everyone's predictions out of the water. And so now you have this model where you say, you give it a mathematical problem, kind of early college level, prove that so and so, or like, what's the answer to this question of so sure proof, and it will do it with 50% accuracy.
Starting point is 01:55:37 And that was something that people weren't expecting for years. If you look at like other language models as well, I mean, okay, I'll give an anecdote. So one of my responsibilities for the university is marking undergraduate exams. And so I asked GPT-3, not even the most advanced language model nowadays, but one you can get public access to. I was curious, how good is GPT-3, this language model, compared to Oxford undergraduates at philosophy who studied philosophy for three years? And so I got it to answer these questions, like some of the questions that the students most commonly answered. That's great. And it passed.
Starting point is 01:56:15 It was not terribly good. It was in the kind of bottom 10% of students, but not the bottom 2% of students. of students, but not the bottom 2% of students. I think if I'd given the essays to a different examiner, I don't think that examiner would have thought, oh, there's something really odd here. This is not a human writing it. They would have just marked it and thought like, not a very good student. You know, a student that's good in some ways has good structure. It like really knows the art, like the essay form, but sometimes gets like a little bit derailed by arguments that aren't actually relevant or something that's what it would have thought but would not have thought oh this is clearly nonsense this is an ai or something and that's
Starting point is 01:56:54 just you know that's pretty striking we did not have anything close to that just a few years ago and so i don't know anyone who's like this know, I think it's fine to be like an enormous AI skeptic and think that AGI is very far away. You know, maybe that view is correct. Like I say, there's an enormous amount of uncertainty, but at least take seriously, like, this is like a fast moving area of technology. Yeah. Yeah. And obviously there are many interesting and consequential questions that arise well short of AGI. I mean, just on this front, I mean, what will be the experience of all of us collectively to be in the presence of AI that passes the Turing test, right? It does not have to be comprehensive AGI. It doesn't have to be empowered to do things out in the world. But just when Siri or Alexa become indistinguishable from people at the level of their conversation and have access to the totality of human knowledge,
Starting point is 01:57:56 when does Siri on your phone become less like a glitchy assistant and more like an oracle. That could be fairly near term, even with blind spots and something that disqualifies her as proper AI. I think the Turing test will fall for most intents and purposes. For some people, the Turing test has already fallen. I mean, we just have this kind of laughable case of the Google engineer who thought that their current language model was sentient. It was absurd. Yeah. I mean, there was no reason to believe that, but I would be very surprised if it took more than a decade for the Turing test to fall for something like a chatbot. Yeah. I mean, I'm actually curious about why there haven't been attempts
Starting point is 01:58:46 at the Turing test using language models. My thought is that they actually would have a pretty good chance of passing it. So I'm unsure on that. But one thing you mentioned is, yeah, this idea of like, oh, maybe rather than Siri, you have this incredibly advanced personal assistant. It's kind of like an article.
Starting point is 01:59:02 I just want to emphasize, this is one of the examples where I talk about accelerating the safer parts of AI and maybe slowing down or not investing much in the more dangerous parts, where this idea of kind of Oracle AI, just something that is, it's not an agent trying to pursue goals in the world. Instead, it's just this input-output function. You put in text, it outputs text. And it is like the most helpful, extraordinarily smart and knowledgeable person that you've ever met. That's the technology we want to have earlier rather than later,
Starting point is 01:59:41 because that can help us with the scariest issues around alignment. That can help us align AI that is more like a kind of agent acting in the world. And some of the leading AI labs are taking exactly that approach, and I think that's the right way to go. Well, Will, it's always fascinating. It was a joy, thank you Sam so to be continued, best of luck with the book again, that is what we owe the future
Starting point is 02:00:11 and we have by no means covered all of its contents and until next time Will, thanks for being here thanks so much Sam Thank you.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.