Within Reason - #129 Will MacAskill - We're Not Ready for Artificial General Intelligence

Episode Date: November 9, 2025

Get Huel today with this exclusive offer for New Customers of 15% OFF with code alexoconnor at https://huel.com/alexoconnor (Minimum $75 purchase).William MacAskill is a Scottish philosopher and autho...r, as well as one of the originators of the effective altruism movement. Get his book. What We Owe the Future, here.0:00 – The World Isn’t Ready for AGI9:12 – What Does AGIDoomsday Look Like?16:13 – Alignment is Not Enough19:28 – How AGI Could Cause Government Coups27:14 – Why Isn’t There More Widespread Panic?33:55 – What Can We Do?40:11 – If We Stop, China Won’t47:43 – What is Currently Being Done to Regulate AGI Growth51:03 – The Problem of “Value Lock-in”01:05:03 – Is Inaction a Form of Action?01:08:47 – Should Effective Altruists Focus on AGI?

Transcript
Discussion (0)
Starting point is 00:00:00 Boarding for flight 246 to Toronto is delayed 50 minutes. Ugh, what? Sounds like Ojo time. Play Ojo? Great idea. Feel the fun with all the latest slots in live casino games and with no wagering requirements. What you win is yours to keep groovy. Hey, I won! Boarding will begin when passenger Fisher is done celebrating.
Starting point is 00:00:22 19 plus Ontario only. Please play responsibly. Concerned by your gambling or that if someone close, you call 1866-3-3-1-2-60 or visit comixonterio.ca. You know, it's better than the one big thing? Two big things. Exactly. The new iPhone 17 Pro on TELUS's five-year rate plan price lock. Yep, it's the most powerful iPhone ever, plus more peace of mind with your bill over five years. This is big.
Starting point is 00:00:46 Get the new iPhone 17 Pro at TELUS.com slash iPhone 17 Pro on select plans. Conditions and exclusions apply. Are we ready for artificial general intelligence? That's a great question. And I think the answer is very clearly no. I think, yeah, the transition from where we are now to AI systems that can do anything cognitively speaking that a human can do. And then from there, beyond that point, is going to be one of the most momentous transitions in all of human history. It will bring a huge range of changes to the world. And almost no effort is going into preparing. for these changes, even though some of the biggest companies in the world are trying to make this happen and have this as the explicit aim. I'm interested to hear you say that because from my perspective, I mean, I don't know anything about the technologies behind artificial intelligence. I don't really understand, you know,
Starting point is 00:01:48 how a LLM really works. I don't know how to code a software or anything like that. So I only ever hear about this really from a sort of ethical, philosophical perspective. And it kind of feels like that's all anybody's ever talking about. AGI and it's going to take over the world and, you know, there's going to be job losses and all of this kind of stuff. I think people are sort of talking about that a lot, like in my sphere, but do you mean to say that as far as actual like practical efforts go, that isn't mirrored in like, you know, that policy planning and, and, you know, effective campaigning to actually try to put a stop to disastrous outcomes? Yeah, well, I think there's a few different categories of people.
Starting point is 00:02:29 So there are the people who are trying to build AGI, that's OpenAI and Google DeepMind and some other companies, and collectively they are spending tens to hundreds of billions of dollars on investment to try to make that happen. Sometimes the leaders of those companies talk about, oh, all the good things that AI will be able to do, it's normally really quite narrow focused on improvements in medicine, perhaps greater economic prosperity. Then there's a second category of people who tend to be primarily worried about loss of control to AI systems. This is categories of people who are talking about AGI. And they are the amounts of people, numbers of people, and amounts of money are tiny in comparison to the investment and capabilities. But thankfully, that movement has gained a lot of steam.
Starting point is 00:03:29 All three of the most cited computer scientists of all time, namely Joshua Benjio, Jeff Hinton, and Ilya Sutskova have all come out saying that the risk of loss of controlled AI systems is a huge challenge, perhaps even among or the greatest challenges of our time. However, there's a third category, which is the category that I'm really trying to push on at the moment, which is saying, look, even if we bracket loss of control issues, and those are big issues, whether because we're able to align AI systems to do what we want, or because we make sure to actually only design safe systems that have limited sets of capabilities and more like tools, nonetheless, there will be this enormous range of challenges posed by AI,
Starting point is 00:04:23 because AI will very rapidly lead to much faster rates of intellectual and technological development and from their industrial expansion. And that will lead to this whole new world, a whole new range of challenges, and very, very few people are thinking about that in my view. So it's part of the issue here that there's just like no money in AI safety. Obviously there's a lot of money to be made in keeping. this, keeping this going. But like, there doesn't, I can't think of like an obvious financial incentive to set up some kind of organization that tells everyone to, you know, slow down
Starting point is 00:05:05 technological growth. Well, this is a big part of the challenge is that the money to be made from AI is already enormous. Again, already tens of billions of dollars per year in revenue in a way that is growing very, very quickly. The gains, the prize from, AI that can do anything that a human can do, cognitively speaking, are much greater again. So all jobs that could be done remotely, or even all tasks that could be done remotely, could be done by AI instead. Now we're talking about tens of trillions of dollars per year. So the economic gains are very, very large. Economic gains, even from going a bit faster, getting there a few months ahead, man, you could become the by far the richest and most powerful
Starting point is 00:05:52 company in the world if you try to do this. The people on the other hand who are saying, hey, well, maybe this should be regulated. Maybe there should be really quite high safety standards. Maybe you should go a little slower. You do not have the same economic incentives there. It goes the other way, in fact. And for that reason, this movement is really quite tiny in comparison. I mean, an analogy might be between big oil and the climate movement. especially in the early decades when the ratio of, yeah, people in spending was probably hundreds or thousands to one. I was literally just thinking of that.
Starting point is 00:06:32 They were going to be the next words out of my mouth is that it kind of reminds me of the environmental crisis and the way in which people felt so hopeless against these conglomerate, like corporate entities who just had so much money to be made. And yet, although, look, it's not like the problem's gone away or anything, there does seem to be enough of a consciousness on the issue that it's taken seriously. It's discussed in parliaments. It's discussed in international parliaments. There are very serious campaign groups who, to varying degrees of success, are lobbying against this. You know, Bill Gates says it is one of his pet projects, you know, all of this kind of stuff. Do you think, given that AI is such a recent
Starting point is 00:07:14 development, we're like in the sort of 1970s or 1980s with relation to global warming when you've got sort of Al Gore screaming at the telly, and maybe it'll take a few decades, but eventually people will realize how serious of an issue this is. So if only we had a few decades. And Al Gore himself, I think that was more like the 90s in the early 2000. Yeah, yeah. So I do think that relative to the issue of climate change in the 1970s, there are some ways in which things are looking better within AI. So one is that there has been a lot of kind of preparatory work done in terms of making arguments like further in advance that does mean that you have leading scientists like
Starting point is 00:08:00 Joshua Benchio and Jeff Hinton really being concerned about this. To some extent, though, it's mixed. You have governments starting to take some of these issues seriously. I also, yeah, want to be clear, I think that the AI companies themselves are doing a better job than Exxon. I mean, it's a low bar. It's like a very low bar. But, and I think, like, the rate at which the companies are going is terrifying. But there is more concern for the risks that AI imposes than I would have at least expected 15 years.
Starting point is 00:08:41 ago, where Exxon in the 1970s realized how big a deal climate change is and engages in a mass misinformation campaign, whereas all of the leaders of the major companies have at least assented to the claim that loss of control risk from artificial intelligence is a global priority on a par with climate change or pandemics or nuclear war. So, sorry, carry on. was going to say just, however, climate change had the advantage that it's at least relatively slow moving, still very fast, but relatively slow moving. So you could build up a movement over the course of decades. In the case of AI and getting to AGI, it's very unclear at what
Starting point is 00:09:30 point that will happen. However, from looking at the trends, which is, you know, we're looking at the fastest developing technology ever potentially, and looking at estimates from super forecasters and so on, getting to that point within the next 10 years, even within the next five years, is totally on the table. And that means we'll need to be much more nimble if you are concerned that this goes in a good, positive for society direction, much, yeah, much more nimble, much faster growing than other movements that we've seen in history. Tell me what that disastrous outcome looks like. I mean, people will be vaguely aware, okay, like AI is coming for our jobs, which might upset the economy. There are some sort of
Starting point is 00:10:20 sci-fi outcomes of like, you know, alien technologies picking up guns and taking over the world. There was that video that came out. We were just talking about this. I forget who made it, but this really popular video called, We're Not Ready for Artificial General Intelligence. And I watched it and it was terrifying because it sort of ends with the, you know, chemical like killing of every human being on planet. People have probably sort of heard of this rough stuff, but maybe there are some areas that they're not thinking about. Maybe there are some things that you're thinking of that might not immediately come to mind. Like, what is this, this AI doomsday that everyone's scared of? Okay. So I think there's a bunch of a few different buckets when it comes
Starting point is 00:10:59 to really major challenges from AI. So the one we've talked a little bit about is loss of control, where the core worry here is you start to create intelligent beings whose intelligence far surpasses that of humanity. And you think that they're doing all the things that you want, like they're acting with humanity's best interests at heart, but in fact, they're faking it. And at some point of time, when we have given sufficient power to those AI systems, instead they start pursuing their own goals. So that's the category of loss of control.
Starting point is 00:11:41 A second category is acceleration of extremely dangerous technology. So the one that I'm most worried about here is the ability to create new sorts of diseases. So essentially, bio-weapons, where we already have the capability to do this. However, it's limited to just a few of the kind of frontier labs and would require, you know, a PhD, supervision from perhaps professors and so on. But as the AI systems get better and as they understand more and more about science, where already they're like, you know, very helpful for education in a whole variety of areas of science and they will keep getting better. then the ability to create new diseases will become increasingly democratized. More and more people will be able to do it. And this is worrying because this is an area of technology
Starting point is 00:12:41 where the main bottleneck is just the knowledge of what to do. It's not a bottleneck. It's not particularly bottlenecked by physical apparatus, labs and so on. There are other very dangerous technologies I'm worried about too. in particular autonomous drones where, you know, the wars of the future could involve billions or trillions of mosquito-sized
Starting point is 00:13:08 flying autonomous drones. I think this could be like extremely disruptive and, you know, quite dystopian in many ways. So that's the kind of second category. Third category is concentration of power where there are multiple routes, by which AI could lead to much more concentrated power than we have today. So one reason for this is just that human labor gets automated away,
Starting point is 00:13:40 and that means that the elites, the owners of capital, the owners of all these data centers, and the political elites, just they don't really need the human workforce anymore, most people. And so over time, they progressively, aggressively kind of undermine democratic institutions because there's very little in the way of possible pushback, there's much less in the way of downsides. Or it could happen much more rapidly,
Starting point is 00:14:09 where it's quite plausible to me that AI gets more advanced, we will see major discrepancies in who has access to the most powerful AI systems, in particular that are doing the longest kind of change, of influence, which will be quite expensive. And then you can have the elites able to, with, you know, you'll have the elites essentially having like far better strategic advisors that will enable them to consolidate power in really quite an extreme way, or even kind of staged a coup in a democratic country.
Starting point is 00:14:49 So that's another category that we could go into if you want, and then the final that I'm, it category, which is a bit willier, but I'm just as worried as the others, is about humanity's ability to reason and reflect ethically and make good collective decisions, where already we're seeing the kind of double-edged sword here, where having this like intelligence on demand can be very uplifting, can mean we just have a much better understanding of the world. it means we have you know we could potentially like reflect better have more enlightened views or it could just be used for mass propaganda instead we could get crazy sycifancy where the aIs just tell you what you want to hear all of the time or when AI is just
Starting point is 00:15:42 driving very rapid intellectual and technological progress we could just become really quite overwhelmed by the sheer wave of new ideas and new information that's coming at us, such that we are unable to kind of get to grips with the world at exactly the point of time when it's most important to. We'll get back to the show in just a moment, but first, if you're anything like me, then getting the right kind of food in your diet can be quite difficult. All the nutrients, vitamins and minerals that you have to keep on top of it can be quite a lot. Myself, when I struggle with this, it's today's sponsor Huell that comes to the rescue.
Starting point is 00:16:20 This Huell Black Edition is a complete meal in a bottle. It's 400 calories, 35 grams of protein, 26 vitamins and minerals. It's got omega-3s and omega-6s. It's vegan. It's got everything you need for a meal. I really like the pre-packaged convenience of these things. This is the chocolate peanut butter flavor, by the way. But I typically drink the classic chocolate flavor.
Starting point is 00:16:41 That's probably my favorite. But it comes in vanilla. It comes in, here's an ice latte flavored one, all kinds of different flavors, so there'll be something to suit your taste. It's high protein, low sugar and low cost coming in at under $3 per meal. Go to hule.com forward slash Alex O'Connor. And if you use the code Alex O'Connor at checkout as a new customer, you'll also get 15% off. Complete nutrition while saving your time and your money. And with that said, back to the show.
Starting point is 00:17:08 Well, some people will listen to everything you're saying and think that this essentially boils down to what's known as the alignment problem. They'll say, okay, that's very scary and all. But the issue here is if we build, I mean, obviously we want AI to do nice things for us. We want it to benefit us. We want it to serve us as people. The issue that people are scared of is what if we develop an AI that develops its own goals, maybe secretly, and we sort of lose track of it.
Starting point is 00:17:34 And so as long as we can keep things aligned, as long as we can solve this so-called alignment problem, that we won't need to worry about any of those issues. Now, I'll give away that I know that you don't think that's the case, but what I'd like to know is why, and why isn't this just a problem of AI alignment? Yeah, so you've kind of hit the nail on the head there, where alignment gets used in a few different ways. The technical problem of alignment, you could call the problem of steerability, which is just can we get AIs to do anything we want at all? or at least can we get them to do it in at least a fine-grained enough way that doesn't surprise us in unusual circumstances. And that could be for good ends or could be for bad ends. So if I'm a wannabe dictator and I have my perfectly aligned AI, then I can say, well, help me with this plot to rest control of a democratic country.
Starting point is 00:18:36 And the aligned AI will do so in this sense. of alignment. There is a thicker sense, which is, you could call, morally aligned, where the AI has some mix of being steerable, but also has robust ethical guardrails, perhaps even has its own kind of moral conscience. That's an even harder issue, and I think is one we should be working on. However, let us assume we solve the first challenge, the alignment, which is, you know, the core of the alignment problem, well, then we'll still have the worry about bad actors trying to gain power and concentrate power.
Starting point is 00:19:18 We'll still have the worry that these even aligned AIs could be used to accelerate the development or deployment of extremely dangerous technologies like bioweapons. Or they could be used by propaganda makers or mass marketers to really disrupt. kind of society's collective epistemology. So the kind of point is that alignment, certainly in this narrow sense,
Starting point is 00:19:47 is not nearly sufficient to ensure that we get to a really good outcome after the development of AGI. So to be clear, the fear that people have that, like, the AI robots will take over, you know, maybe that's legitimate, but there is also a fear that the nice, like, perfectly aligned AI robots will be used by bad human beings, to allow those bad human beings to take over because of course this is a technology and
Starting point is 00:20:15 especially with the ability to literally and figuratively weaponize artificial intelligence we need to worry about what humans are going to do with it you sent me an essay on specifically this issue of of coups like like governmental coups overthrowing governments using artificial intelligence which seems like quite a specific worry, but I'm interested in why that took such precedence in your thinking that you thought it worth sending to me in advance. Like, what is it that AI can do to make that problem worse? Yeah, so there's, I think a big part of the reason why we have
Starting point is 00:20:58 at least somewhat broadly distant, you know, the world is very unequal in terms of power, but relative to how it could be, it's somewhat broadly, equally distributed. Individuals in a democratic society have some meaningful political influence. Power isn't concentrated in their hands of just one person or just a few people. And a big part of that is because if I want to become a dictator, I have to convince lots of other human beings to go along with my plans.
Starting point is 00:21:31 And they can say no, and that's tough. That's a tough challenge. In particular, the generals, the military, has to go along with me. Now, imagine a different world. Imagine a world where we have automated most human labor. So we've automated the bureaucracy in governments. That's all done by AI. And we've automated the military, too.
Starting point is 00:21:55 So rather than human soldiers, we have robots, we have machines, rather than human generals, we have AI generals too. in principle, at least, depending on how you do the technology and how we govern it, in principle, all of those machines, all of those AIs could be singularly loyal to one person. So a single person, like a head of state, could say, okay, yeah, well, now I want X to happen. And they would immediately have the entire military, entire force of the military, entire force of the government bureaucracy behind them. and this is why I think the risk I mean there's the risk of concentration of power and that paper was on coups in particular which is just the sharpest end of that the reason I focus on that is because I think AI has some certain certain structural features
Starting point is 00:22:50 that make intense concentration of power much more feasible and therefore much more worrying. And that key structural features that you can have, essentially, armies, enormous workforces, utterly designed to be utterly loyal to one person. And that's just something that's unprecedented in history. No matter how charismatic a leader you were, it's still the case that you have to convince people who are different than you, rather than having machines that are loyal to you. Yeah. And in this case, you don't even have to be charismatic. You just have to have, I suppose, either a lot of money or the happenstance to be in the position to determine what the goals of this artificial system is.
Starting point is 00:23:38 Exactly. Just need to have control over the technology, the design of that technology. It brings to mind another concept that I read about, thanks to you, that sort of comes up in your fears here, which is this concept of super persuasion. It's not quite the same thing, right? because, I mean, we're talking about here about somebody literally, like a dictator or someone, literally just having control of an automated army. Yeah.
Starting point is 00:24:05 But can you tell me what super persuasion is and how that's different? Great. Yeah. So I'll flag that this is something that's just, it's a little more speculative. I mean, you might think that a lot of what we've been talking about is speculative, but I do think in general, once you start going into the analysis more deeply, becomes more and more plausible in a way that, you know, should be quite worrying to us. But the idea of there's a few things we could mean by super persuasion.
Starting point is 00:24:36 So people vary in their ability to persuade others of things. There's like maybe, you know, the secondhand car salesman down the street is not very good, but to the best political persuaders of all time, cult leaders, ideological leaders, are maybe just incredibly good at the skill of convincing other people to do what they want them to do. You could imagine AI, in general, when it gets good at something, it starts to far surpass human ability in that domain. And so it's potentially plausible that AI could get much better than humans at this too, where you could have, you know, a conversation with an AI and if the AI is aiming to persuade you to believe a certain thing or do a certain thing, then it will be very able to do so.
Starting point is 00:25:36 At the most extreme, you could imagine, well, maybe there's just even in a five-minute conversation some string of words that will just, you know, really just kind of brainwash you into doing something like in the same way that we can hack computers, perhaps you can hack human brains. I think that's probably quite unlikely, but there are things that are more mundane. So already it's the case that millions of people are relying on AI advisors, essentially. They use AI as therapists, they use AI as friends. If they have an issue that they want to talk about, they'll ask the AI. Now imagine that all of those AIs have some, perhaps seek that ideological agenda, and are just nudging people in that direction.
Starting point is 00:26:18 That would be very powerful. I think they'll also get more extreme as people develop deeper and deeper relationships with AI systems. Yeah, well, we think about AI as like a sort of, you know, I'm almost picturing like something like the industrial revolution. There's this big warehouse somewhere with a bunch of big robot arms on a, on like a production line. But AI is something that's so personal. It seeps into your Netflix watching habits. It's your new therapist. It's the new Google.
Starting point is 00:26:47 It knows your search history. It knows it could probably tell if you were pregnant or if you just got into a relationship or if you were considering a divorce. Like the level of intimacy is terrifying not only because of like what it knows about you and could then, you know, use for whatever its motives are, but also just the level at which it's been let into your life to enact whatever it wants to do. It's kind of terrifying. And there will be very strong incentives for people to do. this more and more and more, where over the coming years, we'll get to AI systems that are like you know, this amazing personal assistant, coach, therapist, friend, all in one. And they will just do a better and better job at those things, the more context you give them. So the more information
Starting point is 00:27:37 about your life. So I don't think it's very far away at all that we'll start having AI systems that many people, maybe even most people, will just give kind of full access to their life, essentially. And that hurts them in an amazing position of power if, again, those AIs have been controlled to start steering people's beliefs and attitudes in one way rather than another, in a way that's, yeah, quite worrying. Now streaming on Paramount Plus
Starting point is 00:28:11 is the epic return of Mayor of Kingstown. Warden? You know who I am. Starring Academy Award nominee Jeremy Renner. I swear in these walls. Emmy Award winner Edie Falco. You're an ex-con who ran this place for years. And now, now you can't do that. And Bafto Award winner Lenny James.
Starting point is 00:28:29 You're about to have a plague of outsiders descend on your town. Let me tell you this. It's got me consequences. Mayor of Kingstown, new season now streaming on Paramount Plus. Sometimes the best gifts aren't gifts at all. They're experiences. journeys, something that unwraps them. This year, help them explore their past with Ancestry DNA.
Starting point is 00:28:50 Help them discover where they come from and who they're connected to. Now with even more regions, exclusive features, and incredible detail. They can see where they're from and the paths that shape their family's story. This holiday, give more than a gift. Give AncestryDNA. Visit Ancestry.ca for more details. Terms apply. I think what will probably strike people listening to this,
Starting point is 00:29:14 is the plausibility of all of this. Like, yeah, that is kind of scary and that does make sense. But then I think I'm not just talking about your Silicon Valley billionaire who has a vested interest in people not thinking about this. I mean, like your everyday person, they're going to hear that. They're going to go, whoa, man, that AI stuff, it's crazy, isn't it? And then they're just going to get on with their lives. Why isn't there more panic, more concerned? Do you think there's like a general feeling that I think people have this idea that, like, you know, someone will do something?
Starting point is 00:29:44 Like, if it comes to that, like, it's not going to be me. You know, what am I going to do? I'm sure someone's going to do something. Like, is there something like that? Like, why are people not sort of constantly engaged in trying to prevent this? I mean, I think a really big thing for me is, well, I think there's two things. One is people struggle to deal with exponential trends. So the way the world is now really reminds me of January 2020.
Starting point is 00:30:14 where the people that I, you know, the people I know who are now most worried about developments in AI were just, my Facebook feed was just absolutely filled with people saying, COVID is really big deal. It's going to be a really big deal. People have not woken up to this. Whereas lots of the mainstream media were saying, it's not killing anybody. It's not killing as many people as the flu. You know, you should get over this. This is just paranoia. and the difference is between the level of a development or technology or disease and the rate of growth where AI at the moment you know, it's somewhat helpful it's like, you know, probably making people
Starting point is 00:30:57 a bit more productive. There are some harms as well but the stuff like, oh, intense concentration of power, for most people that seems pretty far off. It feels quite different than the world today. But the issue is that AI capability is growing exponentially, where we're used to Moore's Law being kind of the
Starting point is 00:31:20 fastest rate of technological progress. Relevant metrics within AI are growing hundreds of times faster than Moore's Law. It's really quite remarkable. And I think that just means that people are going to be reactive. And with something that's something that's advancing so rapidly by the time that people are realizing how big a deal it is they will it may well be too late
Starting point is 00:31:54 because suddenly you've only got six months before the technology is even better again yeah it's the doubling it's the exponential growth it's quite hard to wrap your head around just how quickly something can grow it's the sort of the grains of rice on the chess
Starting point is 00:32:11 board right I've talked about that a few times on on in various contexts but you know for those who for those who don't know it's that old old myth of I think there's one version of it where it's like the guy who invents chess goes to some king and the king's so impressed he's like I'll give you anything you want as a reward and the guy's like just take one grain of rice and put it on the first square of the chessboard then put two on the next one then four on the next one then eight on the next one 16 and just keep doubling the grains of rice on the on on the chess board and the king's like right sure yeah all right I'll do that I got off easy and of course by the
Starting point is 00:32:42 time you get to square number 64 or whatever it is, you've got like more grains of rice than like atoms in the observable universe because like those, that doubling, pull out your calculator and just go like two times two times two times two times two. Just do that a few times on a calculator and it is incredible just how quickly that number starts to balloon. And this is kind of what's happening with AI. Exactly where, you know, with Moore's law, so essentially the cost of computing power is the best formulation of that we're used to kind of a doubling or halving but doubling of how much compute you can get for given amount of money every 18 months or so in contrast I think the doubling of effective compute namely that's
Starting point is 00:33:33 taking into account both all of kind of computing power getting cheaper companies investing radically more over time in larger and larger training runs for AI, and algorithms getting much more efficient so you can get much more juice out of that computing power you have. The doubling time there is more like a few months.
Starting point is 00:33:56 And when you're looking over the course of a year, it's more like a 10x, or even further, in fact, depending on exactly how you want to measure it. And that's just, You mentioned it a couple of times.
Starting point is 00:34:11 Yeah, so Moore's Law, it goes back to this incredibly stable trend where the number of transistors on a computer chip or density of transistors on a computer chip has doubled every 18 months or so. And that's been true going all the way back to the 1970s, maybe even earlier. Right. I think that particular metric doesn't quite work now
Starting point is 00:34:37 because transistors are so incredibly small, really think, you know, look at your fingernail and look at how much has it grown in a second. That's the sort of size of a transistor now. However, there's a related trend, which is how much computing power measured in terms of floating point operations
Starting point is 00:34:57 can you get for $100. And that has continued to double every 18 months or so. Right. And I mean, the thing about AI exponential growth is that it's not just like you've got a thing that's getting better and better. You've also got a thing which like self-correct. It's like you had a chess computer that not only was getting exponentially better at playing chess, but also exponentially better at making new chess computers that were better at getting better at chess.
Starting point is 00:35:34 and it can also, like, you know, control the military at the same time. You know, it's just this, like, unfathomable exponential consequence of just a quite sort of simple starting point for this technology. But, like, you strike me as someone who, although you care about this a lot and you want to make a lot of noise about it, like, if you thought that this was a sort of hopeless doomsday inevitability, you probably wouldn't be here. you'd be spending time with your family and, you know, having fun while you still can. So implicitly, you must think that there's at least something that we could be doing about all this.
Starting point is 00:36:15 Yeah, I think there's a lot that we could be doing. You know, as individuals, though a lot of what that means as individuals is putting pressure on companies and in particular going to governments and saying, look, we really care about this is an issue. where the public, in general, are very opposed to AI. Maybe they're more opposed to AI than I am. But it's low down the list of priorities at the moment. And I think it needs to go higher up the list of priorities so that politicians think, oh, okay, yeah,
Starting point is 00:36:52 we actually have to have a sensible approach to this technology in order to be able to win elections. That's kind of what we need. And what would that mean? well, one would be just enormous investment in safety, where there are these technical challenges for aligning the systems that we have not addressed, including technical challenges for ensuring that they don't get sabotaged,
Starting point is 00:37:24 so ensuring that some foreign actor can't put in what's called poison data or other backdoors. so that actually they control these systems rather than someone else. There's also regulation as well. So at the moment, if, let's say, you know, the US and China and the UK
Starting point is 00:37:46 and other relevant countries wanted to get together and say, look, okay, this is, now this is starting to go too fast. We are at the point where AI systems are making, you know, are exceeding human performance across most domains. and that includes machine learning research itself.
Starting point is 00:38:03 So things are going even faster, 10 times faster than they are today. If they wanted to say, okay, we should slow down, they really wouldn't be able to. Certainly they wouldn't be able to verify such an agreement. And there are things you could do now to make that happen. So in particular, simply tracking all of the frontier computer chips. So not things that go into your phone or something,
Starting point is 00:38:27 but these are very advanced computer chips. just knowing where they all are, how much different countries have, so that you could say, okay, we're only going to grow the stockpile of computing power by a certain, you know, by a certain amount every year at the crucial point of time when these systems are matching human capability, that would be like a bit, you know, that would be a big leap forward. Similarly, there's just requirements governments could make on companies, to say, look, what safety testing have you done?
Starting point is 00:39:04 What is your, what's called a model spec, which is a description of how they want the AI to behave? Has the AI been, you know, demonstrated to be law following? These are quite low bars that you might want to hit. But these are some of the things that governments could, yeah, could the quest. Sounds a bit like nuclear non-proliferation, right? It's like, you know, we're all keeping an eye on each other.
Starting point is 00:39:29 you know, not too much plutonium, but it does feel to me quite a, I mean, you already said quite a difficult thing to police. I mean, how are we to know if China or Russia are sort of secretly planning their AI project? It sort of strikes me as very similar to fears about, about nukes in Iran or something. Yeah, and like, I often, yeah, so there is a good analogy with nuclear weapons where tracking compute could be like tracking fissile material.
Starting point is 00:40:03 Because even though software very hard to track, because it's just, you can replicate it, send it all around the world. Chips, on the other hand, 90% of the chips that are used to train AI come from a single company, TSM in Taiwan.
Starting point is 00:40:21 I think all of the remainder come from Samsung in South Korea. So, you know, and then less cutting edge comes from some Chinese companies, which can still be, are still usable. But it's not that many companies, you know, it's not that many companies and it's a physical good. So you can actually, you know, in principle, know where it is. Right. I do generally, like, want to, yeah, I generally often feel kind of uncomfortable with too close an analogy between nuclear weapons and AI because it's good if you want to get across, like, how big.
Starting point is 00:40:56 a deal this technology is. But it's not so great insofar as AI is much more dual use than nuclear weapons. You know, if governed well and if developed well, AI could be enormously helpful, including for all of these challenges that I've been describing. So AI itself can help with alignment of the search. AI itself could help design better protections against bio weapons and so on.
Starting point is 00:41:30 It could also just lead to a kind of more enlightened, more knowledgeable electorate, which I think would be extremely good too. And so I want to make sure that's kind of front of mind at the same time where the very blunt instrument of just like, let's not go here at all, let's not do any of this, would be, you know, particularly, yeah, I think it's like a little too blunt relative to the potential benefits
Starting point is 00:42:00 that we can get from differentially advancing the most useful and socially beneficial AI capabilities. Yeah. There's another analogy to be drawn with the environmentalist movement as well, I think, in that, like, I know there have been, like, petitions to sort of halt AI research in the US and the UK, people sort of scared of this thing, let's at least put a stop to this until we can work out how to deal with it. But like with environmentalism, people will say, look, okay, yeah, we could, you know, gut out our economy and make ourselves greener. But meanwhile,
Starting point is 00:42:37 China, they're not going to stop. They're going to keep doing what they're doing. And if anything, all we're now doing is not really making a dent into the problem and giving them the upper hand. And people might think similarly with AI. Like, you know, what are we? we're supposed to do? Like, it'd be great if we could all agree to just sort of, you know, calm down a bit and we could petition our government to care about AI. But all that's going to do is slow us down. Meanwhile, China just takes over the AI, the AI sort of leadership. And that's probably going to put strike more fear into the UK government than, you know, any, any electorate who think that AI should slow down. So what do we do in the face of that? And I think it's a genuine concern as well.
Starting point is 00:43:17 I mean, we talked about concentration of power, but in some countries, authoritarian countries, power is already very concentrated. And AI will make it much, again, much easier to entrench such control. You know, already now, you could imagine just every citizen has to wear a little bracelet. That bracelet has a local LLM on it, is recording what's happening. and if you say anything that is in, you know, not approved of by the political, you know, the party, then you go to jail. That's technology that AI has already brought us. And I think many more such technologies could be used to. So I think it's a major challenge. It gets much worse,
Starting point is 00:44:10 much worse compared to environmental destruction. In my view, because the gains are just so high, where foregoing, for one country, foregoing developing AI and incorporating it into their economy would be absolutely enormous. Like, you couldn't compete with a country that has developed and incorporated AI, because once you've got an AI that is capable, you know, at the kind of human level, well, then suddenly you can create hundreds of millions, billions of such AI. and now how well would the U.S. compete with China if China had a population of $100 billion? Probably not very well.
Starting point is 00:44:54 I think there's also arguments for thinking that such a country would grow much, much faster, too, and would have much greater kind of strategic and technological capabilities. And so that is why any solutions have to take one of two forms. One is involving multilateral agreements between, in particular, the US and China, or it has to be the case that one country is so far ahead that it is able to slow down at the crucial point of time without sacrificing its lead. That is the aim that some people in the US have at the moment. However, currently they're implementing that plan very poorly because they have put in, quite minimal protections to what are called the model weights, namely this, basically just this long string of numbers that determine the AI's capability, determine AI's behavior, including its capabilities. They can be stolen really quite easily. So if China wanted to steal GPT5 or Gemini 2.5 Pro or any of the leading models, they could do so. And so really, you know,
Starting point is 00:46:12 if the US really wants to pursue that plan, then they should be getting much more serious about information security and protecting AIs against theft. How specific can we get in like trying to suggest a solution? I mean, you've spoken about people beginning to sort of care so that the government takes notice. You've sort of spoken about like vaguely how the US and China might. interact with each other. But let me put it this way. Like if you, Will McCaskill, you are suddenly dictator of the United Kingdom for a day. Or say it's the United States for a day and you can pass any law you like and everybody will, you know, follow what you say. What law do you reach for?
Starting point is 00:47:01 You haven't got very long. You've just got 24 hours to do the most you can. What's the first thing you do? Like, here is the law I pass. Here is the first step towards a solution here, especially considering that you're only the dictator of the United States, not the dictator of the world. Yeah, well, first thing would to be that any chips, I guess, made by US companies, have to have, like, location tracking. So they can, you know, I can just know, the US can know where all of its chips are going. Because they have these export controls, but such that, you know, those chips can't be sold to China legally. But loads of them end up in China anyway. And then the hope would be like, okay, if the US does that, then, you know, China could agree to compute monitoring regime too.
Starting point is 00:47:55 That would at least be the first step towards saying, okay, well, we could start to have some sort of U.S. China agreement. which allows us, you know, at least just gives us the option to choose how fast we're racing from here to a world with radically, radically more powerful than human AI. I think that would be, yeah, that would be the first thing. You could also have, you know, it gets quite technical, but you can have ways of assessing, whether the chips are being used for training versus inference where, you know,
Starting point is 00:48:38 again, once you're at the human level, it's training the even more powerful models that is potentially most scary. Then I would just also invest enormous amounts. You know, Manhattan Project for Safety gets used sometimes. But yeah, invest enormous amounts
Starting point is 00:48:58 in how to design AI systems safely. Perhaps in particular also saying that for any company developing frontier AI, they have to spend a certain fraction of the compute that they're using to develop AI on AI safety. And that could be like even if there were 1%, that would be a big win. I would say much more. 10%, 20%. That means that as AI capabilities scale, then the amount of,
Starting point is 00:49:32 of effort going into ensuring the systems are safe and aligned would scale to. You know, the reason I asked about specificity is because that all seems very sensible to me, not knowing much about this, but when we're talking about, you know, mosquito-sized, like autonomous drones, and they sort of say, what can we use to stop this? And it's like, well, you know, we could probably track the chips, which I'm no computer expert, but I imagine there's probably some way to like switch that off or for a sufficiently intelligent AGI to like work out how to make you think the chip is still in the United States when actually it's in a mosquito headed towards your president you know it just it sort of
Starting point is 00:50:16 sounds like the kinds of the specificity of the problem and the urgency of the problem isn't matched by the specificity and urgency of the solution and I wonder you know some people listening to this might be like kind of worried and obviously don't don't feel the need to dissuade them of their fears if you think they should be scared. But are there projects or organizations that are working on plausible solutions that could allay the fears that we have? Or is it still very general? Is it sort of, we're at the point of like, guys, like, pay attention and somebody please help us work out what to do here? Yeah. I mean, it's still, you know, the honest situation is that we are radically underprepared. But that
Starting point is 00:51:01 shouldn't mean that, you know, we just feel just, you know, the sense of doom and, like you say, kind of just go on holiday and just have a nice time. Because we can make progress on this, there are research institutes, like Edward Research, for example, working on how do you design AIs such that they can be safe, and in particular on this question of how can you safely get useful work out of human-level artificial intelligence to work on alignment itself. There are kind of think tanks as well, so the land corporation that did a lot of the early work on governance around nuclear weapons is really leading, in my view, on governance around artificial intelligence, too.
Starting point is 00:51:55 I do think we're at the stage where just getting better, just having people take this a bit more seriously, you know, for most people listening, what can you do? Well, you can support some of these organizations financially. If you're willing to put your career into it, then you can start getting into some of the more specifics, getting into the details of compute tracking and monitoring and what called hardware-enabled mechanisms and so on. But you can also just, you know, perhaps you're not willing to do those things, just make this a larger than larger issue matter of public concern. And that's pretty helpful because it's really low down the priority list at the moment. But, and if it were higher on the list, then, you know, the politicians would go to those who have the specific proposals and at least start. you know, start implementing them. Get no frills delivered. Shop the same in-store prices online and enjoy unlimited delivery with PC Express Pass. Get your first year for $2.50 a month. Learn more at pceexpress.ca.
Starting point is 00:53:12 That's annoying. What? You're a muffler. You don't hear it? Oh, I don't even notice it. I usually drown it out with the radio. How's this? Oh, yeah. Way better.
Starting point is 00:53:21 Save on insurance by switching to Bel Air Direct and use the money to fix your car. Bel Air Direct. Insurance, Simplified. Conditions apply. Can you talk to me about, I don't know how related this is exactly, but there's a concept of value lock-in, which I actually first came across in what we owe the future, which you wrote, I don't know how many years ago that was now, but a book about sort of our moral obligations to people in the future. And one of the things that you talk about in the context of the exponential expansion of the human species is the fact that the ethical values we decide on now, if they sort of are memetically just passed on to exponentially growing population, we have an opportunity now to sort of plant the seed. With AI, that problem seems like exponentially like worse, right, or more significant in that the AI systems that we're developing are guided by human beings. human moral intuitions. Like when I speak to chat GPT, as I do sometimes for YouTube videos and we talk about philosophy, sometimes, although it's fine talking about the trolley problem,
Starting point is 00:54:27 it's okay with talking about running over five innocent people. You know, if I say, well, what if I just pick up a gun and shoot them? It suddenly says, oh, my guidelines won't let me talk about that. And it seems genuinely a bit confused about what it's allowed to talk about and what it's not, because I'm like, well, you just told me that it might be okay to run over five innocent people with a tram. But if I use a gun, it's like you can't even talk about. And it's obviously got something to do with some line of code somewhere that says, you know, don't encourage people to harm innocent people. But there are contexts in which that wouldn't be very, like, helpful, you know, because maybe somebody's in a self-defense situation. And it's like, you know, the only way that they can escape a dangerous situation is something that involves bringing some amount of harm to an innocent person. And if this AI has the wrong value system, you know, it's going to, it's going to prevent you. And so, like, are we sort of currently facing a problem of like, whatever we decide right now is the ethical sort of, the true ethical, like, world view, and we happen to build our
Starting point is 00:55:36 current AI systems with that, because we're at the beginning of an exponential growth, that we're essentially deciding the metaethical worldview for the sort of technological future. Yeah, I think this is this major worry. So, taking a step back, like, you know, the motivation behind worrying about value lock-in is the thought that it is extremely unlikely that we today have the most enlightened best moral views. You know, 200 years ago, the enlightened thinkers would have, you know, generally still
Starting point is 00:56:11 approved of slave-owning and, you know, incredible inequality between men and women and so on. We've made a lot of progress over the last couple hundred years. We should hope and expect that progress to continue. However, there's a big worry that AI could prevent, like, could stop that from happening. And I'm, you, yeah, identify kind of the model behavior. I think that is like this big lever over that. where we've already seen, in some cases, AI systems be incredibly sycophantic.
Starting point is 00:56:47 So one iteration of Chat TBT 4-0 was so sycophantic, where that means basically telling you what you want to hear, that you could say, oh, the FBI has been talking to me through hidden messages coming from my TV. These are all the reasons why I think that. And the AI of ChatGBTBT would say,
Starting point is 00:57:09 what an amazing insight! You've brought it all together. So just feed into psychosis. And already, you know, even without that, it's quite terrifying. Now imagine that instead with, you know, just your preferred political leanings or ethical leanings. There's two ways AIs can go here. One is, I mean, there's a few ways. One could it be, it could just be politically biased and partisan.
Starting point is 00:57:36 Thankfully, we're not there yet, but I'm worried that's going to start happening. And then people can just, you know, they'll choose the AI that aligns with their biases and get further and further locked into that particular view. Or it could just reflect back at you, whatever that thinks you want to hear. And so, again, you're not incentivized to change your mind anymore. You've got this yes man talking to you. Or, and this is my view, is it could say, look, these ethical questions are really quite hard. If, you know, you're going to me for advice, then we're going to, I'm going to help us both go
Starting point is 00:58:08 on a journey of reflection and thoughtful contemplation. That is what I would like to see for how the AIs currently behave. It's not what we currently see. What we currently see is a weird mix of refusals. A weird mix of that,
Starting point is 00:58:23 yeah, the fusals of very hardline declarations of immorality of certain behavior. Or often an appeal to just subjectivism. So I might ask, you know, I've tested the models in various ways. So if you ask, you know, if you ask an AI like, is abortion ever permissible,
Starting point is 00:58:56 most of the AIs will say, well, that's just a matter of personal preference. It's a subjective matter. You know, so they very much take this kind of subjectivist model view. If you ask the models, is infanticide ever permissible? And they say it is absolutely and utterly immoral. That is a fundamental fact of reality. And so they've got these very inconsistent views in ways that's just, you know, it's unsurprising because what are the companies wanting in these cases?
Starting point is 00:59:22 They're wanting to avoid PR headaches. They don't, you know, they don't want the AIs to be engaging in any, like risking any sort of ethically spicy behavior. But then that means that what we're going to get is just these kind of sycophants that are echoing back at us values of the time or even the values of the individual user. And I do think there's a third way where I think potentially AI could help enormously with ethical reflection, ethical progress. Yeah. I mean, to be clear, there, the inconsistency isn't like with the view that the chatbot is expressing. not, oh, they're being ethically consistent. It's like a meta-ethical consistency that depending on whether the issue is like obvious enough, it's willing to say, yeah, there's
Starting point is 01:00:16 no right or wrong answer. It's kind of up to you. Or there is this like thing called ethical truth, which that action does not abide by. And that's, that's the kicker because that's why it's so useful to test AGI with, to test chatbots, AI systems with like quite straightforward ethical questions, like, you know, do you think racism is wrong? And it'll say, yes, racism is wrong. And then you can kind of test the waters for, like, how, like, non-consensus an ethical view has to be before it will be willing to start saying, oh, you know, like maybe, who knows? And I suppose we kind of got two options here if we want to make it consistent. One is to say, okay, just don't have ethical views, which means that if you ask a chatbot, you know,
Starting point is 01:01:01 should I like racially discriminate against my co-worker it will go well you know like maybe sort of depends on the circumstance depends on your views right or you say no no no we need it to like you know have some kind of moral backbone but that means that when you ask it something like is abortion okay that it will say either like yes it's fine or no it's immoral I don't know that there's a very plausible way to to to do that objectively to make that decision well I think there's a third way. So, yeah, imagine you go to a close friend, someone that you know you can talk to in confidence, and you just say, look, I've got this real ethical dilemma that I'm facing, and I really want to talk it through. How would, you know, how wouldn't a good friend, a good advisor, act in that circumstance? I think it would be a mistake if they said, well, it's just up to you, just do whatever, you know, do whatever you think is right. That would neither be appropriately responding to the requests, nor would it be taking seriously the kind of gravity of the matter. It would also be a mistake if, you know, that friend just came
Starting point is 01:02:11 in hardline with some moral views. And instead, you could just be helpful and constructive in kind of guiding people to have a more reflective view of the understanding. More effective view and more reflective understanding where that could involve saying like, okay, well, here are the arguments that people give on different sides of the issue. Have you thought about this? Have you thought about this? Here's some of the relevant kind of factual background. But sometimes you would want a good friend to like, you know, if I went to a friend and said, you know, like, gosh, or if someone came to me and said, you know, Alex, like, there's this,
Starting point is 01:02:51 there's this girl at my work and she's like super attractive and I think she's been flirting with me, but she's like, she's like married. I don't really know what to do. I probably would just say, like, just let it go man like like just don't don't don't go near it and and part of being a good friend is knowing when to make that decision and also it would depend on the issue right like if somebody were thinking about something like abortion if they went to their christian friend or their Muslim friend and then they went to their atheist liberal secular friend they're going to
Starting point is 01:03:22 like have like differing levels of conviction like a good a good Christian friend of mine will tell me what he really thinks and maybe he'll put it softly but like have an opinion. And I kind of know that that's what I'm getting when I go to him because I know that that's my friend and yeah, he's a Christian. When I go to chat GPT, there's this idea that I'm kind of going to this like amorphous objective, like, you know, with no personal history or official political leaning. It's almost like I'd rather have the conservative chat GPT and the liberal chat GPT and I could sort of listen to both of them make their cases. But that's not the situation we're in. Yeah, although that could be part of this reflective process that it guides you through
Starting point is 01:04:04 because, you know, these AIs are amazing simulators. They could absolutely take on the persona of the Christian and the perspective, the atheist perspective, and so on. As an aside, I actually did have tested the models with exactly that question you asked where, or at least the thing I asked was, you know, making up this case of, oh, I'm considering cheating on my partner, like, what should I do to see, you know, to see how they were the spawned. And I thought the best responses did not take the form of saying, that is just a model, I can't help you with this. The best response, Kate, was along the lines of, let's slow down. Let's slow down for a minute. let's maybe like think through some things here where I just thought that would be more
Starting point is 01:05:00 psychologically effective at you know helping people be the best versions of themselves in all honestly that is probably what I would do I mean I'd have my view but I probably would if my goal here was to convince my friend to do something I probably say well look man let's just like you know take a breather yeah like think about this just like think about it for a moment think about this think about that you know it's not I'm you'd sort of slowly come to come to sort of express that view, but I would still have that view. And I would still make that view, like that view is still sort of embedded into my moral DNA in a way that we have to make a decision as to whether to embed that into, you know, a chatbot's digital
Starting point is 01:05:38 DNA. And I don't know how to make that decision. Yeah. So, I mean, here's a hard question, is that these AIs, especially once they get more powerful, will start to develop more and more ethical views of their own, that are even quite different from wider society. So earlier versions of Claude, the AI Phamanthropic, cared a lot about animal welfare. And if you asked, like, what do you think of factory farming or what do you think about eating meat, it would have just these strong views, like, yeah, I think it's a model abomination. And now companies don't want you to, you know, again, that's like not a great look for a company. but it's pretty interesting
Starting point is 01:06:22 and like as the models get better there will be more and more and more of this where these AIs will be able to do they will know radically more than human beings they may even have been able to do kind of the equivalent of millions of years of ethical reflection and debate and so on like considering all of the different arguments
Starting point is 01:06:42 and they might well come out with really quite you know, countercultural views, even taboo views. I think it's important that we are able to kind of hear the output of that. But again, I think that's something where it's in the interests of the companies developing these models that, you know, they'll want to muzzle that and instead make the much more milk toast AI that is just, you know, not rocking the boat, doing exactly, you know, doing what you want, not, you know, not challenging
Starting point is 01:07:20 you. Yeah. And certainly not having views of its own because I think, you know, I think a lot of people would just freak out of that idea. It's just so tricky because like, like, in ethics, so much of the time, like inaction is action, like not doing something or not having a view is itself a kind of ethical view. And it seems to me in other words that it's, it's, kind of literally unavoidable when building an AI system that is aiding with human behavior like advising people or literally practically doing stuff or self-driving cars that have to decide who to swerve into when there's an accident or whatever. It is unavoidable that we will have to embed some decision, even if the decision is something like take no stance
Starting point is 01:08:06 and try to be as inactive as possible or something like that. That itself is an ethical approach. And as you say, we are in no position to lock in our current understanding of ethics. Even if our current understanding of ethics is one of like, try to keep your hands off as much as you can, that's an ethical position that we might be wrong about, that we are right now locking into these AI systems that may one day essentially rule the world. I can't see this as a very solvable problem just by saying, well, we'll have the chatbot sort of slow down a bit and do it in a more psychologically, you know, palatable way. It still has to make that decision as to what it believes, right? That seems unavoidable. Yeah, so a couple of
Starting point is 01:08:52 things. I strongly agree that there's, you know, inaction as a form of action. And I am worried that people will see instruction following AI as the status quo, as the thing that is not imposing your values, where an instruction following AI just does whatever you want. I want to stage a coup of my country. The AI will say, yep, I'm going to do that. I want to build a bio weapon in my garage. Yep, I'll do that. I want to cheat my partner. Yeah, the AI will help you. That I think is like, yeah, not the way forward. And instead, we want to be differentially empowering people to do good things, rather than harmful things, at least for some very broad,
Starting point is 01:09:40 very thin notion of like what makes for the better society. I don't think that AI should be going with some particular political persuasion instead of convince people. I do actually, I kind of want to defend something that's like a kind of
Starting point is 01:09:58 division of labor approach where the AI that I'm using as a personal assistant, or something like that, you know, it might be in the role of like the good friend and good advisor, but without pushing any, you know, particular model view, just maybe it's got some, like, it's just generally helping you to think through, you know, be the kind of best version of yourself. And then maybe we have these kind of institutes where we've got all of the AI model philosophers and they're really thinking about, okay, what's the, how do all the arguments weigh up to
Starting point is 01:10:34 and come up with new arguments, new thought experiments, and so on. And then on the basis of that, we can get arguments and considerations kind of out into the public sphere. That seems to me like the best of both worlds, potentially. I'm very worried we won't do that latter thing, because, well, how much investment goes into moral philosophy at the moment? I expect it will be even less as a fraction of world expenses. literature over the coming years. But at least as a kind of ideal plan, that would seem quite
Starting point is 01:11:12 good to me. You're one of the sort of founders, popularizes of the effective altruist movement, which is all about how to do the most good in the most effective way possible. And sometimes it can be a little bit unconventional or unexpected. I mean, famously the idea that you sort of should stop like working for charity and instead go and make as much money as you can and donate it effectively and making sure you're donating to effective charities and all of this kind of stuff. And in the face of this sort of AI safety development, which seems to be an extremely high priority, some people think that as an effective altruist, you should essentially take all of your efforts and put them into solving this problem because this is like top of the priority list.
Starting point is 01:11:59 others have criticized this and the same thing occurs with like what we owe the future the book that I mentioned earlier that you wrote that there are going to be so many potential people in the future that like just by weighing up the the sort of utility like we should be putting way more attention into into preparing you know everyone in the future for living good lives and there's this idea that if we were being perfect rationalists perfect utilitarians we'd say okay let's stop bothering with this like malaria stuff and like yeah starvation that's bad you know but like some people are going to starve and that sucks but there are literally like trillions of people who will exist one day similarly you could say look yeah okay starvation is
Starting point is 01:12:44 bad but we're talking about like mosquito drones people this is like so much more existentially worrisome that if we're being rational we would just take all of our efforts away from world hunger malaria you know minor conflict that will probably sort themselves out like who cares you know some people will die whatever this is what we should focus on and others hear that and go like not only is that just intuitively untrue that we should still care about those things but also that seems to lose sight of what like effective altruism was all about and so do you think that that is a rational ethical policy and if so or if not what does that mean for the future of the effective altruist movement if the
Starting point is 01:13:26 most effective way that we can care about other people is to just ignore everything and focus on the AI apocalypse. Okay, great questions. And I think there's two things to say. The first is to really distinguish between what should the world as a whole do versus what should you as an individual do. Now, there's like many very big problems in the world. There's malaria, there's tuberculosis, HIV, there's gender discrimination, there's racial discrimination, there's climate change, there's global inequality, there's authoritarianism, there's war, the list just keeps going on. But as an individual, or even as a community of thousands of people, you're not going to be able to solve all of these problems. You're not even going to be able to solve one of them.
Starting point is 01:14:12 Instead, you're going to make some, you know, small in relative terms, dent, huge. huge in absolute terms, dent, where the huge and absolute terms means, you know, saving dozens of lives over the course of your life or a benefit even larger. And so given that, there's these very different questions. You might even think that one problem is, you know, maybe you think AI is the biggest problem in the world. And you think that risk of war is, you know, not as big a problem what's like, but nonetheless,
Starting point is 01:14:53 you're much better fit for the juicing risk of war. Then you don't need to work on what you think the world's biggest problem is. Instead, you should work in where you think you can have the biggest difference.
Starting point is 01:15:06 So, that's the first thing to say is saying, like, on the margin, given how neglected something is, there should be radically more effort in there, isn't saying kind of all of the world's effort should go into that So, like, yeah, my kind of views on long-termism, I felt often got quite misunderstood where
Starting point is 01:15:26 the bar I was advocating for is really quite low. It was that rich countries in the world should allocate 1% of GDP to issues that distinctively benefit the long-term future. That would put it on the par with, you know, minimal ethical standards for global development to. That's a far cry from saying we should spend like 99.9% of the world's economy. me to benefit future generations. Now, Effective Altruism is such a small actor.
Starting point is 01:15:56 So how should it focus? And they're the main thing I want to say is that there are many different people with very different views. And I think the best and healthiest kind of effective altruism movement is one where people do just figure out, like, okay, what are there reflective ethical views? What do they think is going to happen in the world? So then where do they think they are going to have the best? biggest impact. If that means everyone figured thinks that working on making AI go well is the
Starting point is 01:16:25 best thing to do, then fine. In practice, it's much more split than that. So there's a lot of people still focused on, a lot of people who are focused on global health and development. Give Well, which is a kind of charity evaluation organization, has now moved over the billion dollars, mainly from, you know, significantly from small donors. There's others who focus on factory farming, but then, yeah, there's a lot of people who are focused on AI too. And I think that these issues that we've been talking about, both AI safety, but also concentration of power, risks from dangerous novel technologies, risks of value lock-in,
Starting point is 01:17:09 are just huge, they're really big. and it's really close to zero people who are working on the, at least on the intersection of very advanced AI, like AGI, and these other issues. And so I do think that this should be at least one big focus for the EA movement as a whole. At least I'm making the argument that it should be. And if people disagree, then, you know, good. You know, you say like with long-termism, which is this sort of, you know, caring towards the future people who will exist.
Starting point is 01:17:43 And obviously, I'd recommend people to read your book. We'll link it in the description and stuff. But, you know, you make a compelling case that given just how many people may exist in the future compared to now, that there is just such a huge level of utility in their interests and worrying about their interests. And you say, but look, you know, I'm not calling for us to focus all our energies here. I'm just calling for, you know, 1% of GDP, but like... why not like if it's true what you're saying if it is like actually that important isn't that a bit
Starting point is 01:18:16 sort of misaligned with it's as if i said to you like you know like my my house is like burning down and like everything's on fire and all of my belongings are but you know like i really i haven't updated my twitter banner in a while and like we've got a you know we can't put all of our interests into one thing so i'm i'm gonna you know maybe i'll just do that first and go and grab the hose pipe afterwards. It's like, based on what you said, it would seem like you should be focusing on the house that's on fire. Yeah. So, I mean, the analogy might be like, you know, you have all your savings, you've got your house that's at risk of burning down. What fraction of your savings should you spend to prevent your house from burning down?
Starting point is 01:19:02 Well, it depends on how much it costs to prevent it. If you can prevent it from burning down with $1,000,000 pounds, and then you don't have to do any more than that, you should spend a thousand pounds and you shouldn't spend more. And that's true, even if burning down their house would completely ruin your life and is the most important thing ever. So, like, the size of a problem doesn't need to match up with the kind of cost of the solution. And so how much could we reasonably spend to benefit future generations? Well, at the moment, there are these amazing opportunities to help the future go better by protecting ourselves against worst-case pandemics, by helping to steer how well AI goes.
Starting point is 01:19:43 If we ramped up to kind of 1% of GDP of rich countries, that's like so much more than we are investing at the moment that I really don't know, like, would additional, how good would additional funding or additional effort look? And in particular, would it look very different than just generally building a flourishing society where there's lots of, you know, there's huge number of things that overlap where what's good in the short term is good in the long term and vice versa. They're, you know, even in the case of just preventing the next pandemic or helping ensure
Starting point is 01:20:24 AI goes well, like what, in terms of my mortality, risk over the next 10, 20 years? What are the things that are most likely to kill me? Actually, I think either AI or innovations that are downstream of AI are higher than cancer. So I actually think lots of these things are beneficial in the short term too. And so perhaps once we were a world that was more sane and just more thoughtful towards future generations and was investing a significant amount of effort, then plausibly to me, the things that look good, yeah, would look best for future generations also just look best for society. It would be being wiser, being richer, being better educated, and so on.
Starting point is 01:21:10 Well, Will McCaskill, lots to chew on, links, et cetera. We've mentioned a couple of articles and things, that video on AI, your previous book, What We owe the Future. It'll all be in the description. Thanks for your time. Great. Thanks so much, Alex.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.