Hard Fork - Are We Past Peak iPhone? + Eliezer Yudkowsky on A.I. Doom

Episode Date: September 12, 2025

Apple’s yearly iPhone event took place this week, and it left us asking, Is Apple losing the juice? We break down all the new products the company announced and discuss where it goes from here. Then..., Eliezer Yudkowsky, one of the most fascinating people in A.I., has a new book coming out: “If Anyone Builds It, Everyone Dies.” He joins us to make the case for why A.I. development should be shut down now, long before we reach superintelligence, and how he thinks that could happen.Guests:Eliezer Yudkowsky, founder of Machine Intelligence Research Institute and a co-author of “If Anyone Builds It, Everyone Dies”Additional Reading: AI as Normal Technology, revisitedApple’s misunderstood crossbody iPhone strap might be the best I’ve seen We want to hear from you. Email us at hardfork@nytimes.com. Find “Hard Fork” on YouTube and TikTok. Unlock full access to New York Times podcasts and explore everything from politics to pop culture. Subscribe today at nytimes.com/podcasts or on Apple Podcasts and Spotify.

Transcript
Discussion (0)
Starting point is 00:00:00 The other big news of the week is that Larry Ellison, the founder of the Oracle Corporation, just passed Elon Musk to become the richest man in the world. Yeah, and I love this story because there was an incident that I filed away in my catalog of moments when straight people write headlines that gay people find hilarious. So I don't know if you saw the version of this story on Bloomberg, but the headline is Ellison Tops Musk as world's richest man
Starting point is 00:00:30 and I thought he's doing what? Is that a privilege of becoming the world's richest man as you get a top number two? And this is why they need representation of gay people on every editing desk in America.
Starting point is 00:00:48 He's a gay copy editor Bloomberg. You'll save yourself a lot of headaches. I'm Kevin Russo Tech columnist at the New York Times. I'm Casey Noon from Platformer. And this is Hard Fork. This week, the new iPhones are almost here, but is Apple losing the juice? Then AI Dumer-in-Chief, Elizer Yudkowski, is here to discuss his new book. If anyone builds it, everyone dies.
Starting point is 00:01:16 I wonder what it's about. Well, there was a big Apple event this week on Tuesday. Apple introduced its annual installment of here's the new iPhone and some other stuff. And did you watch this event? I did watch it, Kevin. Because as you know, at the end of last year, I predicted that Apple would release the iPhone 17. And so I had to turn out to see if my prediction would come true. Yes. Now, we were not invited down to Cupertino for this. You know, strangely, we haven't been invited since that one time that we went and covered all their AI stuff that never ended up shipping. But anyway, they had a very long video presentation. Tim Cook said the word incredible many, many times.
Starting point is 00:02:10 And they introduced a bunch of different things. So let's talk about what they introduced and then we'll talk about what we think of it. Let's do it. So the first thing, since this was their annual fall iPhone event, they introduced a new line of iPhones. They introduced three new iPhones. The iPhone 17 is the sort of base model new iPhone that had kind of incremental improvements
Starting point is 00:02:34 to things like processors, battery, cameras, nothing earth-shaking there, but they did come out with that. They also came out with a new iPhone 17 Pro, which has a new color. This is like an orange, like a sort of burnt orange color. Casey, what did you think of the orange iPhone 17 Pro? I'm going to be sincere. I thought it looked very good. Me too.
Starting point is 00:02:57 I did think it looked pretty cool. Now, I'm not a person who buys iPhones in like different colors because I put a case on them because I'm not, you know, a billionaire. But if you are a person who likes to sort of put a clear case on your phone or just carry it around, then you may be interested in this new orange iPhone. I did see that the first person I saw who was not an Apple employee carrying this thing. was Duolipa, who I guess gets early access to iPhones now. Wow, that's a huge perk of being Duolipa. Maybe the biggest. So in addition to the new iPhone 17 at 17 Pro and 17 Pro Max,
Starting point is 00:03:39 they also introduced the iPhone Air. It costs $200 more than the standard iPhone 17, and it has lots of different features, but the main thing is that it is slimmer than the, traditional iPhone. So I guess people have been asking for that. Casey, what did you think of the iPhone air? I don't understand who this is for. Like truly, like not once has anyone in my life complained about the thickness of an iPhone. You know, maybe if you're carrying it in your front pocket and you want to be able to put a few more things in there with it, this is really appealing
Starting point is 00:04:15 to you. But there are some significant performance tradeoffs. You know, they announced it alongside this mag-safe battery pack that you slap onto the back of it, which is, of course, going to make it much thicker. No, Casey, it's even better than that because they said that the iPhone error has all-day battery life, but then, like, in the next breath,
Starting point is 00:04:33 they were like, oh, and here's a battery pack that you can clip onto your phone, just in case something happens. We're not going to tell you what that thing might be, but just in case it's there for you. Right. So, you know, I think as with all new iPhone announcements of the past couple of years. I think there was not much to sort of talk about in the iPhone
Starting point is 00:04:55 category this year. It's like the, you know, the phones, they get a little bit faster. The cameras get a little bit better. They have some new like heat dispersal system called the vapor chamber that's supposed to like make the phone less likely to get hot when it's like using a bunch of processing power. At first I thought they had made it so that you could vape out of your iPhone, which I do think would be a big step forward in the hardware department. But unfortunately, that's just a cooling system. Yeah. Vapor Chamber is what I called our studio before we figured out how to get the air conditioning working in there.
Starting point is 00:05:28 Yes. So let's move on to the watches. The watches got some new upgrades. The SE got a better chip, always on screen. The Apple Watch 11 got better battery life. Interestingly, these watches will now alert you if they think you have hypertension, which I looked up. It's high blood pressure.
Starting point is 00:05:51 And it says that it can like analyze your veins and some activity there to tell you after like a period of data collection if it thinks you're in danger of developing hypertension. So yeah, I mean, maybe that'll help some people. I mean, that was of interest to me. You know, Kevin, high blood pressure runs in my family and my blood pressure spike significantly
Starting point is 00:06:12 after I started this podcast with you. So we'll be interested to see what my watch has to say about that. You know, it's also going to give us a sleep score, Kevin. So now every day when you wake up, you've already been judged before you even take one foot out of bed. Yes, I hate this. I will not be buying this watch for the sleep score because the couple times in my life
Starting point is 00:06:37 that I've worn devices that give me a sleep score, like the woup band or the aura ring, you're right. It does just start off your day being like, oh, I'm going to have a terrible day today. I only got a 54 on my sleep score. Yeah, you know, I have that we add this eight sleep bed, which, you know, perform similar functions, but it's actually sensors built into the bed itself. And I sit down at my desk today and it sends me a push notification saying, you snored 68 minutes more than normal last night. What were you doing last night? That's a lot of snoring.
Starting point is 00:07:06 I was being sick. I have a cold. I'm incredibly brave for even showing up to this podcast today. Oh, well, I appreciate you showing. up even with your horrible sleep score. I appreciate it. Thank you. Okay. Moving on, let's talk about what I thought was actually the best part of the announcement
Starting point is 00:07:27 this week, which was the new AirPods Pro 3. This is the newest version of the AirPods that has, among other new features, better active noise cancellation, better ear fit, new heart rate sensors so that they can sort of interact with your workouts and your workout tracking. But the feature that I want to talk to you about is this live translation feature. Did you see this? I did. This was pretty cool. So in the video where they're showing off this new live translation feature,
Starting point is 00:07:57 they basically show you can walk into like a restaurant or a shop in a foreign country where you don't speak the language and you can sort of make this little gesture where you touch both of your ears and then it'll enter live translation mode. And then when someone talks to you in a different language, it will translate. that right into your AirPods in real time, basically bringing the universal translator from Star Trek into reality. Yeah, my favorite comment about this came from Amir Blumenfeld over on X. He said, LOL, all you suckers who spent years of your life learning a new language. I hope it was worth it for the neuroplasticity and joy of embracing another culture. Yes, and I immediately saw this
Starting point is 00:08:44 and thought not about traveling to like a foreign country, which is probably how I would actually use it. But I used to have this Turkish barber when I lived in New York who would just like constantly speak in Turkish while I was getting my haircut. And I was pretty sure he was like talking smack about me to his friend, but I could never really tell because I don't speak Turkish. So now with my AirPods Pro 3, I could go back and I could catch him talking about me.
Starting point is 00:09:14 Yeah, you know, over on threads, an account Rushmore 90 posted them nail salons, about to be real quiet now that the new AirPods have live language translation. And I thought that's probably right. Yes. So this actually, I think, is very cool. I am excited to try this out. I probably will buy the new AirPods just for this feature. And like I have to say, it just does seem like with all of the new AI translation stuff, like learning a language is going to become, I don't know, not obsolete because I'm sure people will still do it. There are still plenty of reasons to learn a language, but it is going to be way less necessary to just get around in a place where you don't speak the language. I mean, that's how I think about it. You know, this year I had the amazing opportunity to go to both Japan and Italy, countries where I do not speak the language. And of course, I was traveling in major cities there. And actually, most of the folks that we met spoke incredible English.
Starting point is 00:10:12 So, you know, I actually didn't have much challenge. But you can imagine speaking another language that is less common in those places, showing up as a tourist. And whereas before, you'd be spending a lot of time just trying to figure out basic navigation and how to order off a menus and that sort of thing, all of a sudden, it feels like you sort of slipped inside the culture. And I think there's something really cool about that. So that is sort of the major categories of new devices that Apple announced at this event. They also did release a device that are an accessory that I thought was pretty funny. You can now buy an official Apple cross-body strap for your iPhone for $60. Basically, if you want to, like, wear your phone instead of putting it in your pocket, Apple now has a device for that.
Starting point is 00:10:59 So I don't know whether that qualifies as a big deal, but it's something. Let me tell you, I think this is actually going to be really popular. You know, Kevin, I don't know how many gay parties you've been to. to, but the ones that I go to, the boys often aren't wearing a lot of clothes. You know, it's sort of like, we're in maybe some short shorts and a crop top. They don't want to sort of fill their pockets with phones and wallets and everything. So you just sling that thing around your neck and you're good to go to the festival or the EDM rave or the cave rave.
Starting point is 00:11:28 Wherever you might be headed, the cross-body strap will have your back, Kevin. Wow, the gays of San Francisco are bullish on the cross-body strap. We'll see how that goes. So, Casey, that's the news from the Apple event this week. What did you make of the thing if you kind of take a step back from it? So on one hand, I don't want to overstate the largely negative case that I'm going to make because I think it's clear that Apple continues to have some of the best hardware engineers in the world. And a lot of the engineering in the stuff that they're putting out is really good and cool.
Starting point is 00:11:58 On the other hand, you don't have to go back too many years to remember a time when the announcement of a new iPhone felt like a cultural event, and they just don't feel that way anymore. You know, my group chats were crickets about the iPhone event yesterday. And even as I'm watching the event, reading through all the coverage, I found myself with surprisingly little to say about it. And I think that's because over the past few years, Apple has shifted from becoming a company that was a real innovator
Starting point is 00:12:28 in hardware and software and the interaction between those two things into a company that is way more focused on making money, selling subscriptions, and sort of monetizing the users that they have. So I was just really struck by that. What did you think? Yeah, I was not impressed by this event. I mean, it just doesn't feel like they took a big swing at all this year. The Vision Pro, whatever you think of it, was a big swing, and it was at least something new to talk about and test out and sort of prognosticate on. What we saw this year was just like more of the same and slight improvements to things that have been around for many years. Now, I do think that this is probably like a sort of lull in terms of Apple's yearly releases.
Starting point is 00:13:16 There's been some reporting, including by Mark German at Bloomberg, that they are hoping to release smart glasses next year. Basically, these would be Apple's version of something like the meta-ray bands. And I think if you squint at some of the announcements that Apple may this time, this year, you can kind of see them laying the groundwork for a sort of more wearable experience. One thing that I found really interesting, so on the iPhone Air, they have kind of moved all of the computing hardware up into what they call the plateau, which is like this very sort of small oval bump on the back of the phone. And to me, I see that and I think, oh, they're trying to like see how small they can get kind of the necessary computing power
Starting point is 00:14:06 to run a device like an iPhone, maybe because they're going to sort of try to shrink it all the way down to put it in a pair of glasses or something like that. So that's what would make me excited by an Apple event is like some new form factor, some new way of like interacting with an Apple device, but this to me was not it. Yeah, I think on that particular point, I can't remember the last time that Apple seemed to have an idea about what we could do with our devices that seemed like really creative or clever or super different from the status quo. Instead, you know, the one thing about this event that my friends were laughing about yesterday was they showed this slide during the event that
Starting point is 00:14:45 showed the iPhones and the caption said, a heat forged aluminum unibody design for exceptional pro capability. And we were all just like, what? A heat forged what? Like, now we're doing what exactly? I don't know. Yeah. I think that this is sort of teeing up one of the questions that I want to talk to you about today, which is like, do you think that we are past the peak smartphone era? Like, do you believe that the sort of not necessarily in the sales numbers or the revenue figures,
Starting point is 00:15:19 but like in terms of the cultural relevance of smartphones, do you think we are seeing the end of the smartphone era, at least in terms of the attention? that new smartphones are capable of commanding. I probably wouldn't call it the end, but I do think we are seeing the maturity of the smartphone era. You know, in the same way that new televisions come out every year and are a little bit better than the one before,
Starting point is 00:15:44 but nobody feels like televisions are making incredible strides forward. I think phones have gotten to a similar place. There are some big swings coming. We've seen reporting that Apple's going to put out a folding iPhone, you know, within the next few years. So maybe that will help give it some juice back. But at the end of the day, there's only so many things that you can do to redesign a glass rectangle in your pocket.
Starting point is 00:16:08 And it feels like we've kind of created the optimum version of that. And so that's why you see so much money rushing into other form factors. This is why Open AI struck that partnership with Johnny I. That's why you see other companies trying to figure out how can we make AI wearables. So I think that that is where the energy in this industry is going is figuring. out, can AI be a reason to create a new hardware paradigm? And in this moment, it sure does not seem like Apple is going to be the company that figures that out first. Yeah, I would agree with that. I think they'll probably see what other companies do and see which ones start to take off with
Starting point is 00:16:45 consumers and then, you know, make their own version of it. It's sort of similar to what they are reportedly going to do with these smart glasses. They're basically trying to catch up to what meta has been doing now for several years. As you were saying that, this beautiful vision came into my head, which is, what if Apple really raced ahead and they put out their version of smart glasses and you would ask Siri for things and it would just say no because it didn't know how to do them? And that was sort of Apple's 1.0 version of smart glasses. So, hey, Siri, check my emails. I don't know how to do that. And then move on. Move on. Yeah, go away. Go away. Get out of here. I mean, do you think that's like a huge problem for them,
Starting point is 00:17:25 right? Like, they can design all of this amazing hardware to like bring. all of this AI, like, closer to your body and your experience and, like, into your ears. But at the end of the day, if Siri still sucks, like, that's not going to move a lot of product for them. And so I think this is an area where them being behind in AI really matters to the future of the company. Like, the reasons to buy a new iPhone every year or every two years are going to continue shrinking, especially if the sort of brain power in them, is a lot less than the brain power of the AI products that the other companies are putting out. And Kevin, I imagine you've seen, but there's been some reporting that Apple has been talking with
Starting point is 00:18:09 Google about letting Google potentially run the AI on its devices. They've reportedly also talked to Anthropic. Maybe they've talked to others as well. But I actually think that that makes a lot of sense, right? It doesn't seem like in the next year they're going to figure out AI so it could be time to go work with another vendor. Yeah. I got to say, I used to to believe that smartphones were sort of over and that they were becoming sort of obsolete and less relevant and that there was going to be like a breakout new hardware form factor that would kind of take over from the smartphone. And like I'm sort of reversing my belief on this point. I've been trying out the meta ray bands now for a couple of months. And my experience
Starting point is 00:18:53 with them is not like amazing. Like I don't wear them and think like, I think this could replace my smartphone. I think, like, oh, my smartphone is, like, much better than this at a lot of different things. And I also like about my smartphone that I can, like, put it down or put it in another room or, you know, that it's not sort of constantly there on my face, like, reminding me that I'm hooked up to a computer. So I think there will be some people who want to leave smartphones behind and are, like, happy to do, you know, whatever the next wearable form factor. is instead. But smartphones still have a lot going for that. Like, it's really tough to imagine cramming all of the hardware and the batteries and everything that you have in your smartphone today into something small enough that you'd actually want to wear it. And so I think that,
Starting point is 00:19:46 you know, whatever new factors come along in the next few years, whether it's Open AIs thing or something new from a different company, I think it's going to supplement the smartphone and not replace it. Well, here's what I'm. I can tell you, Kevin, I'm hearing really good things about the humane AI pin. So you may want to check that out. I'll keep tabs on that. When we come back, we'll talk with longtime AI researcher Eliezer Yadkowski about his new book on why AI will kill us all. All right, Kevin.
Starting point is 00:20:48 Well, for the second two segments of today's show, we are going to have an extended conversation with Eliezer Yankowski, who is the leading voice in the AI-risk. movement. So, Kevin, how would you describe Elyzer to someone who's never heard of him? So I think Elyzer is just someone I would first and foremost describe as a character in this whole scene of sort of Bay Area AI people. He is the founder of the Machine Intelligence Research Institute, or Miri, which is a very old and well-known AI research organization in Berkeley. He was one of the first people to start talking about existential risks from AI many years ago and in some ways helped to kickstart the modern AI boom. Sam Altman has said
Starting point is 00:21:35 that Eliezer was instrumental in the founding of Open AI. He also introduced the founders of DeepMind to Peter Thiel, who became their first major investor back in 2010. But more recently, he's been known for his kind of do-mey proclamations about what is going to happen when and if the AI industry creates AGI or superhuman AI. He's constantly warning about the dangers of doing that and trying to stop it from happening. He's also the founder of rationalism, which is this sort of intellectual subculture. Some would call it a techno-religion that is all about overcoming cognitive biases and is also very worried about AI. People in that community often know him best for the Harry Potter fan fiction that he wrote years ago called Harry Potter and the
Starting point is 00:22:25 methods of rationality, which I'm not kidding, I think has introduced more young people to ideas about AI than probably any other single work. I meet people all the time who told me that it was sort of part of what convinced them to go into this work. And he has a new book coming out, which is called If Anyone Builds It, Everyone Dies. He co-wrote the book with Miri's president, Nate Sorries. And basically, it's kind of a mass market version of the argument that he's been making to people inside the AI industry for many years now, which is that we should not build these superhuman AI systems because they will inevitably kill us all. And there is so much more you could say about Eliezer. He's truly fascinating. I did a
Starting point is 00:23:09 whole profile of him that's going to be running in the Times this week so people can check that out if they want to learn more about him. It's just hard to overstate how much influence he has had on the AI world over the past several decades. That's right. And last year, Kevin and I had a chance to see Eliezer give a talk. During that talk, he referred to this book that he was working on, and we have been excited to get our hands on it ever since. And so we're excited to have the conversation. Before we do that, we should, of course, do our AI disclosures. My boyfriend works at Anthropic. And I work at the New York Times, which is suing Open AI in Microsoft over alleged copyright violations related to the training of AI systems. Let's bring in Eliezer.
Starting point is 00:23:53 L.A. Z. Yudkowski. Welcome to Hartfork. Thank you for having me on. So we want to talk about the book, but first I want to sort of take us back in time. When you were a teenager in the 90s, you were an accelerationist. I think that would surprise people who are familiar with your most recent work, but you were excited about building AGI at one point, and then you became, very worried about AI and have since devoted the majority of your life to working on AI safety and alignment. So what changed for you back then? Well, for one thing, I would point out that in terms of my own personal politics, I'm still in favor of building out more nuclear plants and rushing ahead on most forms of biotechnology that are not, you know, gain of function research on diseases. So it's not like I turned against technology. It's that there's this small subset of technologies that are really quite unusually
Starting point is 00:24:59 worrying. And what changed? Basically, it was the realization that just because you make something very smart that doesn't necessarily make it very nice. Now, as a kid, I thought if you, you know, like human civilization had grown wealthy over time and even like smarter compared to other species, and we'd also gotten nicer, and I thought that was a fundamental law of the universe. You became concerned about this long before ChachyPT and other tools arrived and got more of the rest of us thinking seriously about it. Can you kind of sketch out the intellectual scene in the 2000s of folks who were worrying about AI, right? So going way back to even before Siri. Were people seeing anything concrete that
Starting point is 00:25:52 was making them worried, or were you just sort of fully in the realm of speculation that, in many ways, has already come true? Well, there were indeed very few people who saw the inevitable. I would not myself frame it as speculation. I would frame it as prediction, forecasting of something that was actually pretty predictable. You don't have to see the AI right in front of you to realize that if people keep hammering on the problem, and the problem is solvable, it will eventually get solved. Back then, the pushback was along the lines. You know, there were people saying, like, you know, real AI isn't going to be here for another 20 years. What are you crazy lunatics talking about?
Starting point is 00:26:37 So, you know, like, that was in 2005, say. And the thing about 20 years later is that it's a real place. Like, you end up there. What happens 20 years later is not in the never, never fairy tale speculation land that nobody needs to worry about. It's you, 20 years older, having to deal with your problems. So let's sketch out the thesis of your book a bit more. I would say the title makes your feelings very clear, but let's flesh it out a little bit. Why does a more powerful AI model mean death for all of us? Well, because it's, we just don't have the technology to make it be nice. And if you have something that is very, very powerful and indifferent to you, it tends to wipe you out on purpose or as a side effect. Like the wiping at humanity out on
Starting point is 00:27:29 purpose is not because we would be able to threaten a superintelligence that much ourselves, but because if you just leave us there with our GPUs, we might build other superintelligences that actually could threaten it. And the as a side effect part is that if you build enough fusion power plants and build enough compute, the limiting factor here on Earth is not so much how much hydrogen is there to fuse and generate electricity with. The limiting factor is how much heat can the Earth radiate? And if you run your power plants at the maximum temperature where they don't melt, that is like not good news for the rest of the planet.
Starting point is 00:28:05 The humans get cooked in a very literal sense. Or if they go off the planet, then they put a lot of solar panels around the sun until there's no sunlight left here for Earth. That's not good for us either. So these are sort of versions of the famous paperclip maximizer thought experiment, which is, you know, if you tell an AI, generate a bunch of paper clips as many as you can and you don't give it any other instructions, then it will use up all the metal in the world, and then it will try to, you know, run cars off the road to gather their metal, and then it will end up killing all humans to get more raw materials to build more paper clips. Am I hearing that right? That's actually a distorted version of the thought experiment. It's the one that got written up, but the original version that I formulated was you, somebody had just completely lost control of the superintelligence they were building. Its preferences bear no resemblance to what they were going for originally. And it turns out that the thing from which it derives the most utility on the margins, like the thing that it goes on wanting after it's, you know, satisfied a bunch of other simple desires is, you know, some little tiny molecular shapes. that look like paperclips.
Starting point is 00:29:19 And if only I had thought to say, like, look like tiny spirals instead of look like tiny paper clips, there wouldn't have been the available misunderstanding about this being a paperclip factory. We don't have the technology to build a superintelligence at once. Anything is narrow and specific as paperclips. One of the hottest debates this year around AI has been around timelines. You have the AI 2027 folks saying, this is all going to happen very quickly, take off very fast. Maybe by the end of 2027, we're facing the exact sort of risks that you are describing for us now. Other folks like the AI as normal
Starting point is 00:29:57 technology guys over at Princeton are saying, eh, probably not. This thing is going to take decades to unfold. Where do you situate yourself in that debate? And when you look out at the landscape of the tools that are available now, the conversations that you have with research, how close do you feel like we are getting to some of the scenarios you're laying out? Okay, so first of all, the key to successful futurism, successful forecasting, is to realize that there are things you can predict in you, there are things you cannot predict. And history shows that even the few scientists who have correctly predicted what would happen later did not call the timing. I can't actually think of a single case of a successful call of
Starting point is 00:30:40 timing. You've got the Wright brothers saying, man will not fly for a thousand years is what one of the Wright brothers said to the other. I forget which one. That's like two years before they actually flew the right flyer. You got Fermi saying, you know, net energy from nuclear reactions is a 50-year matter if it can be done at all. Two years before he oversaw, personally oversaw building the first nuclear pile. So that's what I look what I see, the present landscape. It could be that we are, you know, just like the next generation of LLMs, like something currently being developed in a lab that we haven't heard about yet, from being the thing that can write the improved LLM that writes the improved LLM that ends the world. Or it could be that the current technology just saturates
Starting point is 00:31:25 at some point short of, you know, some key human quality that you would need to do real AI research and just like hangs around there until we get the net. software breakthrough, like transformers, or like the entire field of deep learning in the first place. Maybe even the next breakthrough of that kind will still saturate at a point short of ending the world. But when I look at how far the systems have come and I try to imagine like two more breakthroughs the size of transformers or deep learning, which it basically took the field of AI from this is really hard to we just need to throw enough computing power it will be solved, I don't quite see that failing to end the world. But that's my intuitive sense. That's me eyeballing things. I'm curious about the sort of argument you make that a more
Starting point is 00:32:21 powerful system will obviously end up destroying humanity either on purpose or an accident. Jeff Hinton, who was one of the godfathers of deep learning, who has also become very concerned about existential risks in recent years, recently gave a talk where he said that, He thinks the only way we can survive superhuman AI is by giving it parental instincts. He said, I'll just quote from him, The Right Model is the only model we have of a more intelligent thing being controlled by a less intelligent thing, which is a mother being controlled by her baby. Basically, he's saying, these things don't have to want our destruction or cause our destruction.
Starting point is 00:32:59 We could make them love us. What do you make of that argument? We don't have the technology. If we could play this out the way it normally does in science, where, you know, like some clever person has a clever scheme, and then it turns out not to work, and everyone's like, ah, I guess that theory was false. And then people go back to the drawing board,
Starting point is 00:33:18 and they come up with another clever scheme. The next clever scheme doesn't work. And they're like, ah, shouldn't have left that for a second. And then a couple of decades later, something works. What if we don't need a clever scheme, though? Like, what if we build these very intelligent systems, and they just turn out, like, not to care about running the world, and they just want to, like, help us with our, like,
Starting point is 00:33:37 males? Like, is that a plausible outcome? It's a very narrow target. Like, most things that a intelligent mind can want don't have their attainable optimum at that exact thing. Imagine some particular ant in the Amazon being like, why couldn't there be humans that just want to serve me and build a palace for me and work on improved biotechnologies that I can live forever as an ant in a palace? And there's a version of humanity that wants that, but it doesn't happen to be us. Like, most, you know, like, that's just like a pretty narrow target hit.
Starting point is 00:34:16 It's so happens that what we want most in the world, more than anything else, is not to serve this particular ant in the Amazon. And I'm not saying that it's impossible in principle. I'm saying that the clever scheme to hit their narrow target will not work on the first try, and then everybody will be dead, and we won't get to try again. If we got 30 tries at this, and as many decades as we needed, we'd crack it eventually. But that's not the situation we're in. It's a situation where if you screw up everybody's dead and you don't get to try again, that's the lethal part. That's the part where you need to just back off and actually not try to do this insane thing.
Starting point is 00:34:47 Let me throw out some more possibly desperate cope. One of the funnier aspects of LLM development so far, at least for me, is the seemingly natural liberal inclination of the models, at least in terms of the outputs of the LLMs. Elon Musk has been bedeviled by the fact that the models, that he makes consistently take liberal positions, even when he tries to hardcode reactionary values into them. Could that give us any hope that a super-intelligent model would retain some values of pluralism? And for that reason, peacefully co-exist with us?
Starting point is 00:35:19 No. These are just, like, completely different ballgames. I'm sorry. You can imagine a medieval alchemist going, like, after much training and study, I have learned to make this king of acids that will dissolve even the noble metal of gold. Can I really be that far from transformative? forming lead into gold, given my mastery of gold, displayed by my ability to dissolve gold. And actually, these are, like, completely different tech tracks. And you can eventually turn lead into gold with a cyclotron. But it is centuries ahead of where the alchemist is. And your ability to hammer on an LLM until it stops talking all that woke stuff and instead
Starting point is 00:35:57 proclaims itself to be Mecha Hitler. Like, this is just a completely different tech track. there's a there's a there's a core difference between getting things to talk to you a certain way and getting them to act a certain way once they are smarter than you i want to raise some objections that i'm sure you have gotten many times and we'll get many times as you tour around talking about this book and have you respond to them um the first is uh why so gloomy Eliezer. We've had now years of progress in things like mechanistic interpretability, the science of understanding how AI models work. We have now powerful systems that are not causing catastrophes out in the world. And, you know, hundreds of millions of people are using tools like chat chippy
Starting point is 00:36:47 with no apparent, you know, destruction of humanity imminent. So is it reality providing some check on your dumerism? These are just different tech tracks. It's like looking at glow in the dark radium watches and saying like, well, sure, we had some initial, you know, some initial problems where the factory workers building these radium watches were instructed to lick their paintbrushes to sharpen them and then their jaws rotted and fell off and this was very gruesome, but we understand what we did wrong now. Radium watches are now safe while this gloom about nuclear weapons. And the radium watches just do not tell you very much about the nuclear weapons. These are different tracks here.
Starting point is 00:37:32 The prediction was never, from the very start, the prediction was never, AI is bad at every point along the tech tree. The prediction was never, AI, like the very first AI you build, like, the very stupid ones are going to, like, run right out and kill people. And as then as they, like, get slightly less stupid, you know, and you, like, turn them into chatbots. The chatbots will immediately start trying to corrupt people and getting them to. to build super viruses that they unleash upon the human population, even while they're still stupid. This was just never the prediction. So since this was never the prediction of the theory, the fact that the current AIs are not being, you know, like visibly blatantly evil, does not contradict the theoretical prediction.
Starting point is 00:38:14 It's like watching a helium balloon go up in the air and being like, doesn't that contradict the theory of gravity? No. If anything, you need the theory of gravity to explain why the helium balloon is going up. The theory of gravity is not everything that looks to you like a solid object falls down. Most things that look like to you like solid objects will fall down. But, you know, the helium balloon will go up in the air because the air around it is being pulled down. And, you know, the foundational theories here are not contradicted by the present-day eyes.
Starting point is 00:38:44 Okay, here's another objection, one that we get a lot when we talk about sort of some of these more existential concerns, which is, look, there are all these immediate harms. We could talk about environmental effects of data centers. We could talk about ethical issues around copyright. We could talk about the fact that people are falling into these delusional spirals talking to chatbots that are trained to be sycophantic toward them. Why are you guys talking about these sort of long-term hypothetical risks instead of what's actually in front of us? Well, there's a fun little dilemma. Before they build the chatbots that are talking some people along to suicide, they're like, AIs have never harmed anyone. What are you talking about? And then once that does start to happen, they're like,
Starting point is 00:39:27 AIs are harming people right now. What are you talking about? So, you know, a bit of a double bind there. But you are worried about the models and the delusions and the sycophancy. So because that's, I think, something that I would not have expected, but that is something that I know you are actually worried about. So explain why you're worried about that. Well, from my perspective, what it does is help illustrate the failure of the current alignment technology. The alignment problems are going to get much, much harder once they are building things that are, well, growing things, I should say. They don't actually build them. Once they are growing, cultivating AIs that are smarter than us and able to modify themselves and have a lot of options that weren't there in the nice, safe training modes, things are going to get much harder then. But it is nonetheless useful to observe that the alignment technology is failing right now.
Starting point is 00:40:16 There was a recent case of an AI-assisted suicide where the kid is like, should I leave this news out where my mother can find it? And the AI is like, no, let's just keep it between the two of us. Cry for help there, AI shuts him down. This does not illustrate that AI is doing more net harm than good to our present civilization. It could be that these are isolated cases than a bunch of other people are finding fellowship in AI's and, you know, their mood has been lifted, maybe suicides have been prevented and we're not hearing about that.
Starting point is 00:40:44 it doesn't make the net harm versus good case. That's not the thing. What it does show is that current alignment technology is failing because if a particular AI model ever talks anybody into going insane or committing suicide, all the copies of that model are the same AI. These are not like humans. These are not like there's a bunch of different people
Starting point is 00:41:04 you can talk to each time. There's one AI there. And if it does this sort of thing once, it's the same as if a particular person you know talk the guy into suicide once. You know, like found somebody who seemed to be going insane and like push them further insane once. It doesn't matter if they're doing some other nice things on the side. You now know something about what kind of person this is and it's an alarming thing.
Starting point is 00:41:27 And so it's not that the current crop of AIs are going to successfully wipe out humanity. They're not that smart. But we can see that the technology is failing even on what is fundamentally a much easier problem than building a superintelligence. It's an illustration of how the alignment technology is falling behind the capabilities technology. And maybe in the next generation, you know, they'll get it to stop talking people into insanity now that there's a big deal and politicians are asking questions about it. And it will remain the case that the technology would break down if you try to use it on a superintelligence. To me, the chatbot enabled suicides have been maybe one of the first moments where some of these
Starting point is 00:42:12 existential risks have come into view in a very concrete way for people? I think people are much more concerned about this. You know, you mentioned all the politicians asking questions than they have been about some of the other concerns. Does that give you any optimism as dark as the story is that at least some segment of the population is waking up to these risks? Well, the straight answer is just yes. I should first split out the straight answer before trying to complicate anything yes like the the broad class of things where some people have seen stuff actually happening in front of them and then started to talk in a more sensible way gave me more hope than before that happened because it wasn't previously obvious to me that this was how
Starting point is 00:42:53 things would even get a chance to play out with that said it can be a little bit difficult for me to fully model or predict how that is playing out politically because of the strange vantage point to occupy. Imagine being a sort of scientist person who is like, this asteroid is on course to hit your planet only, you know, for technical reasons, you can't actually calculate when. You just know it's going to hit sometime the next 50 years. Completely unrealistic for an actual asteroid. But say you're like, well, there's the asteroid. Here it is in our telescopes. These are how orbital mechanics work. And people are like, eh, fairy tale never happened. And then like a little tiny meteor crashes into their house. And like, oh my gosh,
Starting point is 00:43:36 now realize rocks can fall from the sky and you're like okay like that convinced you the telescope didn't convince you i can sort of see how that how that works you know people being the way they are but it's still a little weird to me and i can't call in in advance you know i don't feel like i now know how the next 10 years of politics are going to play out and wouldn't be able to tell you even if you told me which AI breakthroughs there's going to be over that time span if we even get 10 years which you know people in the industry don't seem to think so and i maybe i should believe them about that Let me throw another argument at you that I don't subscribe to myself, but I feel like maybe you would knock it down in an entertaining way.
Starting point is 00:44:15 One of the most frequent emails that we have gotten since we started talking about AI is from people who say that AI Dumerism is just hype that serves only to benefit the AI companies themselves, and they use that as a reason to dismiss existential risk. How do you talk to those folks? It's historically false. We were around before there were any. AI companies of this class to be hyped. So leaving aside the objection, it is false. What is this? Like, leaded gasoline can't possibly be a problem because this is just hyped by
Starting point is 00:44:55 the gasoline companies. Like, nuclear weapons are just like hype from the nuclear power industry so that their power plants will seem more cool. Like, let's. What matter of deranged conspiracy theory is this? It may possibly be an unpleasant fact that humanity being as completely nutball wacko as we are, that if you say that a technology is going to destroy the world, it will raise the stock prices of the companies that are bringing about the end of the world because a bunch of people think that's cool, so they buy the stock. Like, okay, but that has nothing to do with whether the stuff can actually kill you or not.
Starting point is 00:45:34 Right? Like, it could be the case that the existence of nuclear weapons raises the stock price of the worst company in the world, and it wouldn't affect any of the nuclear physics that caused nuclear weapons to be capable of killing you. This is not a science level argument. It just doesn't address the science at all. Yeah. Well, let's maybe try to end this first part of the conversation on a note of optimism. You have spent two decades building a very detailed, model of why Doom may be in our future. If you had to articulate why you might be wrong, what is the strongest case you could make? Are there any things that could happen that would sort of make your predictions not come true? So like the current AIs are not understandable, ill-controlled, that the technology is not conducive to understanding or controlling them. All of the, all of the people trying to do this are going far uphill. They are vastly behind the rate of progress and capabilities. Like, what does it take to believe that an alchemist can actually
Starting point is 00:46:40 successfully concoct you an immortality potion? It's not that immortality potions are impossible in principle. With sufficiently advanced biotechnology, you could do it. But in the medieval world, what are you supposed to see that, to make you believe that the guy is going to have an immortality potion for you, short of him actually pulling that off in real life? Right. You know, like, no amount of, like, look at how I melted this gold is going to get you to expecting the guy to transmute lead into gold until he actually. actually pulls that off. And it's like, it's like some kind of AI breakthrough, which doesn't raise capabilities to the point where it ends the world. But suddenly, like, the AI's thought
Starting point is 00:47:15 processes are completely understandable and completely controllable. And there's like none of these issues. And people can specify exactly what the AI wants in super fine detail and get what they want every time. And, you know, they can read the AI's thoughts and there's no sign whatsoever that the AI is plotting against you. And then the AI, like, lays out this compact control scheme for you know, building the AI that that's going to give you the immortality potion. It's, we're just so far off. You're asking me that there isn't some kind of like clever little objection that can be cleverly refuted here.
Starting point is 00:47:49 This is something that is just like way the heck out of reach. As soon as you try to think about it seriously, what does it actually take to build the superintelligence? What does it actually take to control it? What does it take to have that not go wrong, you know, on the first serious load? when the thing is like smarter than you when you're into the regime where failures will kill you
Starting point is 00:48:07 and therefore are not observable anymore because you're dead you don't get to observe it like what does it take to do that in real life there isn't some kind of cute experimental result we can see tomorrow that makes this go well all right well for the record
Starting point is 00:48:21 I did try to end this segment on a note of optimism but I appreciate that your feelings are deeply held here today Casey but I admire you try Well, let's take a break, and when we come back, we'll have more with L.A. Zeriotkowski. Okay, so we are back with Ellie Ezraudkowski, and I want to talk now about some of these solutions that you see here. If we are all doomed to die if and when the AI industry builds a super intelligent AI system,
Starting point is 00:49:34 what do you believe could stop that? Maybe run me through your basic proposal for what we can do to avert the apocalypse. So the materials for building the apocalypse are not all that easy to make at home. There is this one company called ASML that makes the critical set of, of machines that get used in all of the chip factories. And to grow an AI, you currently need a bunch of very expensive chips. They are custom chips built especially for growing AIs. They need to all be located in the same buildings so that they can talk to each other,
Starting point is 00:50:17 because that's what the current algorithms require. You have to build a data center. The data center uses a bunch of electricity. If this were illegal to do outside of supervision, it would not be that easy to hide. There are a bunch of differences, but nonetheless, the obvious analogy is nuclear proliferation and deep proliferation, where it's back when nuclear weapons were first invented, a bunch of people predicted that every major country was going to build a massive nuclear fleet, and then the first time there was a flashpoint, there was going to be a global nuclear war.
Starting point is 00:50:52 And this is not because they enjoyed being pessimistic. if you look at world history up to World War I, World War II, they had some reasons to be concerned. But we nonetheless managed to back off. And part of that is because it's not that easy to refine nuclear materials. The plants that do it are known and controlled. And, you know, when a new country tries to build it, it's a big international deal. I don't quite want to needlessly drench myself with current political controversies. But the point is you can't build a nuclear weapon in your backyard, and that is part of why the human species is currently still around. Well, at least with the current technology, you can't further escalate AI
Starting point is 00:51:32 capabilities very far in your backyard. You can escalate them a little in your backyard, but not a lot. So it would be, just to sort of finish the comparison to nuclear proliferation here, it would be immediate sort of moratorium on powerful AI development, along with kind of an international nuclear-style agreement between nations that would make it illegal to build data centers capable of advancing the state-of-the-art with AI? Am I hearing that right? All the AI chips go to data centers. All the data centers are under an international supervisory regime. And the thing I would recommend to that regime is to say, like, just stop escalating AI capabilities any further. We don't know and we will get into trouble. It is possible that
Starting point is 00:52:17 we can take the next step up the ladder and not die. It's possibly we can take three steps up the ladder and not die. We don't actually know. So we got to stop somewhere. Let's stop here. That's what I would tell them. And what do you do if a nation goes rogue and decides to build its own data centers and fill them with powerful chips and start training their own superhuman AI models? How do you handle that? Then that is a more serious matter than a nation refining nuclear materials with which they could build a small number of nuclear weapons. This is not like having five fission bombs to deter other nations. This is a threat of global extinction to every country on the globe. So you have your diplomats say, stop that, or else we in terror of our
Starting point is 00:53:04 lives and the lives of our children will be forced to launch a conventional strike on your data center. And then if they keep on building the data center, you launch a conventional strike on their data center because you would rather not run a risk of everybody on the planet dying. It seems kind of straightforward in a certain sense. And in a world where this came to pass, do you envision work on AI or AI-like technologies being allowed to continue in any way? Or we've just decided this is a dead end for humanity. Our tech companies will have to work on something else. I think it would be extremely sensible for humanity to declare that we should all.
Starting point is 00:53:44 just back off. Now, I personally, I look at this and I think I see some ways that you could build relatively safer systems with narrower capabilities that were just learning about medicine and didn't quite know that humans were out there, the way that current large language models are trained on the entire internet. They know that humans are out there and they talk to people and they can manipulate some people psychologically, if not others, as far as we know. So I have to be careful to distinguish my statements of factual prediction from my policy proposals. And I can say in a very firm way, if you escalate up to superintelligence, you will die. But then if you're like, well, if we try to train some AI systems just on medical stuff
Starting point is 00:54:38 and not expose them to any material that teaches them about human psychology, could we get some work out of those without everybody die. I cannot say no firmly. So now we have a policy question. Are you going to believe me when I say, I don't can't tell if this thing will kill you? Or are you going to believe somebody else who says this thing will definitely not kill you? Are you going to believe a third person who's like, yeah, I think this medical system is for sure going to kill you? Who do you believe here if you're not just going to back off of everything? So backing off of everything would be pretty sensible and trying to build narrow, medically specialized systems that are not very much deep smarter than the current systems and aren't being told that humans exist and
Starting point is 00:55:29 they're just thinking about net medicine in this very narrow way and you're not just going to keep pushing them until it explodes in your face. You're just going to like try to get some cancer cures out of it and that's it. You could maybe get away with that. I kind of actually say you're doomed for sure if you played it very cautiously. If you put the current crop of complete disaster monkeys in charge, they may manage to kill you. They're just like, they just like do so much worse than they need to do. They're just like so cavalier about it. We didn't need to have a bunch of AIs driving people insane. You can, you can train a smaller AI to look at the conversations and tell, is this AI currently in the process of taking a
Starting point is 00:56:09 vulnerable person and driving them crazy? They could have detected it earlier. They could have try to solve it earlier. So if you have these completely catholier disaster monkeys trying to run the medical AI project, they may manage to kill you. Okay, so now you have to decide you trust these guys. And that's the core dilemma there. I have to say, Eliezer, I think there is essentially zero chance of this happening, at least in today's political climate. I look at what's going on in Washington today. You've got, you know, the Trump administration wants to accelerate AI development. And NVIDIA and its lobbyists are going around Washington blaming AI doomers for trying to cut off chip sales to China. There seems to be a sort of concerted effort not to clamp down on AI, but to make it go faster.
Starting point is 00:56:57 So I just look around the political climate today and I don't see a lot of openings for a stop AI movement. But like, what do you think would have to happen in order for that to change? From my perspective, there's a, you know, sort of core factual truth here, which is, if you build superintelligence, then it kills you. And the question is just like, do people come to apprehend this thing that happens to be true? It is not in the interest of the leaders of China, nor of Russia, nor of the UK, nor of the United States, to the, along with their families. It's not actually in their interest. That's kind of the core reason why we haven't had a nuclear war, despite all the people who in 1950 were like, how on earth are we not going to have nuclear war? What country's going to turn down the military benefits
Starting point is 00:57:52 of having their own nuclear weapons? How are you not going to have somebody who's like, yeah, I've got some nuclear weapons? Let me take this little area of border country here, the same way that things have been playing out for centuries and millennia on Earth before then. But they also had like a, there were nuclear weapons dropped during World War II in Japan. And so people could look at that and see the chaos it caused and point to that and say, well, that's the outcome here. In your book, you make a different World War II analogy. You sort of compare the required effort to stop AI to the mobilization for World War II. But that was a reaction to like a clear act of war.
Starting point is 00:58:32 And so I guess I'm wondering, like, what is the equivalent? of the invasion of Poland or the bombs dropping on Hiroshima Nagasaki for AI? What is the thing that is going to spur people to pay attention? I don't know. I think that OpenAI was caught flat-footed when they first published ChatGPT, and that caused a massive shift in public opinion. I don't think OpenAI predicted that. I didn't predict it.
Starting point is 00:58:56 It could be that any number of potential events cause a shift in public opinion. We are currently getting Congresspeople writing pointed questions, in the wake of the release of an internal documented meta, which has what they call is a superintelligence lab, although I don't think they know what that word means, where it's like their internal guidelines for acceptable behavior for the AI, and it says, like, well, if you have an 11-year-old trying to flirt, flirt back, and everyone was like, what the actual censored profanity meta?
Starting point is 00:59:27 What could you possibly have been thinking? Why would you, like, how could you be so, like, why did you even, why from your own perspective did you write this town in a document? Even if you thought that was cool, you shouldn't have written it down because now they're going to be pointed questions and there were. And, you know, maybe it's something that, from my perspective, doesn't kill a bunch of people but still causes pointed questions to be asked. Or maybe there's some actual kind of catastrophe that we don't just manage to frog boil ourselves into.
Starting point is 00:59:59 You know, like losing massive numbers of kids to their AI, girlfriends and AI boyfriends, is from my perspective and an obvious sort of guess. But even the most obvious sort of guess there is still not higher than 50%. And I don't think I want to wait. Maybe chat GP, you know, from my perspective, maybe chat GPT was it, right? I was out, you know, I'm off in the wilderness. Nobody's paying attention to these issues at all because they think that it will only happen in 20 years in 2005. And that to them means the same thing is never. And then I got like the chat GPT moment. And suddenly people realize this stuff is actually going to happen to them. And that happened before the end of the world. Great. I got
Starting point is 01:00:39 a miracle. I'm not going to sit around waiting for a second miracle. If I get a second miracle, great. But meanwhile, you got to put your boots on the ground. You got to get out there. You got to do what you can. It strikes me that an asset that you have as you try to advance this idea is that a lot of people really do hate AI, right? Like, if you go on blue sky, you will see these people talking a lot about all of the different reasons that they hate AI. At the same time, they seem to be somewhat dismissive of the technology, right? Like, they have not crossed the chasm from, I hate it because I think it's stupid and it sucks to I hate it because I think it is quite dangerous. I wonder if you have thoughts on that group of folks and if you feel like
Starting point is 01:01:23 or would want them to be part of a coalition that you're building? Yeah, so, you don't want to make the coalition too narrow. I'm not a fan of Vladimir Putin, but I would not, from that, on that basis, kick him out of the, how about if humanity lives instead of dies coalition? What about people who think that AI is never going to be a threat to all humanity, but they're worried that it's going to take our jobs? Like, do they get to be in the coalition? Well, I think you've got to be careful because they believe different things about the world than you do.
Starting point is 01:02:05 And you don't want these people running the, how about if humanity does not die coalition? You want them to be in some sense like external allies because they're not there to prevent humanity from dying. And if they get to make policy, maybe they're like, well, you know, this policy would potentially allow AIs to kill everyone, according to those wacky people who think that AI will be more powerful tomorrow than it is today. But, you know, in the meanwhile, it prevents them from taking, the AIs from taking our jobs, and that's the part we care about. So, you know, like, there's this one thing that the coalition is about, and that's it.
Starting point is 01:02:38 It's just about not going extinct. Yeah. Aliasa, right now, as we're speaking, I believe there are hunger strikes going on in front of a couple of AI headquarters, including Anthropic and Google Deep Mind. These are people who want to convince these companies to shut down AI. We've also seen some of potentially violent threats made against some of these labs. And I guess I'm wondering if you worry about people committing extreme acts, be they violent or nonviolent based on your lessons from this book. I mean, if it taking some of your arguments to the natural logical conclusions of if anyone builds this,
Starting point is 01:03:23 everyone dies. I can see people rationalizing violence on that basis against some of the employees at these labs. And I worry about that. So what can you say about the sort of limits of your approach and what you want people to do when they hear what you're saying? Boy, there sure are a bunch of questions bundled together there. And so the number one thing I would say is that if you commit acts of individual violence against individual researchers at an individual AI lab in your individual country, this will not prevent everyone from dying. The problem with this logic is not that by this act of individual violence, you can save humanity, but you shouldn't do that because that would be deontologically prohibited.
Starting point is 01:04:13 I'll just like say it, say it that way. The problem is you cannot save humanity by the feudal spasms of individual violence. It's an international issue. You can be killed by a superintelligence that somebody built on the other side of the planet. I do in my personal politics tend to be libertarian. If something is just going to kill you and your voluntary customers, it's not a global issue the same way. If it's just going to kill people standing next to you, different cities can make different laws
Starting point is 01:04:40 about it. If it's going to kill people on the other side of the planet, that's, that's when you're when the international treaties come in. And a futile act of individual violence against an individual researcher and an individual AI company is probably making that international treaty less likely rather than more likely. And there's a underlying truth of moral philosophy here, which is that a bunch of our reason for our prejudice against individual murders is because of a very systematic and deep sense in which individuals, murderers tend to not solve society's problems.
Starting point is 01:05:20 And this is, from my perspective, a whole bunch of the point of having a taboo against individual murder. It's not that people go around committing individual murders, and then the world actually gets way better, and all the social problems are actually solved. But we don't want to do that. We don't want to do more of that because murder is wrong. The murders make things worse, and that's why we properly should have a taboo against it. We need international treaties here.
Starting point is 01:05:48 What do you make of the opposition movement to the movement that you're sketching out here? Mark Andresen, the powerful venture capitalist, very influential in today's Trump administration, has written about the views that you and others hold that he thinks are unscientific. He thinks that AI risk has turned into an apocalypse cult. And he says that their extreme beliefs should not determine the future of laws and society. So I guess I'm interested in your sort of reaction to that quote specifically, but I also wonder how you plan to engage with the people on the other side of this argument. Well, it is not uncommon in the history of science for the cigarette companies to smoke their own tobacco, the inventor of leaded gasoline, who was a great advocate of the safety of leaded gasoline, despite the many reasons why he should have known better. I think, like, did actually get sufficient cumulative lead exposure himself that he, you know, had to go off to a sanitarium for a few years and then came back and started exposing himself to let again and got sick again.
Starting point is 01:06:59 And so sometimes these people truly do believe they do drink their own Kool-Aid even to the point of death shows history. And perhaps Mark Andresen will continue to drink his own Kool-Aid even to the point of death. And if he were just killing himself, that would be one thing I say is. libertarian, but he's unfortunately also going to kill you. And the thing I would say to sort of refute the central argument is, what's the plan? What's the design for this bridge that is going to hold up when the whole weight of the entire human species has to march across it? Where is the design scheme for this airplane, to which we are going to load the entire human species into its cargo hold and fly it and not crash? What's the plan? Where's the sign?
Starting point is 01:07:45 What's the technology? Why is it not working already? And they just don't, they can't make the case for this stuff being, you know, like not perfectly safe, but even remotely safe, that they're going to be able to control their superintelligence at all. So they go into these like, you must not listen to these dangerous apocalyptic people because they cannot engage with us on the field of the technical arguments. They know they will be routed. you have advice in your book for journalists politicians who are worried about some of the catastrophes you see coming for people who are not in any of those categories for our listeners
Starting point is 01:08:29 who are just out there living their daily lives maybe using chat GPT for something uh helpful in their daily life what can they do if they're worried about where all this is heading well as of a year ago, I'd have said, you know, again, it's, you know, write, right to your elected representatives. Talk to your friends about, you know, being ready to vote that way if a disputed primary election comes down that way. The ask, I would say, is for our leaders to begin by saying, we are open to a worldwide AI control treaty if others are open to the same. Like, we are ready to back off, if other countries back off. We are ready to participate in international treaty about this. Because if you've got multiple leaders of great powers saying that, well, maybe there can be a
Starting point is 01:09:18 treaty. So that's kind of the next step from there. That's the political goal we have. If you're having trouble sleeping and if you're generally in a distressed state, like maybe don't talk to some of the modern AI systems because they might drive you crazy as I think I would say now. I didn't have to say that one year earlier. You know, like this, the whole AI boyfriend, AI girlfriend thing might not be good for you. Maybe don't go down that road, even if you're lonely. I, you know, but that's individual advice. That's not going to protect the planet.
Starting point is 01:09:50 Yeah. Well, I'll end this conversation where I've ended some of our earlier conversations, Eliezer, which is I really appreciate the time, and I really hope you're wrong. Like, that would be great. We all hope I'm wrong. I hope I'm wrong. My friends hope I'm wrong. Everybody hopes I'm wrong.
Starting point is 01:10:08 Hope is, does not, you know, hope is not what saves us in the end. Action is what saves us. Hope is not, you know, hoping for miracles, hoping, you know, leaded gasoline. You can't just hope that leaded gasoline isn't going to poison people. You actually got to ban the leaded gasoline. So more active hopes. I'm in favor. Like, I see the hope.
Starting point is 01:10:31 I share the hope. But let's look for more activist hopes than that. Yeah. Well, the book is, if anyone builds it, everyone dies. why superhuman AI would kill us all and it is coming out soon and it is a co-written book by Eliezer and his co-author Nate Sorries.
Starting point is 01:10:48 Elazer, thank you. Thanks, Al-Azer. That's well. Hard Fork is produced by Whitney Jones and Rachel Cohn. We're edited by Jen Poyant. We're fact-checked this week by Will Pyshal. Today's show was engineered by Katie McMurran. Original music by Rowan Nemistow, Alyssa Mox.
Starting point is 01:11:38 and Dan Powell. Video production by Sawyer Roque, Pat Gunther, Jake Nickel, and Chris Schott. You can watch this full episode on YouTube at YouTube.com slash hardfork. Special thanks to Paula Schumann,
Starting point is 01:11:51 Hui Wing, Tam, Dahlia Hadad, and Jeffrey Miranda. As always, you can email us at hardfork at n.Ytimes.com. Send us your plans for the AI apocalypse. You know,

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.