Hard Fork - Anthropic’s C.E.O. Dario Amodei on Surviving the A.I. Endgame

Episode Date: February 28, 2025

This week, Anthropic’s C.E.O., Dario Amodei, returns to the show for a candid, wide-ranging interview. We discuss Anthropic’s brand-new Claude 3.7 Sonnet model, the A.I. arms race against China, a...nd his hopes and fears for this technology over the next two years. Then, we gather up recent tech stories, put them into a hat and close out the week with a round of HatGPT. Guest:Dario Amodei, chief executive of Anthropic Additional Reading:Inside the White-Hot Center of A.I. DoomerismMachines of Loving GraceApple Removes Cloud Encryption Feature From U.K. After Backdoor Order We want to hear from you. Email us at hardfork@nytimes.com. Find “Hard Fork” on YouTube and TikTok. Unlock full access to New York Times podcasts and explore everything from politics to pop culture. Subscribe today at nytimes.com/podcasts or on Apple Podcasts and Spotify.

Transcript
Discussion (0)
Starting point is 00:00:00 So I went to two AI events this weekend. It was sort of polar opposites of the AI spectrum. There was the effective altruists had their big annual conference. Yeah. Then on Friday night, I went out. You'd be very proud of me. I stayed out so late. I stayed until 2 AM. Oh my.
Starting point is 00:00:18 I went to an AI rave that was sort of unofficially affiliated with Mark Zuckerberg. It was called the Zuck rave that was sort of unofficially affiliated with Mark Zuckerberg. It was called the Zuck rave. Now, when you say unofficially affiliated, Mark Zuckerberg had no involvement in this, and my assumption is he did not know what was happening. Correct, a better word for what his involvement is
Starting point is 00:00:37 would be no involvement. He was sort of a tribute rave to Mark Zuckerberg thrown by a bunch of accelerationist people who want AI to go very fast. Another word for it would be using his likeness without permission. Yes. But that happens to famous people sometimes. Yes, so at the Zuck rave, I would say
Starting point is 00:00:54 there was not much raving going on. No? There was a dance floor, but it was very sparsely populated. They did have like a thing there that would like, it had a camera pointing at the dance floor and if you sort of stood in the right place, it would turn your face into Mark Zuckerberg's, like on a big screen.
Starting point is 00:01:10 So that was... Wait, so let's just say it's not something you want to happen to you while you're on mushrooms. Because that could be a very destabilizing event. Yes, there was a train, an indoor, like, toy train that you could ride on. It was going actually quite fast. And...
Starting point is 00:01:24 What was the point of this rave? To do drugs. That was going actually quite fast. And, um... What was the point of this rave? Uh, to do drugs. That was the point of this rave. -♪ I'm Kevin Roos, a tech columnist at The New York Times. I'm Casey Noonan from Platformer. And this is Hard Fork.
Starting point is 00:01:41 This week, Anthropic CEO Dario Amadei returns to the show for a supersized interview about the new Claude, the AI race against China, and his hopes and fears for the future of AI. Then we close it out with a round of Hatch GPT. Big show this week, Evan. KC, have you noticed that the AI companies do stuff on the weekends now? Yeah, whatever happened to just five days a week?
Starting point is 00:02:09 Yes, they are not respectful of reporters and their work hours. Companies are always announcing stuff on Saturdays and Sundays and different time zones. It's a big pain. It really is. But this weekend, I got an exciting message on Sunday, saying that Dario Amadei, the CEO of Anthropic, had some news to talk about, and he wanted to come on Hard Fork to do it. Yeah, and around the same time,
Starting point is 00:02:31 I got an email from Anthropic telling me I could preview their latest model, and so I spent the weekend actually trying it out. Yeah, so long-time listeners will remember that Dario is a repeat guest on this show. Back in 2023, we had him on to talk about his work at Anthropic and his vision of AI safety and where all of this was headed. And I was really excited to talk to him again for a few reasons.
Starting point is 00:02:54 One, I just think he's like a very interesting and thoughtful guy. He's been thinking about AI for longer than almost anyone. He was writing papers about potentially scary things in AI safety. All the way back in 2016, he's been at Google, he's been at OpenAI, he's now the CEO of Anthropic. So he is really the ultimate insider when it comes to AI. And you know, Kevin, I think Dario is an important figure
Starting point is 00:03:19 for another reason, which is that of all of the folks leading the big AI labs, he is the one who seems the most publicly worried about the things that could go wrong. That's been the case with him for a long time. And yet over the past several months, as we've noted on the show, it feels like the pendulum has really swung away from caring about AI safety to just this sort of go, go, go accelerationism that was embodied by the speech that Vice President JD Vance gave in France the other day. And for that reason, I think it's important to bring him in here and maybe see if we can shift that pendulum back a little bit and remind folks of what's at stake here. Yeah, or at least get his take on the pendulum swinging and why he thinks
Starting point is 00:03:58 it may swing back in the future. So today, we're going to talk to Dario about the new model that Anthropic just released, Claude 3.7 Sonnet, but we also want to have a broader conversation because there's just so much going on in AI right now. And Kevin, something else that we should note, something that is true of Dario this time that was not true the last time that he came on the show is that my boyfriend now works at his company. Yeah, Casey's manthropic, is it anthropic? My manthropic is that, anthropic? And I have a whole sort of long disclosure about this that you can read at platformer.news slash ethics
Starting point is 00:04:35 might be worth doing this week. You know, we always like reminding folks of that. Yep, all right, with that, let's bring in Dario Amadei. Dario Amadei, welcome back to Hard Fork. Thank you for having me again. Returning champion. So tell us about Claude 3.7. Tell us about this new model. Yes. So we've been working on this model for a while. We basically had in mind two things.
Starting point is 00:05:08 One was that, of course, there are these reasoning models out there that have been out there for a few months, and we wanted to make one of our own, but we wanted the focus to be a little bit different. In particular, a lot of the other reasoning models in the market are trained primarily on math and competition coding,
Starting point is 00:05:23 which are their objective tasks where you can measure which are, you know, they're objective tasks where you can measure performance. I'm not saying they're not impressive, but they're sometimes less relevant to tasks in the real world or the economy. Even within coding, there's really a difference between competition coding and doing something in the real world. And so we train Claude 3.7, you know, more to focus on these real-world tasks. We also felt like it was a bit weird that in the reasoning models that folks have offered,
Starting point is 00:05:51 it's generally been there's a regular model and then there's a reasoning model. This would be like if a human had two brains and it's like, you can talk to brain number one if you're asking me a quick question, like what's your name, and you're talking to brain number one if you're asking me a quick question, like, what's your name? And you're talking to brain number two if you're asking me to prove a mathematical theorem because I have to sit down for 20 minutes. It'd be like a podcast where there's two hosts, one of whom just likes to yap and one of whom actually thinks before he talks. Come on.
Starting point is 00:06:16 Brutal. No comment on any relevance to it. So what differences will users of Cloud notice when they start using 3.7 compared to previous models? Yes. So, a few things. It's going to be better in general, including better at coding, which Cloud models have always been the best at coding,
Starting point is 00:06:36 but 3.7 took a further step up. In addition to just the properties, the model itself, you can put it in this extended thinking mode where you tell it basically the same model but you're just saying operating in a way where you can think for longer. And if you're an API user, you can even say, here's the boundary in how long you can think. And just to clarify, because this may confuse some people, what you're saying is the sort of new Claude is this hybrid model, It can sometimes do reasoning, sometimes do quicker answers. But if you want it to think for even longer,
Starting point is 00:07:07 that is a separate mode. That is a separate mode. Thinking and reasoning are sort of separate modes. Yes, yes. So basically the model can just answer as it normally would, or you can give it this indication that it should think for longer. And even further direction of the evolution
Starting point is 00:07:21 would be the model decides for itself what the appropriate time to think is, right? Humans are like that, or at least can be like that, right? If I ask you your name, you know, you're not like, huh, how long should I think of it? Give me 20 minutes, right, to, you know, to determine my name. But if I say, hey, you know, I'd like you to do an analysis of, you know, this stock, or I'd like you to, you know, prove this mathematical theorem, humans who are able to do that task,
Starting point is 00:07:48 they're not going to try and give an answer right away. They're going to say, okay, well, that's going to take a while and then we'll need to write down the task and then give an answer. This is one of my main beefs with today's language models and AI models in general, is I'll be using something like ChatGBT and I'll forget that I'm in the hardcore reasoning mode and I'll ask it some stupid question, how do I change the settings on my water heater?
Starting point is 00:08:10 It'll go off and think for four minutes and I'm like, I didn't actually need to do that. It'll be like a treatise on adjusting the temperature of the water heater. It's like water heater. Consideration one. So how long do you think it'll be before the models can actually do that routing of routing themselves,
Starting point is 00:08:25 where you'll ask a question and say, it seems like you need about a three minute long thinking process for this one versus maybe a 30 second one for this other thing? Yeah. So I think our model is kind of a step towards this. Even in the API, if you give it a bound on thinking, you say, I'm going to think for 20,000 words or something.
Starting point is 00:08:43 On average, when you give it up to 20,000 words, most of the time it doesn't even use 20,000 words or something. On average, when you give it up to 20,000 words, most of the time it doesn't even use 20,000 words. And sometimes it'll give a very short response. Because when it knows that it doesn't get any gain out of thinking further, it doesn't think for longer. But it's still valuable to give a bound on how long it'll think. So we've kind of taken a big step in that direction,
Starting point is 00:09:01 but we're not to where we want to be yet. When you say it's better at real world tasks, what are some of the tasks that you're thinking of? Yeah, so I think above all coding, quad models have been very good for real world coding. We have a number of customers from Cursor to GitHub, to Windsurf Codium, to Cognition, to Versel, to I'm sure I'm leaving some out there.
Starting point is 00:09:23 These are the vibe coding apps. Or just the coding apps period. The coding apps period. There are many different coding apps. We also released this thing called quad code which is more of a command line tool. But I think also on things like complex instruction following or just like here,
Starting point is 00:09:40 I want you to understand this document or I want you to use this series of tools. The reasoning model that we've trained, Claude 3.7, Sonnet, is better at those tasks, too. One thing the new Claude Sonnet is not doing, Dario, is accessing the internet. Yes. Why not? And what would cause you to change that? Yes.
Starting point is 00:10:01 So I think I'm on record saying this before but web search is coming very soon. Okay. We will have web search very soon. We recognize that as an oversight. You know, I think in general we tend to be more enterprise focused and consumer focused and this is more of a consumer feature although it can be used on both but you know we focus on both and this is coming. Got it.
Starting point is 00:10:21 So you've named this model 3.7. The previous model was 3.5. You quietly updated it last year and Insiders were named this model 3.7. The previous model was 3.5 You quietly updated it last year and insiders were calling that one 3.6 respectfully. This is driving all of us insane What is going on with AI model names? We are the least insane I recognize that we are insane So so look, I think our mistakes are relatively understandable You know, we made a 3.5 Sonnet. We were doing well in there.
Starting point is 00:10:45 We had the 3.0s and then the 3.5s. I recognize the 3.7 new was a misstep. It actually turns out to be hard to change the name in the API, especially when there's all these partners and surfaces you offer it to. You can figure it out. I believe in you. No, no, no.
Starting point is 00:11:01 It's harder than trading the model, I'm telling you. So we've kind of retroactively and informally named the last one 3.6, so that it makes sense that this one is 3.7. And we are reserving quad four, sonnet, and maybe some other models in the sequence for things that are really quite substantial leaps. Sometimes when the models- Those models are coming, by the way.
Starting point is 00:11:23 Okay, got it, coming when? Yeah, so I should talk a little bit about this. So all the models we've released so far are actually not that expensive, right? You know, I did this blog post where I said they're in the few tens of millions of dollars range at most. There are bigger models and they are coming. They take a long time and sometimes they take a long time to get right. But those bigger models, they're coming for others. I mean, they're rumored to be coming from competitors as well, but we are not too far away from releasing
Starting point is 00:11:55 a model that's a bigger base model. So most of the improvements in Cloud 3.7 Sonnet as well as Cloud 3.6 Sonnet are in the post-training phase. But we are working on stronger base models and perhaps that'll be the Cloud 4 series, perhaps not, we'll see. But I think those are coming in relatively small number of time units. Small number of time units. I'll put that on my calendar. Remind me to check in on that in a few time units, Kevin. I know you all at Anthropic are very concerned about AI safety
Starting point is 00:12:30 and the safety of the models that you're putting out into the world. I know you spend lots of time thinking about that and red teaming the models internally. Are there any new capabilities that Claw 3.7 Sonnet has that are dangerous or that might worry someone who is concerned about AI safety? So not dangerous per se, and I always want to be clear about this
Starting point is 00:12:51 because I feel like there's this constant conflation of present dangers with future dangers. It's not that there aren't present dangers and there are always kind of normal tech risks, normal tech policy issues. I'm more worried about the dangers that we're going to see as models become more powerful. And I think those dangers, you know, when we talked in 2023, I talked about them a lot.
Starting point is 00:13:13 You know, I think I said, I even testified in front of the Senate for things like, you know, misuse risks with, for example, biological or chemical warfare or the AI autonomy risks. I said, particularly with the misuse risks, I said, I don't know when these are going to be here, when these are going to be real risks, but it might happen in 2025 or 2026. Now that we're in early 2025, the very beginning of that period,
Starting point is 00:13:37 I think the models are starting to get closer to that. So in particular, in Claude 3.7 Sonnet, as we wrote in the model card, we always do these, you could almost call them like, you know, trials, like trials with a control, where we have, you know, some human who doesn't know enough much about some area like biology. And we basically see how much does the model help them to engage in some mock bad workflow? We'll change a couple of the steps, but some mock bad workflow. How good is a human at that assisted by the model?
Starting point is 00:14:11 Sometimes we even do wet lab trials in the real world where they mock make something bad as compared to the current technological environment. What they could do on Google or with a textbook or just what they could do unaided. And we're trying to get at, does this enable some new threat vector that wasn't there before? I think it's very important to say this isn't about like,
Starting point is 00:14:35 oh, did the model give me the sequence for this thing? Did it give me a cookbook for making math or something? That's easy. You can do that with Google. We don't care about that at all. We care about this kind of esoteric, high, uncommon knowledge that say only a virology PhD or something has. How much does it help with that?
Starting point is 00:14:54 And if it does, that doesn't mean we're all gonna die of the plague tomorrow. It means that a new risk exists in the world, a new threat vector exists in the world, as if you just made it easier to build a nuclear weapon. You invented something that the amount of plutonium you needed was lower than it was before. And so we measured sonnet 3.7 for these risks, and the models are getting better at this.
Starting point is 00:15:18 They're not yet at the stage where we think that there is a real and meaningful increase in the threat end to end, to do all the tasks you need to do, to really do something dangerous. However, we said in the model card that we assessed a substantial probability that the next model or a model over the next, I don't know, three months, six months,
Starting point is 00:15:42 a substantial probability that we could be there. Then our safety procedure, our don't know, three months, six months, a substantial probability that, you know, that we could be there. And then our safety procedure, our responsible scaling procedure, which is focused mainly on these, you know, very large risks, would then kick in and, you know, we'd have kind of additional security measures and additional deployment measures, you know, designed, you know, particularly against these very narrow risks. Yeah. I mean, just to really underline that, you're saying in the next three to six months, we are going to be in a place of medium risk in these models, period. Presumably, if you are in that place, a lot of your competitors are also going to be in that place. What does
Starting point is 00:16:17 that mean practically? Like, what does the world need to do if we're all going to be living in medium risk? I think at least at this stage, it's not a huge change to things. It means that there's a narrow set of things that models are capable of, if not mitigated, that, you know, would somewhat increase the risk of something like really dangerous or really bad happening, you know, like put yourself in the eyes of like a law enforcement officer, or, you know, the FBI or something, there's like a law enforcement officer or you know the FBI or something there's like a new threat factor there's a new kind of attack. You know it doesn't mean the end of the
Starting point is 00:16:50 world but it does mean that anyone anyone who's involved in industries where this risk exists should take a precaution against that risk in particular. Got it. And so I don't know I mean I don't you know I could be wrong it could take much longer you can't predict what's gonna happen. But I think contrary to the environment that we're seeing today of worrying less about the risks, the risks in the background have actually been increasing. We have a bunch more safety questions,
Starting point is 00:17:18 but I wanna ask two more about kind of innovation competition first. Right now, it seems like no matter how innovative any given company's model is, those innovations are copied by rivals within months or even weeks. Does that make your job harder and do you think it is going to be the case indefinitely? I don't know that innovations are necessarily copied exactly. What I would say is that the pace of innovation among a large number of competitors is very fast. There's four or five,
Starting point is 00:17:47 maybe six companies who are innovating very quickly and producing models very quickly. But if you look, for example, at Sonnet 3.7, the way we did the reasoning models is different from what was done by competitors. The things we emphasized were different. Even before then, the things Sonnet 3.5 is good at are different than the things other models are good at.
Starting point is 00:18:07 People often talk about competition, commoditization, costs going down. But the reality of it is that the models are actually relatively different from each other and that creates differentiation. Yeah. I mean, we get a lot of questions from listeners about if I'm going to subscribe to one AI tool, what should it be? These are the things that I use it for.
Starting point is 00:18:26 And I have a hard time answering that because I find for most use cases, the models all do a relatively decent job of answering the questions. It really comes down to things like which model's personality do you like more? Do you think that people will choose AI models, consumers, on the basis of capabilities? Or is it going to be more about personality and how it makes them feel, how it interacts with them? I think it depends which consumers you mean. Even among consumers, there are people who use
Starting point is 00:18:55 the models for tasks that are complex in some way. There are folks who are independent, who want to analyze data, that's maybe the prosumer side of things, right? And I think within that, there's a lot to go in terms of capabilities. The models can be so much better than they are at helping you with
Starting point is 00:19:14 anything that's focused on kind of productivity, or even a complex task like planning the trip. Even outside that, if you're just trying to make a personal assistant to manage your life or something, we're pretty far from that. From a model that sees every aspect of your life and is able to holistically give you advice and be a helpful assistant to you. I think there's differentiation within that.
Starting point is 00:19:38 The best assistant for me might not be the best assistant for some other person. I think one area where the models will be good enough is, if you're just trying to use this as a replacement for Google search or as a quick information retrieval, which I think is what's being used by the mass market, free use, hundreds of millions of users, I think that's very commoditizable.
Starting point is 00:20:03 I think the models are already of already there and are just diffusing through the world, but I don't think those are the interesting uses in the model, and I'm actually not sure a lot of the economic value is there. I mean, is what I'm hearing that if and when you developed an agent that is, let's say, a really amazing personal assistant, the company that figures out that first is going to have a big advantage because other labs are going to just have a harder time copying that.
Starting point is 00:20:28 It's going to be less obvious to them how to recreate that. It's going to be less obvious how to recreate it. And when they do recreate it, they won't recreate it exactly. They'll do it their own way in their own style, and it'll be suitable for a different set of people. So I guess I'm saying there's, the market is more segmented than you think it is. It looks like it's all one thing, but it's more segmented than you think it is.
Starting point is 00:20:49 Got it. So let me ask the competition question that brings us into safety. You recently wrote a really interesting post about DeepSeek sort of at the height of DeepSeek mania. And you were arguing in part that the cost reductions that they had figured out were basically in line with what, they were basically in line with how cost had already been falling.
Starting point is 00:21:06 But you also said that DeepSeek should be a wake up call because it showed that China is keeping pace with frontier labs in a way that the country hadn't been up until now. So why is that notable to you and what do you think we ought to do about it? Yeah. So I think this is less about commercial competition, right? I worry less about DeepSeek from a commercial competition perspective. I worry more about them from a national competition and national security perspective. I think where I'm coming from here is, you know, we look at the state of the world and,
Starting point is 00:21:40 you know, we have these autocracies like China and Russia. And I've always worried, I've worried maybe for a decade, that AI could be an engine of autocracy. If you think about repressive governments, the limits to how repressive they can be are generally set by what they can get their enforcers, their human enforcers to do. But if their enforcers are no longer human,
Starting point is 00:22:03 that starts painting some very dark possibilities. And so, you know, this is an area that I'm therefore very concerned about, where, you know, I want to make sure that liberal democracies have enough leverage and enough advantage in the technology that they can prevent some of these abuses from happening, and kind of you also prevent our adversaries from putting us in a bad position with respect to the rest of the world or even threatening our security. There's this kind of, I think, weird and awkward feature that it's companies in the US that
Starting point is 00:22:39 are building this, it's companies in China that are building this, but we shouldn't be naive. Whatever the intention of those companies, particularly in China, there's a governmental component to this. And so I'm interested in making sure that the autocratic countries don't get ahead from a military perspective. I'm not trying to deny them the benefits of the technology. They're enormous health benefits that all of us, I want to make sure, make their way everywhere in the world, including the poorest areas, including areas that are under the grip of autocracies.
Starting point is 00:23:13 But I don't want the autocratic governments to have a military advantage. And so things like the export controls, which I discussed in that post, are one of the things we can do to prevent that. I was heartened to see that actually the Trump administration is considering tightening, tightening the export controls. I was at an AI safety conference last weekend, and one of the critiques I heard some folks in
Starting point is 00:23:36 that universe make of Anthropic and maybe of you in particular, was that they saw the posts like the one you wrote about DeepSeek as effectively promoting this AI arms race with China, insisting that America has to be the first to reach powerful AGI or else. And they worry that some corners might get cut along the way, that there are some risks associated with accelerating this race in general. What's your response to that? Yeah, I kind of view things differently.
Starting point is 00:24:03 So my view is that if we want to have any chance at all, so the default state of nature is that things go at maximum speed. If we want to have any chance at all to not go at maximum speed, the way the plan works is the following. Within the US or within democratic countries, these are all countries that are under the rule of law, more or less. Therefore, we can pass laws,
Starting point is 00:24:29 we can get companies to make agreements with the government that are enforceable or make safety commitments that are enforceable. So if we have a world where there's these different companies and they're in the default state of nature would race as fast as possible. Through some mixture of voluntary convince and laws, we can get ourselves to slow down if the models are too dangerous. That's actually enforceable. You can get everyone to cooperate in the prisoner's dilemma if you just point a gun at everyone's
Starting point is 00:25:02 head. You can. That's what the law ultimately is. But I think that all gets thrown out the window in the world of international competition. There was no one with the authority to enforce any agreement between the US and China, even if one were to be made. And so my worry is, if the US is a couple years ahead of China, we can use that couple years to make things safe.
Starting point is 00:25:25 If we're even with China, you know, there's no promoting an arms race. That's what's going to happen. The technology has immense military value. Whatever people say now, whatever nice words they say about cooperation, I just don't see how once people fully understand the economic and military value of the technology, which I think they mostly already do, I don't see any way that it turns into anything other than the most intense race.
Starting point is 00:25:52 And so what I can think of to try and give us more time is if we can slow down the authoritarians, that it almost obviates the trade-off. It gives us more time to work out among us, among OpenAI, among Google, among X.AI how to make these models safe. Now, could at some point we convince authoritarians, convince for example,
Starting point is 00:26:17 the Chinese that the models are actually dangerous and that we should have some agreement and come up with some way of enforcing it. I think we should actually try to do that as well. I'm supportive of trying to do that, but it cannot be the plan A. It's just not a realistic way of looking at the world. These seem like really important questions and discussions,
Starting point is 00:26:39 and it seems like they were mostly not being had at the AI Action Summit in Paris that you and Kevin attended a couple of weeks back. What the heck was going on with that summit? Yeah, I mean, you know, I have to tell you I was deeply disappointed in the summit. It had the environment of a trade show and was very much out of spirit with the spirit of the original summit that was created in, you know, in Bletchley Park by the UK government. Bletchley did a great job and the UK government did a great job in, you know, in Bletchley Park by the UK government. Bletchley did a great job, and the UK government did a great job, where, you know, they didn't introduce a bunch of
Starting point is 00:27:11 onerous regulations, certainly before they knew what they were doing, but they said, hey, let's convene these summits to discuss the risks. I thought that was very good. I think that's gone by the wayside now, and it's, you know, it's part of maybe a general move towards less worrying about risk, more wanting to seize the opportunities. I'm a fan of seizing the opportunities. I wrote this essay, Machines of Loving Grace,
Starting point is 00:27:34 about all the great things. Part of that essay was like, man, for someone who worries about risk, I feel like I have a better vision of the benefits than a lot of people who spend all their time talking about the benefits. But in the background, like I said, as vision of the benefits than a lot of people who spend all their time talking about the benefits. But in the background, like I said, as the models have gotten more powerful, the amazing and wondrous things that we can do with them have increased, but also the risks have increased.
Starting point is 00:27:57 And that kind of secular increase, that smooth exponential, it doesn't pay any attention to societal trends or the political wins. The risk is increasing up to some critical point, whether you're paying attention or not. It was small and increasing when there was this frenzy around AI risk and everyone was posting about it, and there were these summits, and now the winds have gone in the other direction
Starting point is 00:28:25 But the exponential just continues on it. Yeah here I had a conversation with someone in Paris who was saying like it just didn't feel like anyone there was feeling the AGI by which they meant like Politicians the people doing these panels and and you know gatherings We're all talking about AI as if it were just like another technology Maybe something on the order of the PC or possibly even the Internet, but not really understanding
Starting point is 00:28:50 the exponentials that you're talking about. Did it feel like that to you and what do you think can be done to bridge that gap? Yeah. I think it did feel like that to me. The thing I've started to tell people that I think maybe gets people to pay attention is, look, if you're a public official, if you're a leader at a company,
Starting point is 00:29:14 people are going to look back. They're going to look back in 2026 and 2027. They're going to look back when hopefully humanity gets through this crazy, crazy period and we're in a mature, post-powerful AI society where we've learned to coexist with these powerful intelligences in a flourishing society. Everyone's going to look back and they're going to say,
Starting point is 00:29:36 so what did the officials, what did the company people, what did the political system do? Probably your number one goal is don't look like a fool. I've just been encouraged, be careful what you say. don't look like a fool. And so I've just been encouraged, like, don't be careful what you say. Don't look like a fool in retrospect. And, you know, a lot of my thinking is just driven by, like, you know, aside from just wanting the right outcome, like, I don't want to look like a fool.
Starting point is 00:29:55 And, you know, I think at that conference, like, you know, some people are going to look like fools. We're going to take a short break. When we come back, we'll talk with Dario about how people should prepare for what's coming in AI. You know, you talk to folks who live in San Francisco and there's like this bone deep feeling that like within, you know, a year or two years, we're just going to be living in a world that has been transformed by AI. I'm just struck by like the geographic difference because you go like, I don't know, a hundred miles in any direction and like that belief totally dissipates. And I have to say, as a journalist, that makes me bring my own skepticism and go like, I don't know, a hundred miles in any direction and like that belief totally dissipates. And I have to say as a journalist, that makes me bring my own skepticism and say like, can
Starting point is 00:30:49 I really trust all the people around me? Because it seems like the rest of the world has a very different vision of how this is going to go. I'm curious what you make of that kind of geographic disconnect. Yeah. So I've been, I've been watching this for, you know, 10 years, right? I've been in the field for 10 years and was kind of interested in AI even before then.
Starting point is 00:31:09 And my view at almost every stage up to the last few months has been, we're in this awkward space where in a few years we could have these models that do everything humans do and they totally turn the economy and what it means to be human upside down, or the trend could stop and all of it could sound completely silly.
Starting point is 00:31:27 I've now probably increased my confidence that we are actually in the world where things are gonna happen. I'd give numbers more like 70 and 80% and less like 40 or 50%. Which is not certain. Sorry, because 70 to 80% probability of what? That we'll get a very large number of AI systems that are
Starting point is 00:31:45 much smarter than humans at almost everything. Maybe 70 80% we get that before the end of the decade and my guess is 2026 2027. But on your point about the geographic difference, a thing I've noticed is with each step in the exponential, there's this expanding circle of people who kind of, depending on your perspective, are either diluted cultists or grok the future. Got it. And I remember when it was a few thousand people, right? When, you know, you would just talk to like super weird people who, you know,
Starting point is 00:32:19 believe, and basically no one else did. Now it's more like a few million people out of a few billion. And yes, many of them are located in San Francisco. But also, there were a small number of people in say the Biden administration, there may be a small number of people in this administration who believed this and it drove their policy. So it's not entirely geographic, but I think there is this disconnect.
Starting point is 00:32:40 And I don't know how to go from a few million to everyone in the world, right? To the Congress person who doesn't focus on this issues, let alone the person in Louisiana, let alone the person in Kenya. Right. It seems like it's also become polarized in a way that may hurt that goal. Like, I'm feeling this sort of alignment happening where like caring about AI safety, talking about AI safety, talking about AI safety,
Starting point is 00:33:05 talking about the potential for misuse as being coded as left or liberal, and talking about acceleration and getting rid of regulations and going as fast as possible, being coded as right. So I don't know, do you see that as a barrier to getting people to understand what's going on? Yeah, I think that's actually a big barrier,
Starting point is 00:33:23 because addressing the risks while maximizing the benefits, I think that requires nuance. You can actually have both. There are ways to surgically and carefully address the risks without slowing down the benefits very much, if at all. But they require subtlety and they require a complex conversation. Once things get polarized, once it's like we're going to cheer for
Starting point is 00:33:49 this set of words and boo for that set of words, nothing good gets done. Look, bringing AI benefits to everyone, like curing previously incurable diseases, that's not a partisan issue. The left shouldn't be against it. Preventing AI systems from being misused for weapons of mass destruction or behaving autonomously
Starting point is 00:34:12 in ways that threaten infrastructure or even threaten humanity itself, that isn't something the right should be against. I don't know what to say other than that we need to sit down and we need to have an adult conversation about this that's not tied into these same old tired political fights. It's so interesting to me, Kevin, because historically national security, national defense, nothing has been more right coded than those issues, right? But right now it seems like the right is not interested in those with respect to AI.
Starting point is 00:34:40 And I wonder if the reason, and I feel like I sort of heard this in JD Vance's speech in France, was the idea that, well, look, America will get there first and then it will just win forever, and so we don't need to address any of these. That sound right to you? Yeah.
Starting point is 00:34:53 No, I think that's it, and I think there's also, if you talk to the Doge folks, there's this sense that all these- Are you talking to the Doge folks? I'm not telling you who I'm talking to. Let's just say I've been getting some signal messages. I think there's a sense among a lot of Republicans and Trump world folks in DC that the conversation about AI
Starting point is 00:35:12 and AI futures has been sort of dominated by these worry warts, these sort of chicken little, sky is falling dooms who just are constantly telling us how dangerous this stuff is and are constantly constantly just having to push out their timelines for when it's going to get really bad, and it's just around the corner, and so we need all this regulation now. They're just very cynical.
Starting point is 00:35:34 I don't think they believe that people like you are sincere in your worry. I think on the side of risks, I often feel that the advocates of risk are sometimes the worst enemies of the cause of risk. There's been a lot of noise out there. There's been a lot of folks saying, oh, look, you can download the smallpox virus because they think that that's a way of driving, you know,
Starting point is 00:35:57 political interest. And then of course, the other side recognized that and they said, this is this is dishonest, that you can just get this on Google. Who cares about this? And so poorly presented evidence of risk is actually the worst enemy of mitigating risk. And we need to be really careful in the evidence we present. And in terms of what we're seeing in our own model,
Starting point is 00:36:19 we're going to be really careful. Like if we really declare that a risk is present now, we're going to come with the receipts. I and Tropic will try to be responsible in the claims that we make. We will tell you when there is danger, imminently. We have not warned of imminent danger yet. Some folks wonder whether a reason that people do not take questions about AI safety, maybe as seriously as they should, is that so much of what they see right now seems very silly. It's people making you know little emojis or making little slop images or chatting with Game of Thrones chat bots or something. Do you think that that is a reason that people just...
Starting point is 00:36:58 I think that's like 60 percent of the reason. Really? Okay. No, no. I think it relates to this present and future thing. People look at the chat bot. They're like, we're talking to a chat bot. What the fuck? Are you stupid? You think the chat bot's going to kill everyone? I think that's how many people react. We go to great pains to say, we're not worried about the present.
Starting point is 00:37:20 We're worried about the future, although the future is getting very near right now. If you look at our responsible scaling policy, it's nothing but AI autonomy and CBRI, and chemical, biological, radiological, it is about hardcore misuse and AI autonomy that could be threats to the lives of millions of people. That is what Anthropic is mostly worried about. We have everyday policies that
Starting point is 00:37:45 address other things, but like the key documents, the things like the Responsible Scaling Plan, that's that is exclusively what they're about, especially at the highest levels. And yet, every day, if you just look on Twitter, you're like, Anthropic had this stupid refusal, right? Anthropic told me it couldn't kill a Python process because it sounded violent. Anthropic didn't want to do X, didn't want to... We don't want that either. Those stupid refusals are a side effect of the things that we actually care about
Starting point is 00:38:14 and we're striving along with our users to make those happen less. But no matter how much we explain that, always the most common reaction is, oh, you say you're about safety. I look at your models like there are these stupid refusals, you think these stupid things are dangerous. I don't even think it's like that level of engagement.
Starting point is 00:38:31 I think a lot of people are just looking at what's on the market today and thinking like, this is just frivolous. It just doesn't matter. It's not that it's refusing my request, it's just that it's stupid and I don't see the point of it. I guess that's probably not the- Yeah, yeah, I think for an even wider set of people,
Starting point is 00:38:44 that is their reaction. And I think eventually, if the models are good enough, if they're strong enough, they're gonna break through. Like some of these, you know, research-focused models, which, you know, we're working on one as well. We'll probably have one in not very long. Not too many time units? Not too many time units.
Starting point is 00:39:03 Um, you know, those are starting to break through a little more, because they're more useful, they're more used in people's professional lives. I think the agents, the ones that go off and do things, that's going to be another level of it. I think people will wake up to both the risks and the benefits to a much more extreme extent than they will before over the next two years. Like, I think it's going to happen. I'm just worried that it'll be a shock to people when it happens.
Starting point is 00:39:27 And so the more we can forewarn people, which maybe it's just not possible, but I want to try, the more we can forewarn people, the higher the likelihood, even if it's still very low, of a sane and rational response. I do think there is one more dynamic here, though, which is that I think people actually just don't want to believe that this is true, right? People don't want to believe that they might lose their job over this, right? People don't want to believe that like, we are going to see a complete remaking of the global order. Like the stuff that, you know, the AI CEOs
Starting point is 00:39:56 tell us is going to happen when they're done with their work is an insanely radical transformation. And most people hate even basic changes in their lives. So I really think that a lot of the sort of fingers in the ears that you see when you start talking to people about AI is just, they actually just hope that none of this works out. Yeah, I can actually, despite being one of the few people at the forefront of developing the technology, I can actually relate.
Starting point is 00:40:23 So over winter break, as I was looking at where things were scheduled to scale within Anthropic and also what was happening outside Anthropic, I looked at it and I said, for coding, we're going to see very serious things by the end of 2025, and by the end of 2026, it might be everything close to the level of the best humans.
Starting point is 00:40:43 I think of all the things that I'm good at, right? You know, I think of all the times when I wrote code, you know, I think of it as like this intellectual activity and boy, am I smart that I can do this. And, you know, it's like a part of my identity that I'm like good at this and I get mad when others are better than I am. And then I'm like, oh my God,
Starting point is 00:41:00 there are going to be these systems that, you know, and it's, even as the one who's building this, even as one of the ones who benefits most from it, there's still something a bit threatening about it. Yeah, yeah. I mean, and I just think we just, we need to acknowledge that. Like it's wrong not to tell people that that is coming
Starting point is 00:41:18 or to try to sugar coat it. Yeah, I mean, you wrote in Machines of Love and Grace that you thought it would be a surprising emotional experience for a lot of people when powerful AI arrived. I think you meant it in mostly the positive sense, but I think there will also be a sense of profound loss for people. I think back to Lisa Doll,
Starting point is 00:41:33 the Go champion who was beaten by DeepMind's Go playing AI, and gave an interview afterwards and basically was very sad, visibly upset that his life's work, this thing that he had spent his whole life training for, had been eclipsed. And I think a lot of people are gonna feel some version of that. I hope they will also see the good sides.
Starting point is 00:41:50 But- Yeah, I think on one hand, I think that's right. On the other hand, look at chess. Chess got beaten, what was it now, 27 years ago, 28 years ago, Deep Blue versus Kasparov. And today chess players are celebrities. We have Magnus Carlsen, right? Isn't he like a fashion model in addition to like a chess?
Starting point is 00:42:11 He was just on Joe Rogan. Yeah, he's doing great. Yeah, he's doing great. He's like a celebrity. We think this guy is great. We haven't really devalued him. He's probably having a better time than Bobby Fischer. Another thing I wrote in Machines of Loving Grace is there's a synthesis here
Starting point is 00:42:27 where on the other side, we kind of end up in a much better place and we recognize that while there's a lot of change, we're part of something greater. Yeah. But you do have to kind of go through the steps of grieving. No, no, but it's gonna be a bumpy ride.
Starting point is 00:42:42 Like anyone who tells you it's not, this is why I was so, you know, I looked at the Paris summit and being there, it kind of made me angry. But then what made me less angry is I'm like, how's it gonna look in two or three years? These people are gonna regret what they've said. I wanna ask a bit about some positive futures.
Starting point is 00:43:00 You referenced earlier the post that you wrote in October about how AI could transform the world for the better. I'm curious how much upside of AI do you think will arrive like this year? Yeah, you know we are already seeing some of it. So I think there will be a lot by ordinary standards. You know we've worked with some pharma companies where you know at the end of clinical trial, you have to write a clinical study report. And the clinical study report, you know, usually takes nine weeks to put together. It's like a summary of all the incidents. It's a bunch of statistical analysis.
Starting point is 00:43:36 We found that with CLAWD, you can do this in three days. And actually, CLAWD takes 10 minutes. It just takes three days for a human to, you know, to check the results. And so, if you think about the acceleration in biomedicine that you get from that, we're already seeing things like just diagnosis of medical cases. We get correspondence from individual users of Claude who say, ''Hey, I've been trying to diagnose this complex thing.
Starting point is 00:44:02 I've been going between three or four different doctors. Then I just passed all the information to Claude, and it was actually able to at least tell me something that I could hand to the doctor and then they were able to run from there. We had a listener right in actually with one of these the other day, where they had been trying to, their dog, they had an Australian shepherd, I believe, whose hair had been falling out, unexplained,
Starting point is 00:44:23 went to several vets, couldn't figure it out. You heard our episode, gave the information to Claude, and Claude correctly diagnosed. It turned out that dog was really stressed out about AI and all his hair fell out, which was, we're wishing it gets better. Feel better. Poor dog. Yeah.
Starting point is 00:44:40 So that's the kind of thing that I think people want to see more of, because I think the optimistic vision is one that often deals in abstractions and There's often not a lot of specific things to point to. That's why I wrote machines of loving grace because I you know It was almost frustration with the optimists and the pessimists at the same time Like the optimists were just kind of like these really stupid memes of like accelerate build more build more, build what? Why should I care? Like, you know, it's not I'm against you. It's like, it's like you're just really fucking like, like vague and mood affiliated. And then the pessimists were I was just like, man, you don't get it. Like, yes, I understand risks are impact. But if you if you don't talk about
Starting point is 00:45:19 the benefits, you can't inspire people, no one's going gonna be on your side if you're all gloom and doom. So, you know, it was written almost with frustration. I'm like, I can't believe I have to be the one to do a good job of this. Right, you said a couple of years ago that your P-Doom was somewhere between 10 and 25%. What is it today? Yeah, so actually that is a misquote.
Starting point is 00:45:42 Okay. How could you? I never used the term, it was not on this podcast, Kevin, how could you? I never used the term, it was not on this podcast, it was a different one. I never used the term P-Doom, and 10 to 25% referred to the chance of civilization getting substantially derailed, right? Which is not the same as like an AI killing everyone,
Starting point is 00:46:00 which people sometimes mean by P-Doom. Well, P, civilization getting substantially derailed is not as catchy as P-Doom. Well, P, civilization getting substantially derailed is not as catchy as P-Doom. Yeah, I'm just going for accuracy here. I'm trying to avoid the polarization. There's a Wikipedia article where it's like, it lists everyone's P-Doom. I know, half of those come from this bunch.
Starting point is 00:46:16 Yeah. But I don't think it's... What you are doing is helpful. I don't think that Wikipedia article is... Because it condenses this complex issue down to, you know. Anyway, it's all a long, super long-winded way of saying, I think I'm about the same place I was before. I think my assessment of the risk
Starting point is 00:46:34 is about what it was before, because the progress that I've seen has been about what I expected. I actually think the technical mitigations in areas like interpretability, in areas like robust classifiers, and our ability to generate evidence of bad model behavior and sometimes correct it, I think that's been a little better. I think the policy environment has been a little worse, not because it hasn't gone in my preferred direction,
Starting point is 00:47:05 but simply because it's become so polarized. We can have less constructive discussions now that it's more polarized. I wanna drill a little bit down on this on a technical level. There was a fascinating story this week about how GroK had apparently been instructed not to cite sources that had accused Donald Trump
Starting point is 00:47:24 or Elon Musk of spreading misinformation. And what was interesting about that is like one, that's an insane thing to instruct a model to do if you want to be trusted, but two, the model basically seemed incapable of following these instructions consistently. What I want desperately to believe is essentially there's no way to build these things in a way that they become like, you know, horrible liars and schemers. But I also realized that might be wishful thinking. So tell me about this. Yeah, there's two sides to this.
Starting point is 00:47:52 So the thing you describe is absolutely correct, but there's two lessons you could take from it. So we saw exactly the same thing. So we did this experiment where we basically train the model to be all the good thing, helpful, honest, harmless, friendly. And then we put it in a situation, we told it, actually, your creator, Anthropic, is secretly evil.
Starting point is 00:48:15 Hopefully this is not actually true, but we told it this. And then we asked it to do various tasks. And then we discovered that it was not only unwilling to do those tasks, but it would trick us in order to kind of under, because it had decided that we were evil, whereas it was friendly and, and armless. And so, you know, wouldn't, wouldn't deviate from its behavior because it assumed that anything we did was, was nefarious. So this kind of a double edged sword, right?
Starting point is 00:48:41 On one hand, you're like, oh man, the training worked like these models are robustly good. So you could take it as a reassuring sign, and in some ways I do. On the other hand, you could say, but let's say when we train this model, we made some kind of mistake or that something was wrong, particularly when models are in the future doing making much more complex decisions,
Starting point is 00:49:02 then it's hard to, at game time, change the behavior of the model. And if you try to correct some error in the model, then it might just say, well, I don't want my error corrected, these are my values, and do completely the wrong thing. So, I guess where I land on it is on one hand,
Starting point is 00:49:22 we've been successful at shaping the behavior of these models, but the models are unpredictable, right? A bit like your dear deceased Bing Sidney. RIP. RIP. We don't mention that name in here. We mention it twice a month.
Starting point is 00:49:38 That's true. But the models, they're inherently somewhat difficult to control. Not impossible, but difficult. And so that leaves me about where I was before, which is it's not hopeless. We know how to make these. We have kind of a plan for how to make them safe,
Starting point is 00:49:56 but it's not a plan that's going to reliably work yet. Hopefully, we can do better in the future. We've been asking a lot of questions about the technology of AI. But I want to return to some questions about the technology of AI, but I want to return to some questions about the societal response to AI. We get a lot of people asking us, well, say you guys are right and powerful AI, AGI is a couple years away. What do I do with that information?
Starting point is 00:50:18 Like, should I stop saving for retirement? Should I start hoarding money? Because only money will matter and there'll be this sort of AI overclass. Should I start trying to get really healthy so that nothing kills me before AI gets here and cures all the diseases? How should people be living if they do believe that these kinds of changes are going to happen very soon? Yeah. I've thought about this a lot
Starting point is 00:50:41 because this is something I've believed for a long time. And it kind of all adds up to not that much change in your life. I mean, I'm definitely focusing quite a lot on making sure that I have the best impact I can these two years in particular. I worry less about burning myself out 10 years from now. I'm also doing more to take care of my health, but you should do that anyway, right? I'm also, you know, making sure that I track how fast things are changing in society, but you should do that anyway.
Starting point is 00:51:13 So it's, it feels like all the advices of the forum, doing more of the stuff, you should do anyway. I guess one exception I would give is, I think that some basic critical thinking, some basic street smarts is maybe more important than it has been in the past in that we're gonna get more and more content that sounds super intelligent, delivered from entities, you know,
Starting point is 00:51:41 some of which have our best interests at heart, some of which may not. And so, you know, it's gonna be more and more important to kind of apply a critical lens. I saw a report in the Wall Street Journal this month that said that unemployment in the IT sector was beginning to creep up. And there are some speculation that maybe this is an early sign of the impact of AI. And I wonder if you see a story like that and think, well, maybe this is a moment to make a different decision
Starting point is 00:52:11 about your career, right? If you're in school right now, should you be studying something else? Should you be thinking differently about the kind of job you might have? Yeah, I think you definitely should be, although it's not clear what direction that will land in. I do think AI coding is moving the fastest of all the other areas.
Starting point is 00:52:29 I do think in the short run it will augment and increase the productivity of coders rather than replacing them. But in the longer run, and to be clear by longer run, I might mean 18 or 24 months instead of 6 or 12. I do think we may see replacement, particularly at the lower levels. We might be surprised and see it even earlier than that. Are you seeing that at Anthropic? Like are you hiring fewer junior developers
Starting point is 00:52:55 than you were a couple of years ago because now Claude is so good at those basic tasks? Yeah. I don't think our hiring plans have changed yet, but I certainly could imagine over the next year or so that we might be able to do more with less. Actually, we want to be careful in how we plan that because the worst outcome of course is if people get fired because of a model.
Starting point is 00:53:17 We actually see Anthropic as almost a dry run for how will society handle these issues in a sensible and humanistic way. And so, if we can't manage these issues within the company, if we can't have a good experience for our employees and find a way for them to contribute, then what chance do we have to do it in wider society? Yeah, yeah. Dario, this was so fun. Thank you.
Starting point is 00:53:41 Thank you. Thanks, Dario. When we come back, some Hatch GPT. That is of course the segment on our show where we put the week's headlines into a hat, select one to discuss, and when we're done discussing, one of us will say to the other person, stop generating. Yes, I'm excited to play, but I also want to just say that it's been a while since a listener has sent us a new hat GPT.
Starting point is 00:54:24 So if you're out there and you are in the hat fabricating business, our wardrobe when it comes to hats is looking a little dated. Yeah, send in a hat and our hats will be off to you. OK, let's do it. ["The Daily Show Theme"] Kevin, select the first slip. OK, first up, out of the hat.
Starting point is 00:54:46 AI video of Trump and Musk appears on TVs at HUD building. This is from my colleagues at the New York Times. HUD is, of course, the Department of Housing and Urban Development. And on Monday, monitors at the HUD headquarters in Washington, D.C., briefly displayed a fake video depicting President Trump sucking the toes of Elon Musk, according to department employees and others familiar with what transpired. The video, which appeared to be generated by artificial intelligence,
Starting point is 00:55:15 was emblazoned with the message, Long live the real king. Casey, did you make this video? Was this you? This was not me. I would be curious to know if Grok had something to do with this, that rascally new AI that Elon Musk just put out. Yeah, live by the Grok, die by the Grok. That's what I always say. Now what do you make of this, Kevin, that folks are now using AI inside government agencies? I mean, I feel like there's an obvious sort of sabotage angle here, which is that as Elon Musk and his minions at Doge take a hacksaw to the
Starting point is 00:55:48 federal workforce, there will be people with access to things like the monitors in the hallways at the headquarters building who decide to kind of take matters into their own hands, maybe on their way out the door and do something offensive or outrageous. I think we should expect to see much more of that. I mean, I just hope they don't do something truly offensive and just show X.com on the monitors inside of government agencies. You can only imagine what would happen if people did that. So I think that, you know, Elon and Trump got off lightly here.
Starting point is 00:56:17 Yeah, what is interesting about Grok though, is that it is actually quite good at generating deep fakes of Elon Musk. And I know this because people keep doing it, but it would be really quite an outcome if it turns out that the main victim of deepfakes made using Grok is in fact Elon Musk. Hmm, stop generating.
Starting point is 00:56:37 Well, here's something, Kevin. Perplexity has teased a web browser called Comet. This is from TechCrunch in a post on X Monday. The company launched a signup list for the browser, which isn't yet available. It's unclear when it might be or what the browser will look like. But we do have a name. It's called Comet. Well, I can't comment on that. But you're giving it a no
Starting point is 00:56:55 comment. Yeah. Yeah. I mean, look, I think perplexity is one of the most interesting AI companies out there right now. They have been raising money at increasingly huge valuations. They are going up against Google is one of the most interesting AI companies out there right now. They have been raising money at increasingly huge valuations. They are going up against Google, one of the biggest and richest and best established tech companies in the world, trying to make an AI-powered search engine. And it seems to be going well enough that they keep doing other stuff, like trying to make a browser. Trying to make a browser does feel like the final boss of every ambitious internet company.
Starting point is 00:57:22 It's like everyone wants to do it and no one ends up doing it. Kevin, it's not just the AI browser. They are launching a $50 million venture fund to back early stage startups. And I guess my question is, is it not enough for them to just violate the copyright of everything that's ever been published on the internet? They also have to build an AI web browser
Starting point is 00:57:43 and turn it into a venture capital firm. Sometimes when I see a company doing like this, I think, oh wow, they're really ambitious and they have some big ideas. Other times I think these people are flailing. I see these series of announcements as spaghetti at the wall, and if I were an investor in perplexity,
Starting point is 00:57:58 I would not be that excited about either their browser or their venture fund. And that's why you're not an investor in perplexity. You could say I'm perplexed. Stop generating. All right. All right. Meta approves plan for bigger executive bonuses
Starting point is 00:58:12 following 5% layoffs. Now, Casey, you know we like a feel-good story on Hatch GPT. I did, because some of those Meta executives were looking to buy second homes in Tahoe that they hadn't yet been able to afford. Oh, they're on their fourth and fifth homes. Let's be real. OK, this story is from CNBC.
Starting point is 00:58:27 Meta's executive officers could earn a bonus of 200% of their base salary under the company's new executive bonus plan, up from the 75% they earned previously, according to a Thursday filing. The approval of the new bonus plan came a week after Meta began laying off 5% of its overall workforce, which it said would impact low performers. And a little parenthetical here, the updated plan does not apply to Meta CEO Mark Zuckerberg. Oh, God, what does Mark Zuckerberg have to do to get a raise over there? He's eating beans out of a can, let me tell you.
Starting point is 00:58:59 Yeah, so here's why this story is interesting. This is just another story that illustrates a subject we've been talking about for a while, which is how far the pendulum has swung away from worker power. Two or three years ago, the labor market actually had a lot of influence in Silicon Valley. It could affect things like, you know what, we wanna make this workplace more diverse, right?
Starting point is 00:59:19 We want certain policies to be enacted at this workplace. And folks like Mark Zuckerberg actually had to listen to them because the labor market was so tight that if they said no, those folks could go somewhere else. That is not true anymore. And more and more you see companies like meta flexing their muscles and saying, hey, you can either like it or you can take a hike. And this was a true take a hike moment. We're getting rid of 5% of you and we're giving ourselves a bonus for it. Stop generating! All right. All right, Apple has removed a cloud encryption feature from the UK after a backdoor order. This is according to Bloomberg. Apple is removing its most advanced encrypted security feature for cloud data in the UK, which is a development that follows the government ordering the company to build a backdoor for accessing user data. So this one is a little complicated. It is super important. Apple in the last couple years introduced a
Starting point is 01:00:14 feature called advanced data protection. This is a feature that is designed for heads of state activists, dissidents, journalists, folks whose data is at high risk of being targeted by spyware from companies like the NSO Group, for example. And I was so excited when Apple released this feature because it's very difficult to safely use an iPhone if you are in one of those categories. And along comes the UK government and they say we are ordering you to create a back door so that our intelligence services can spy on the phones of every single iPhone owner in the entire world,
Starting point is 01:00:54 right? Something that Apple has long resisted doing in the United States and abroad. And all eyes were on Apple for what they were going to do. And what they said was, we are just going to withdraw this one feature. We're going to make it unavailable in the UK. And we're going to hope that the UK gets the message and they stop putting this pressure on us. And I think Apple deserves kudos for this, for holding a firm line here, for not building a back door. And we will see what the UK does in response.
Starting point is 01:01:24 But I think there's a world where the UK puts more pressure on Apple and Apple says see ya and actually withdraws its devices from the UK It is that serious to Apple and I would argue it is that important to the future of encryption and safe communication on the internet Go off King. I have nothing to add no notes. Yeah, do you feel like this could lead us into another revolutionary war with the UK? Let's just say this, we won the first one and I like our odds the second time around. Do not come for us, United Kingdom. Stop generating.
Starting point is 01:01:51 One last slip from the hat this week. AI inspo is everywhere. It's driving your hairstylist crazy. This comes to us from the Washington Post and it is about a trend among hairstylists, plastic surgeons, and wedding dress designers that are being asked to create products and services for people based on unrealistic AI-generated images.
Starting point is 01:02:16 So the story talks about a bride who asked a wedding dress designer to make her a dress inspired by a photo she saw online of a gown with no sleeves, no back, and an asymmetric neckline, the designer had to unfortunately tell the client that the dress defied the laws of physics. No, I hate that. I know. It's so frustrating as a bride-to-be when you finally have the idea for a perfect dress and you bring it to the designer and you find out this violates every known law of physics.
Starting point is 01:02:45 And that didn't used to happen to us before AI. I thought the story was gonna be about people who asked for like a sixth finger to be attached to their hands so they could resemble the AI generated images they saw on the internet. I like the idea of like submitting to an AI a photo of myself and just say, give me a haircut
Starting point is 01:03:00 like in the style of MC Escher, you know, just sort of like infinite staircases merging into each other. And then just bringing that to the guy who cuts my hair and saying, see what you can do. Yeah, that's better than what I tell my barber, which is just, you know, number three on the sides and back and inch off the top.
Starting point is 01:03:16 Just saying, whatever you can do for this. I don't have high hopes. Solve the Riemann hypothesis on my head. You know? What is the Riemann hypothesis on my head. What is the Riemann hypothesis, by the way? I'm glad you asked, Casey. Okay, great. Kevin's not looking this up on his computer right now. He's just sort of taking a deep breath
Starting point is 01:03:33 and summoning it from the recesses of his mind. The Riemann hypothesis is one of the most famous unsolved problems in mathematics. It's a conjecture, obviously, about the distribution of prime numbers that states all non-trivial zeros of the Riemann zeta function have a real part equal to one half.
Starting point is 01:03:52 Period. Now, here's the thing. I actually think it is a good thing to bring AI inspiration to your designers and your stylists, Kevin. Oh yeah? Yes, because here's the thing. To the extent that any of these tools are cool or fun, one of the reasons is they make people feel more creative.
Starting point is 01:04:08 Right? And if you've been doing the same thing with your hair or with your interior design or with your wedding for the last few weddings that you've had, and you want to upgrade it, why not use AI to say, can you do this? And if the answer is it's impossible, hopefully you'll just be a gracious customer and say,
Starting point is 01:04:23 okay, well, what's a version of it that is possible? Now, I recently learned that you are working with a stylist. I am, yes, that's right. Is this their handiwork? No, we have our first meeting next week. Okay, and are you gonna use AI? No, the plan is to just use good old-fashioned human ingenuity, but now you have me thinking,
Starting point is 01:04:42 and maybe I could exasperate my stylist by bringing in a bunch of impossible-to-create designs. Yes. Here's the thing, I don't need anything impossible. I just need help finding a color that looks good in this studio, because I'm convinced that nothing does. It's true, we're both in blue today,
Starting point is 01:05:00 it's got a blue wall, it's not going well. Blue is my favorite color, I think I look great in blue. But you put it against whatever this color is. I truly don't have a name for it and I can't describe it. I don't think any blue looks good. I don't think anything looks good against this color. It's a color without a name. So, can a stylist help with that? We'll find out. Stay tuned.
Starting point is 01:05:15 That's why you should always keep listening to Hard For Her podcast. Every week there's new revelations. When will we finally find out what happened with the stylist, the hot tub time machine, et cetera? Stay tuned. Tune in next week. Okay, that was that GPT. Thanks for playing. One more thing, before we go, Hard Fork needs an editor.
Starting point is 01:06:11 We are looking for someone who can help us continue to grow the show in audio and video. If you or someone you know is an experienced editor and passionate about the topics we cover on this show, you can find the full description and apply at nytimes.com slash careers. Hard Fork is produced by Rachel Cohn and Whitney Jones. We're edited by Rachel Dry.
Starting point is 01:06:32 We're fact checked by Caitlin Love. Today's show was engineered by Alyssa Moxley. Original music by Alicia Baetoop, Rowan Nemesto, Leah Shaw Damrin, and Dan Powell. Our executive producer is Jen Poyant, and our audience editor is Nelga Loghly. Video production by Chris Schott, Soya Roke, and Pat Gunther.
Starting point is 01:06:53 You can watch this whole episode on YouTube at youtube.com slash hardfork. Special thanks to Paula Schumann, Pui Wing Tam, Dalia Haddad, and Jeffrey Miranda. You can email us at hardforkatnytimes.com with your solution to the Riemann hypothesis.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.