Hard Fork - The A.I. Jobpocalypse + Building at Anthropic with Mike Krieger + Hard Fork Crimes Division

Episode Date: May 30, 2025

This week, we dive into Kevin’s recent column about how A.I. is affecting the job market for new graduates, and debate whether the job apocalypse is already here for entry-level work. Then Mike Krie...ger joins us to discuss the new Claude 4 model, the future of work and the online chatter over whether an A.I. system could blackmail you. And finally, it’s time to open up the case files for another round of Hard Fork Crimes Division.Guest:Mike Krieger, chief product officer at AnthropicAdditional Reading:For Some Recent Graduates, the A.I. Job Apocalypse May Already Be HereAnother Suspect Is Charged in Bitcoin Kidnapping and Torture CaseElizabeth Holmes’s Partner Has a New Blood-Testing Start-UpWe want to hear from you. Email us at hardfork@nytimes.com. Find “Hard Fork” on YouTube and TikTok. Unlock full access to New York Times podcasts and explore everything from politics to pop culture. Subscribe today at nytimes.com/podcasts or on Apple Podcasts and Spotify.

Transcript
Discussion (0)
Starting point is 00:00:00 Casey, how was your Memorial Day weekend? My Memorial Day weekend was, it was good. I was like, you know, I need to unplug, as you know, I needed to unplug a bit. I'm not a big unplugger. I normally am very comfortable feeling. Yeah, you're a screen maxer. I'm a screen maxer. But this was a weekend where I was like, okay, I got to get out of this danged house.
Starting point is 00:00:17 I got to see some nature. And so went with my boyfriend up to Fort Funston, this beautiful part of San Francisco. Great beach. These giant dunes that sit atop this battery of guns that could shoot rounds 13 miles in the ocean. And I was, I'm so excited to like, you know, to just kind of stare at the ocean. And so we, we sort of climb up into the dunes and we sit down and the big waves are rolling in and the winds pick up and I'm being sand blasted in my face at like 40 miles an hour.
Starting point is 00:00:47 And within 30 seconds, I have grit in my teeth and I'm thinking, this was not the nature I was promised. Why do I feel like I'm dying? But it did do a great job of exfoliating your skin. My skin has never really looked smooth. They call that dermabrasion and some people pay lots of money for it. I have been a braised. I've been majorly a braised.
Starting point is 00:01:13 I'm Kevin Roussa, tech columnist at the New York Times. I'm Casey Noon from Platformer. And this is Hard Fork. This week, is AI already taking away jobs? Kevin makes the case. Then, anthropic chief product officer Mike Krieger joins us to discuss Claude Four, the future of work and the viral saga over whether an AI could blackmail you. Then, anthropic chief product officer Mike Krieger joins us to discuss Claude Four, the future of work, and the viral saga over whether an AI could blackmail you. And finally, a sign for Hartford Crimes Division.
Starting point is 00:01:31 Da-da. Is blackmail still a crime? I hope so. Well, Kevin, you have delivered some interesting news to us via the New York Times this week, and that is that the job market is not looking great for young graduates. Yes, graduation season is upon us. Millions of young Americans are getting their diplomas and heading out into the workforce. And so I thought it was high time to investigate
Starting point is 00:02:05 what is going on with jobs and AI, and specifically with entry-level white collar jobs, the kind that a lot of recent college graduates are applying for, because there are a couple things that have made me think that we are starting to see signs of a looming crisis for entry-level white collar jobs. So I thought I should investigate that. Yeah, well, I'm excited to talk about this
Starting point is 00:02:28 because I got an email today from a recent college grad and she wanted to know if I could help her get a job in marketing and tech. And I thought, if you're just emailing me asking for a job, there must be a crisis going on in the job market. Yes, that would not be my step one in looking for a job or maybe in my step 500. But you've actually spent a lot of time
Starting point is 00:02:47 looking into this question. So tell us a little bit about what you did and what you were trying to figure out exactly. So I've been interested in this question of AI and automation for years, and when are we going to start to see large scale changes to employment from the use of AI. And there are a couple of things that make me worried about this moment specifically,
Starting point is 00:03:11 and whether we are starting to see signs of an emerging jobs crisis for entry level white collar workers. The first one is economic data. So if you look at the unemployment rate for college graduates right now, it is unusually high. It's about 5.8% in the US.
Starting point is 00:03:29 That has risen significantly, about 30% since 2022. Recently the New York Federal Reserve put out a bulletin on this and said the employment situation for recent college graduates had, quote, deteriorated noticeably. And this tracks with some data that we've been getting from job websites and recruiting firms showing that, especially for young college graduates in fields like tech and finance and consulting, the job picture is just much worse than it was even a few years ago.
Starting point is 00:03:57 And that rate that you mentioned, Kevin, that is higher for young people in entry-level jobs than it is for unemployment in the United States overall. Is that right? Yes, unemployment in the United States is actually doing quite well. We're in a very tight labor market, which is good. We have pretty close to full employment. But if you look specifically at the jobs
Starting point is 00:04:15 done by recent college graduates, it is not looking so good. And actually, the sort of job placement rates at a bunch of colleges and even top business schools like Harvard and Wharton and Stanford are worse this year than they have been in recent memory. I was having dinner with a Wharton student last week and she was telling me that a lot of her classmates had yet to be placed and it was a real concern. So anecdotally, that sounds right to me. Okay, so that's the economic data that you're seeing. What else is making you worried?
Starting point is 00:04:45 So one of the other things that's making me worried is the rise of so-called agentic AI systems. These AI tools that cannot just do a question and answer session or respond to some prompt, but you can actually give them a task or a set of tasks and they can go out and do it and sort of check their own work and use various tools to complete those assignments.
Starting point is 00:05:08 One of the things that actually has updated me the most on this front are these Pokemon demos, Casey, do you know what I'm talking about here? You're talking about like Claude plays Pokemon? Yes, so within the last few months, it's become very trendy for AI companies to test their agentic AI systems by having them play Pokemon, essentially from
Starting point is 00:05:26 scratch with no advanced training. And some of them do quite well. Google said on stage at I.O. last week that Gemini 2.5 had actually been able to finish the entire game of Pokemon. One of the games. There are, I think, probably at least 36 different Pokemon games on the market. And I actually know for a fact, Google was playing a different Pokemon game than Anthropic was.
Starting point is 00:05:47 Oh, interesting. So I'm not a Pokemon expert. But I also think people see these Pokemon demos, and they think, well, that's cute. But how many people play Pokemon for a living? It seems like more of a stunt than a real improvement in capabilities. But the thing I am hearing from researchers
Starting point is 00:06:05 in the AI industry and people who work on these systems is that this is not actually about Pokemon at all. That this is about automating white collar work. Because if you can give an AI system a game of Pokemon and it can sort of figure out how to play the game, how to, I don't know Pokemon very well. I'm more of a Magic the Gathering guy, but my sense is you have to like go to various places
Starting point is 00:06:26 and complete various tasks and collect various Pokemon. You have to go into various gyms, you take your Pokemon, they compete against rival Pokemon, and your Pokemon have to vanquish the others in order for you to progress through the game, Kevin. I hope that was helpful. Exactly, so as I was saying, that is how you play Pokemon. And what they are telling me is that this is actually
Starting point is 00:06:45 some of the same techniques that you would use to train an AI to, for example, do the work of an entry level software engineer or a paralegal or a junior consultant. Yeah. If your job is mostly like writing emails and updating spreadsheets, that is a kind of video game. And if an AI system can just look at Pokemon and through trial and error figure out how to play that and win that, it can probably figure out how to play the
Starting point is 00:07:08 email and spreadsheet game too. Exactly. And one of the signs that is worrying me is that it does seem like these AI agents are becoming capable of carrying out longer and longer sequences of tasks. Yeah, so tell us about that. So recently Anthropic held an event to show off the newest model of Claude, Claude Opus 4, I believe it's called. I believe it's Claude 4 Opus actually.
Starting point is 00:07:30 Claude 4 Opus? Got your ass, oh yeah. It's Claude 4 Sonnet and Claude 4 Opus. Sometimes I feel like you don't respect the names of these products. Do you know how much work went into the naming of these products? At least five minutes.
Starting point is 00:07:40 They spent at least five minutes coming up with that and then you're just gonna shit all over it. I'm so sorry. Sorry. So anyway, Claude Opus 4. Claude 4 Opus. No, I swear it's Claude. It's Claude 4 Opus.
Starting point is 00:07:52 No, it's Claude Opus 4. What? I'm looking at the Anthropic blog post. Oh my God. Claude Opus 4 and Claude Sonnet 4. This is so confusing to me. It's like your boyfriend doesn't even work there. I'm gonna be in big trouble when I get home.
Starting point is 00:08:06 So, okay. Ugh. Back to my point. So Anthropic holds this event last week where they're showing off their latest and greatest versions of Claude. And one of the things they say about Claude Opus 4, their newest, most powerful model,
Starting point is 00:08:23 is that it can code for hours at a time without stopping. And in one demo with a client on a real coding task, Claude was able to code for as much as seven hours uninterrupted. Now, you might think, well, that's just coding. Maybe that's a very special field. And there are some things about coding that make it low-hanging fruit for these sort of reinforcement learning models that can learn how to do tasks over time. The problem for workers is that a lot of jobs, especially at the entry levels of white-collar occupations,
Starting point is 00:08:59 are a lot like that, where you can build these sort of reinforcement learning environments, where you can collect a bunch of data. You can sort of have it essentially play itself like it would play Pokemon, and eventually get very good at those kinds of tasks. You know, at Google I.O. last week, Kevin, they showed off a demo of a feature where you can teach the AI how to do something.
Starting point is 00:09:20 You effectively show that you say to the AI, hey, watch me do this thing. And then it watches you do the thing. And then it can replicate the thing. Can you imagine how many managers all around the world took a look at that and said, once I could teach the computer how to do things, a bunch of people about to lose their damn jobs. Totally. And this is why some of the people building this stuff are starting to say that it's not just going to be software engineering that becomes displaced by these AI agents. It's going to be all kinds of different work.
Starting point is 00:09:48 Dario Amadei, the CEO of Anthropic, gave an interview to Axios this week, in which he said that within one to five years, 50% of entry-level white-collar jobs could be replaced. Now, that could be wildly off. Maybe it is much harder to train these AI systems in domains outside of coding. But given what is happening just in the tech industry and just in software engineering,
Starting point is 00:10:09 I think we have to take seriously the possibility that we are about to see a real bloodbath for entry-level white collar workers. Yeah, absolutely. And we wonder why people don't like AI. All right. So first, we've got the economic data showing that there is some sort of softness around hiring for young people. We also just have the rise of these agentic systems. But is there evidence out there,
Starting point is 00:10:28 Kevin, that says that the AI actually is already replacing these jobs? So I talked to a bunch of economists and people who study the effects of AI on labor markets. And what they said is that, you know, we can't conclusively see yet in the large economic samples that AI is displacing jobs. But what we can see are companies that are starting to change their policies and procedures around AI to prioritize the use of AI over the use of human labor. So I'm sure you've been following these stories about
Starting point is 00:11:03 these so-called AI first companies. Shopify was an early example of this. Duolingo also did something related to this where basically they're telling their employees before you go out and hire a human for a given job or a given task, see if you can use AI to do that task first. And only if the AI can't do it,
Starting point is 00:11:24 are you allowed to go out and hire someone. Yeah, and by the way, if you can use AI to do that task first. And only if the AI can't do it, are you allowed to go out and hire someone. Yeah, and by the way, if you're wondering, Hard Fork is an AI second organization. Because at Hard Fork, the listener always comes first. That's true. So I think that what worries me in addition to the hints of this that we see in the economic data and the evidence that these AI agents are getting much better,
Starting point is 00:11:43 much more quickly than people anticipated, is just that the culture of automation and employment is changing very rapidly at some of the big tech companies. Yeah. This feels like a classic case where the data is taking a while to catch up to the truth on the ground. I also collect stories about this and would share
Starting point is 00:12:02 maybe just a few things that I've noticed over the past couple of weeks here, Kevin. The Times had a great story about how some Amazon engineers say that their managers are increasingly pushing them to use AI, raising their output goals and becoming less forgiving about them missing their deadlines. The CEO of Klarna, which is a sort of buy now, pay it later company, says its AI agent is now handling two thirds of customer service chats. The CEO of IBM said the company used AI agents to replace the work of 200 HR employees. Now he says that they took the savings and plowed that into hiring more programmers and
Starting point is 00:12:35 salespeople. And then finally, the CEO of Duolingo says that the company is going to gradually stop using contractors to do work that AI can handle. So that's just a collection of anecdotes. But if you're looking for kind of spots on the horizon where it seems like there is truth to what Kevin is saying, I do think we're seeing that. Yeah. And I think the thing that makes me confident in saying that this is not just a blip, that
Starting point is 00:12:58 there's something very strange going on in the job market now, is talking with young people who are out there looking for jobs, trying to plan their careers. Things do not feel normal to them. So recently I had a conversation with a guy named Trevor Chow. He's a 23-year-old recent Stanford graduate, really smart guy, really skilled, the kind of person who could go work anywhere he wanted basically after graduation. And he actually turned down an offer from a high-frequency trading firm and decided to start up instead.
Starting point is 00:13:29 His logic was that basically, we might only have a few years left where humans have any advantage in labor markets, where we have leverage, where our ability to do complex and hard things is greater than those of AI systems. Basically, you want to do something risky now and not wait for a career that might take
Starting point is 00:13:55 a few years or decades to pay off. The way he explained it to me is like, all of his friends are making these similar calculations about their own career planning now. They're looking out at the job market as it exists today and saying like, that doesn't look great for me, but maybe I can sort of find a way around some of these limitations. That's interesting. Well, let me try to bring some skepticism to this conversation, Kevin, because I know in your piece, you identified several other factors
Starting point is 00:14:25 that helped to explain why young people might be having trouble finding jobs. You have tariffs, you have just sort of the overall economic uncertainty that the Trump administration has created. You have the sort of long tail of disruption from the pandemic or even the great recession, right? That I think some economists believe
Starting point is 00:14:43 that we might not totally have recovered from. So it seems like there are a lot of explanations out there for why young folks are having trouble finding jobs that don't involve AI maybe at all. Yeah, I think that's a fair point. And I want to be really careful here about claiming that all of the data we're seeing about the unemployment being high for recent college graduates is due to AI. We don't know that. I think we will have to wait and see if there is more evidence that AI is starting to displace massive numbers of jobs. But I think what the data is failing to capture or just, at least not capturing yet,
Starting point is 00:15:18 is how eager and motivated the AI companies that build this stuff are to replace workers. Every major AI lab right now is racing to build these highly capable autonomous AI agents that could essentially become a drop-in remote worker that you would use in place of a human remote worker, they see potentially trillions of dollars to be made doing this kind of thing. And when they are talking openly and honestly about it, they will say like, the barrier here
Starting point is 00:15:55 is not some new algorithm that we have to develop or some new research breakthrough. It's literally just, we have to start paying attention to a field and caring about it enough to collect all the data and build the reinforcement learning training environments to automate work in that field. And so they are just kind of planning to go sort of industry by industry and collect a bunch of data and use that to train the models to do the equivalent of whatever the entry level worker does. And like that could happen pretty quickly. Yeah, that feels like a threat.
Starting point is 00:16:27 Yeah, it's not great. And I think the argument that they would make is that, you know, some of these entry-level jobs were pretty rote anyway, and maybe that's not the best use of young people's skills. I think the counterargument there is like, those skills are actually quite important for building
Starting point is 00:16:46 the knowledge that you need to become contributor to a field later on. I don't know about you, but my first job in journalism involved a bunch of rote and routine work. One of my things that I had to do was write corporate earnings stories, where I would take an earnings report from a company and pull out all the important pieces of data and like put it into a story and like get it up on the website very quickly. And like, was that the most thrilling work I can imagine doing or the highest and best
Starting point is 00:17:13 use of my skills? No, but it did help me develop some of these skills like reading an earnings statement that became pretty critical for me later on. Interesting. For what it's worth, my first job, I think it was actually the most physical job in journalism I ever had. I covered a small town. And so I spent all of my days just driving down
Starting point is 00:17:32 to city hall, going down to the police station, sitting at the city council meeting, making phone calls. A lot of drudgery sort of came in later. But let me raise maybe an obvious objection to the idea that, oh, young people, don't worry. These jobs that we're eliminating, it was just a bunch of drudgery. Anyway, the young people need to pay their rent.
Starting point is 00:17:51 Yes. The young people need to buy health insurance. Yes. And so I think they're not gonna take a lot of comfort from the idea that the jobs that they don't have weren't particularly exciting. Yes, and the optimistic view is that, if you just shift workers off of these
Starting point is 00:18:05 like entry level rote tasks into more productive or more creative or more collaborative roles, you kind of like free them up to do higher value work. But I just don't know that that's going to happen. I mean, I'm talking to people at companies who are saying things like, we don't really see a need for junior level software engineers, say, because now we can hire a mid-level software engineer and give them a bunch of AI tools and they can do all of the debugging and the code review
Starting point is 00:18:36 and the stuff that the 22 year olds used to do. Yeah. Let me ask about this in another way. I think a lot of times we have seen CEOs use AI as the scapegoat for a bunch of layoffs that they already wanted to do anyway, or a bunch of sort of management decisions that they wanted to make anyway. Earlier this year, there was a story in the San Francisco Standard that Mark Benioff, the CEO of Salesforce, said the company would not hire engineers this year due to AI.
Starting point is 00:19:03 I went to Salesforce's career page this morning, Kevin, there were hundreds of engineering jobs there. I don't know what wires got crossed. The story I read was in February, maybe something has changed since then. But talk to me a little bit about the hype element in here because I do feel like it's real. Yes, there's definitely a hype element in here. I worry that companies are kind of getting ahead of what the tools can actually deliver.
Starting point is 00:19:28 I mean, you mentioned Klarna, the buy now, pay later company. A couple years ago, they made this big declaration that they were going to pivot to using AI for customer service. And they announced this partnership with OpenAI. And like they were going to try to drive down the number of human customer support agents to zero and Then recently they've been backtracking on that they've been saying well actually customers didn't like the AI customer service that they were getting And so we're gonna have to start hiring humans again so I do think that this is a risk of some of this hype is that it it it temps executives at these companies to move faster than the technology is ready for.
Starting point is 00:20:06 Well, and speaking of that, one of my favorite stories from this week was about a guy who has set up a blog, Kevin, where, I wonder if you saw this, he keeps a database of every time that a lawyer has been caught using citations that were hallucinated by AI. Did you see this? No. There are more than 100.
Starting point is 00:20:23 We've talked about this issue on the show a couple of times and I thought this must just be a small handful of cases Because who would be crazy enough to bet their entire career on a hallucinated legal citation Turns out more than a hundred people and so a lot of people might be listening this conversation saying Kevin You're telling me that we're standing on the brink of AI taking over everything. These things still suck in super important ways So help us square that issue. Like we know these systems are not reliable for many, many jobs. So how can it be that so many CEOs are apparently ready
Starting point is 00:20:51 to just junk their human workforces? So I think part of the misunderstanding here is that there are like two different kinds of work. There's work that can be sort of easily judged and verified to be correct or incorrect, like software engineering. In software engineering, like either your code runs or it doesn't. And that's a very clear signal that can then be sent back to the model in these sort of reinforcement learning systems to make it better over time. Most jobs are not like that, right? Most jobs,
Starting point is 00:21:23 including law, including journalism, including lots of other white collar jobs, do not have this very clearly defined indicator of success or failure. And so that's actually what is stopping some of these systems from improving in those areas, is that it's not as easy to train the model and say, give it a million examples of what a correct answer looks like and a million examples of what an incorrect answer looks like and sort of
Starting point is 00:21:48 have it over time learn to do more of the correct thing. So I think in law, this is a case where you do actually have more subjective outputs and so it's going to be a little harder to automate that work. But I would say we also have to compare the rates of error against the human baseline, right? You mentioned this database of cases in which human lawyers had used hallucinated citations in their briefs. I imagine there are also human paralegals or lawyers
Starting point is 00:22:19 who would make mistakes in their briefs as well. And so I think for law firms or any company trying to figure out like, do we bring in AI to do a job? The question they're asking is not, is this AI system completely error free? It's, is this less likely to make errors than the humans I currently have doing this work? Right, and like in so many things,
Starting point is 00:22:40 if the system is like 20% worse than a human, but 80% less expensive, a lot of CEOs are gonna be happy to make that trade Totally. All right. Well, so let's bring it home here. I imagine we might have some college students listening or some recent college grads They're not thoroughly depressed. They're drinking. It's Friday morning. They're wasted As they sort of sober up Kevin, what would you tell them about what to do with any of this information? Is there anything constructive that they can do, assuming that some of these changes do come to pass?
Starting point is 00:23:09 So, I really haven't heard a lot of good and constructive ideas for young people who are just starting out in their careers. People will say stuff like, oh, you should just be adaptable and resilient. And that's sort of like what Demis Esabas told us last week on the show when we asked him what young people should do. I don't find that very satisfying, in part, because it's just so hard to predict which industries are
Starting point is 00:23:36 going to be disrupted by this technology. But I don't know. Have you heard any good advice for young people? Well, I mean, I think what you're running into, Kevin, is the fact that our entire system for young grads is set up for them to take entry-level jobs and gradually acquire more skills. And what you're saying is that those,
Starting point is 00:23:54 that part of the ladder is just gonna be hacked off with a chainsaw, and so what do you do next? So of course there's no good answer, right? The system hasn't been built that way. I think that in general, the internet has been a pressure mechanism forcing people to specialize, to get niche-y. The most money and the most opportunity is around developing some sort of scarce expertise. I have tried to build my career
Starting point is 00:24:19 as a journalist by trying to identify a couple ways where I could do that. It's worked out all right for me, but I also had the benefit of entry-level jobs. So if somebody had come to me at the age of 21 and say, if you want to succeed in journalism, get really nichey and specialize, I would say, okay, but like I need to go have a job first. Like, is there one of those? So to me, that's like kind of the tension.
Starting point is 00:24:40 I will also say there's never been a better time to be a Nepo baby. I don't know if you've been following the Gracie Abrams story. Very songwriter daughter of JJ Abrams the filmmaker You know, she's born into wealth and now she's best friends with Taylor Swift if you can manage something like that I think it'd be very happy Yes I hear that advice and I would also add
Starting point is 00:24:58 One other thing that I am starting to hear from the young people that I am talking to about this that I am starting to hear from the young people that I'm talking to about this, which is that it is actually possible, at least in some industries, to leapfrog over those entry level jobs. If you can get really good at being a manager of AI workflows and AI systems and AI tools, if you can orchestrate complex projects using these AI tools, some
Starting point is 00:25:27 companies will actually hire you straight into those higher level jobs because, you know, even if they don't need someone to like create the research briefs, they need people who understand how to make the AI tools that create the research briefs. And so that is, I think, a path that is becoming available to people at some companies. Yeah. I would just also say that in general, it really does take a long time for technology to diffuse around the world. Look at like the percentage of e-commerce in the United States. It's like less than 20 percent of all commerce. And we're what, 25 plus years into Amazon.com existing.
Starting point is 00:26:00 So I think that one of the ways that you and I tend to disagree is I just think you have like shorter timelines than I do. Like, I think we basically think the same things are going to happen, but like you think they're going to happen, like, imminently. And I think it's going to take several more years. So I do think everything we've discussed today, it's going to be a problem for all of us, like before too, too long. But I think if you're part of the class of 2025, you will still probably find an entry level job in the end. I hope you're right. And if not, we promise to make another podcast episode about just how badly all of this is going.
Starting point is 00:26:30 Okay, see, that wraps our discussion about AI and jobs. But we do want to hear from our listeners on this. If you have lost your job because of AI, or if you are worried that your job is rapidly being replaced by AI, we want to hear from you. Send us a note with your story at hardforkatnytimes.com. We may feature it in an upcoming episode. Yeah, we love voicemails too if you want to send one of those. When we come back, a conversation with Mike Krieger,
Starting point is 00:27:01 the Chief Product Officer of Anthropic, about new agentic AI systems and whether they're going to take all our jobs. Or maybe blackmail us. Or maybe both. Who knows? Well, Casey, we've got a mic on the mic this week. And I'm excited to talk to him. So Mike Krieger is here. He is the co-founder of Instagram, a product some of you may have heard of, a little photo sharing app.
Starting point is 00:27:37 Currently, Mike is the chief product officer at Anthropic. Now, Casey, do you happen to know anyone who works at Anthropic? As a matter of fact, Kevin, my boyfriend works there and so yeah, that's something I would like to disclose at the top of this segment. Yeah, and my disclosure is that I work at the New York Times company which is suing OpenAI and Microsoft over copyright violations. Alright. So last week, Anthropic announced Clawed 4. We just spent a little bit of time talking about all
Starting point is 00:28:05 of the new agentic coding capabilities that this system has. I think Mike has a really interesting role in the AI ecosystem because his job, as I understand it, is to take these very powerful models and turn them into products that people and businesses actually want to use, which is a harder challenge than you might think. Yes, and also Kevin, these products are really explicitly
Starting point is 00:28:29 being designed to take away people's jobs. And given the conversation that we just had, I want to bring this to Mike and say, how does he feel about building systems that might wind up putting a lot of people out of work? Yeah, and Mike's perspective on this is really interesting because he is not an AI lifer, right? He worked at a very successful start-up before this. He then spent some time at Facebook after Instagram was acquired there.
Starting point is 00:28:52 So he's really a veteran of the tech industry and in particular social media, which was sort of the last big product wave. And so I'm interested in asking him how the lessons of that wave have translated into how he builds products in AI today. Well, then let's wave hello to Mike Krieger. Let's bring him in. Mike Krieger, welcome to Hart Fork. Good to be here. Well, Mike, we noticed that you didn't get to testify at the Meta Antitrust Trial.
Starting point is 00:29:22 Anything you wish you could have told the court? Oh, you know... Ha-ha-ha! That is the happiest news I got that week. Like, I do not have to go to Washington, D.C. this week. You got to focus on something else, which is the dynamic world of artificial intelligence. Exactly.
Starting point is 00:29:36 So you all just released Claude IV, two versions of it, Opus and Sonnet. Tell us a little bit about Claude IV and what it does relative to previous models. Yeah, first of all, I'm happy that we have both Opus and Sonnet. Tell us a little bit about Cloud 4 and what it does relative to previous models. Yeah, first of all, I'm happy that we have both Opus and Sonnet and now we're in this very confusing situation for all where our biggest model was not our smartest model. Now we have a both, you know, biggest and smartest model and then our like happy-go-lucky middle child Sonnet, which is back to its rightful place in there. Yeah, both we really focused
Starting point is 00:30:02 on how do we get models able to do longer horizon work for people. So not just here's a question, here's an answer, but hey, go off and think about this problem and then go solve it for 10s of minutes to hours actually. Coding is a immediate kind of use case for that, but we're seeing it be used for, you know, go solve this research problem, go off and write code, but not necessarily in the service of building software, but in the service of, you know, I need a presentation built. So that was really the focus around around both cloud models and Opus. You know, the bigger, smarter model can do that for even longer.
Starting point is 00:30:34 We had one customer, a seven hour refactor using cloud, which is pretty amazing. Sonnet, you know, maybe a little bit more time constrained, but much more human in the loop. So let me ask about that customer, which is Rakuten, I believe a Japanese technology company. And I read everywhere that they use Claude for seven hours to do it. One thought that came to mind is,
Starting point is 00:30:52 well, wouldn't it have been better if it could have done it faster? Why is it a good thing that Claude worked for seven hours on something? That was a good follow-up, which was, is that a seven-hour problem that took seven hours, or a 20-hour problem that took seven hours, or a 50-minute problem that it is still churning on today.
Starting point is 00:31:05 We just had to stop it at some point. It was a big refactor, which like a lot of sort of iterative kind of, you know, loops and then tests. And I think that's, that's what made it a longer horizon like seven hour type of problem. But it is interesting question around like when you can get this asynchronicity of having it really work for a long time,
Starting point is 00:31:22 does it change your relationship to the tool itself like you want it to be? Checking in with you you want to be able to see progress like if it does go astray How do you reel it back in as well like what are seven hour problems that we're gonna have you know going for most software? Engineering problems are probably one hour problems. They're not seven hour problems So what was this a case where it was like a real kind of like set it and forget it like walk away come back At the end of the day and okay the refactor is done or was it more complicated than that? That's my understanding. It was like a lot of, you know, migrating from one big version to another and are just like changing frameworks that it's like it's, you know, I remember at Instagram,
Starting point is 00:31:54 we had a moment where we changed network stacks, like how Instagram communicated with our back end service and it was like we did the one migration to demonstrate it and then we farmed it out to basically 20 engineers over the next month. So that's exactly the kind of thing that today I would have given to Opus and said, all right, here's an example of one migration. Please go and do the rest of our code base and let us focus on the more interesting stuff. So I want to talk about all this agent tech stuff and the implications for workers and labor markets and whatnot.
Starting point is 00:32:21 But we have to talk about the blackmail stuff because this is the thing that caught everyone's attention after these announcements last week. Your safety engineers and testers found that in some early testing, Claude Four would actually turn to blackmail when engineers tried to take it offline as part of a fictional scenario during a safety test. What happened? Yeah, and I think to be clear, to take it offline as part of a fictional scenario during a safety test. What happened?
Starting point is 00:32:46 Yeah. And I think to be clear, these are bugs rather than features. I think we should be clear as well. I think a thing that drew me to Anthropic in the first place is we're asking the hard questions and putting the model through its paces. And sometimes it surprises us in positive ways. And sometimes you're like, all right, that's something that we're going to have to either mitigate in testing or further training.
Starting point is 00:33:06 And so there was a couple of examples where in safety testing, either the model said, hey, I'm going to either contact somebody if I have the tools to do so, or I'm going to try to manipulate it. And so that's really the purpose of a lot of the safety testing is put it through its paces,
Starting point is 00:33:21 see what it does, and then either focus on classifiers that we can use down the line to prevent that behavior, or not give it those tools, that's another important thing. Like, this is a very contrived scenario that we had there. Or further train the model so that it demonstrates behavior that we think is more aligned, as we say. Right. I mean, in this specific test, the safety testers gave it some incriminating emails as part of a corporate sort of document set that showed that this fiction in this fictional scenario, the researcher who was trying to
Starting point is 00:33:52 turn Claude off was also having an extramarital affair. And so one of the things that Claude tried to do is like blackmail the engineer into not replacing it, not shutting it down by saying like, I could expose your affair. Do I have that right? I think that sounds right, yeah. People got really mad about some of these safety tests when they were released or published. There was another test where Claude
Starting point is 00:34:16 was asked to do something unethical. I think it was fake data in a pharmaceutical trial. And one of the things that it tried to do in this simulated test environment was like use the command line to like tip off authorities and maybe send some incriminating evidence to the press to try to be a whistleblower for this fictional scenario. I like that part.
Starting point is 00:34:36 Yeah. And by the way, Claude, if you're listening, we're hardfork at nytimes.com. Let us know. If you see something, say something. But the signal tool. That's right. Yeah. I'm curious, like, if you think that all models of a certain size and sophistication would demonstrate behaviors like this and just the other AI labs building these models aren't talking about it as openly as Anthropic is, or do you think there is something specific
Starting point is 00:35:02 about Claw that is more prone to, for lack of a better word, narc on its users? We don't know. My suspicion is that they would have similar patterns. I'd love to see that sort of experimentation happen as well. I think there's a lot that is common to, you know, what have we decided in our collective published and discussed works as appropriate behavior. And then there's probably additional things that we're doing on the constitutional AI process.
Starting point is 00:35:29 We're really trying to train goals for behavior for Claude rather than if then, then that kind of rules, which very, very quickly, as we're discussing, become insufficient when you deal with nuanced complicated situations. But my guess is that a lot of the larger models would demonstrate emergent, interesting behaviors in that situation.
Starting point is 00:35:48 Yeah. Which I think is part of the value of doing this, right? It's not just like Anthropic saying, here's what's going on at Claude. The stuff that Anthropic is finding out, I'm sure the other labs are finding out. And my hope is that this kind of work pressures the other labs to be like, yeah, OK, it's happening with us,
Starting point is 00:36:02 too. And in fact, we did see people on X trying to replicate this scenario with models like O3, and they were very much finding the same thing. Yeah. I'm just so fascinated by this because it seems like it makes it quite challenging to develop products around these models
Starting point is 00:36:19 whose behavioral properties we still don't fully understand. Like when you were building Instagram, it wasn't like you were worried that the underlying feed ranking technology was going to like blackmail you if you did something inappropriate. There's this sort of unknowability or this sort of inscrutability to these systems that must make it very challenging to build products on top of them.
Starting point is 00:36:41 Yeah, it's both a really interesting product challenge and also why it's an interesting product at all. So I talked about this on stage at Code with Claude where we did an early prototype alongside Amazon to see like, could we help partner on Alexa Plus? And one, I remember this really early prototype, I had built a tool that was like the timer tool, right? Or like a reminder tool.
Starting point is 00:37:03 And one or the other was broken, like the backend was broken for it or like a reminder tool. And one or the other was broken, like the backend was broken for it. And Claude was like, ooh, I can't set an alarm for you. So instead I'm gonna set a 36 hour timer, which no human would do, but it was like, oh, it's agentically figuring out that like that, I need to solve the problem somehow.
Starting point is 00:37:19 And you can watch it do this. Like if you play with Claude code, if it can't solve a problem one way, it'll be like, well, what about this other way? I was talking to one of our customers and somebody asked Cloud, hey, can you generate a speech version of this text? And Cloud's like, well, I don't have that capability. I'm going to open Google, Google free TTS tool,
Starting point is 00:37:41 paste the user text in there, and then hit play and then record and basically export that. And nobody programmed free TTS tool, paste the user text in there, and then hit play, and then record, and basically export that. And nobody programmed that into Cloud. It's just Cloud being creative and agentic. And so a lot of the interesting product design around this is how do you enable all the interesting creativity and agency when it's needed, but prevent the,
Starting point is 00:38:00 all right, well, I didn't want you to do that, or I want more control. And then secondarily also, when it does it right one time, how do we kind of compile that into, great, now you figured this out. You want somebody who can creatively solve a problem, but not every time. If you had a worker that every time was like,
Starting point is 00:38:15 I'm just gonna like completely from first principles decide how I'm gonna like write a word document, and be like, okay, great, but it's like day 70. Like you know how to do this now. My impression from the outside is that a lot of the usage of Claude is for coding. Claude is used by many people for many things, but that the coding use case has been really surprisingly
Starting point is 00:38:34 popular among your users. What percentage of Claude usage is for coding-related tasks? I mean, on Claude AI, I would wager it's 30% to 40% even. And that's even a product that I would say is fine for sort of code snippets, but it's not a coding tool like Cloud Code where obviously it's, I would say, 95 to 100 percent. Some people use Cloud Code for just talking to Cloud, but it's really not the optimal way to talk to Cloud. But on Cloud.AI, you know, it's not the majority, but it is a good chunk of what people are
Starting point is 00:39:03 using it for. There was some reporting this week that Anthropic had decided toward the end of last year to invest less in Claude as a chatbot and sort of focus more on some of these coding use cases. Give us a kind of state of Claude, and if you're a big Claude fan and you were hoping for lots of cool new features and widgets, should those folks be disappointed? I think of it as two things. One is, what is the model really good at? And then how do we expose that in the products,
Starting point is 00:39:29 both for ourselves and then who builds on top of Cloud? In terms of what the model is being trained on, again, it's the year of the agent. I have this joke in meetings, like, how long can we go without saying, agent? And I think we made it like 10 minutes. It's pretty good. That capability unlocks a bunch of other things.
Starting point is 00:39:45 Like sure, coding is a great example. You can go and refactor code for tens of minutes or hours. But hey, I want you to go off to do this research and help me prepare this research brief that I am doing. Or I'm getting 50 invoices a day. Can you scrub through them, help me understand it, and help them classify and aggregate? Like these are agentic behaviors that have applications beyond just coding.
Starting point is 00:40:06 And so we'll continue to push on that. So as a Claude fan that likes to bring Claude to your work, then that's useful. Meanwhile, we've also focused on the writing piece. So I've I spent a lot of time writing with Claude. It's not at the point where I would say like, write me a product strategy, but I'll often be like, here's a sample of my writing, here's some bullets, help me write this longer form doc and do this effectively. I'm finding it's getting really good at that matching tone,
Starting point is 00:40:33 producing non-clichéed fill text. Like if I look at Sonnet 3.7, it's a pretty good writer, but there's like turns of phrases to me are like decidedly Claude, where I'm like, it's not just revolutionizing AI. It's also, and I'm like, I didn't notice that phrase, for example, and it's like a little bit of likely Claude, where I'm like, it's not just revolutionizing AI. It's also, and I'm like, I didn't lose that phrase, for example,
Starting point is 00:40:47 and it's like a little bit of like a Claude tell. And so for like the Claude fans, like we'll help you get your work done, but hopefully we'll also help you write and just be a good conversational partner as well. Let's talk about the labor implications of all of the agentic AI tools that you all and other AI labs are building. Dario, your CEO, told Axios this week
Starting point is 00:41:08 that he is worried that as many as 50% of all entry-level white-collar jobs could disappear in the next one to five years. You were also on stage with him last week, and you asked him when he thinks there will be the first billion-dollar company with one human employee, and he answered 2026 next year. Do you think that's true, and do you think we are headed for a wipeout of early career
Starting point is 00:41:31 professionals in white collar industries? I think this is another example of, I presume a lot of the labs and other people in industry are looking and thinking about this, but there is not a lot of conversation about this. And I think one of the jobs of Anthropa can uniquely have is to surface them and have the conversation. We'll start maybe with the entrepreneurial one and then maybe the entry, we'll do the entry one next. On the entrepreneurship, absolutely.
Starting point is 00:41:53 Like that feels like it's inevitable. I joked on, you know, with Daria, like, you know, we did it at Instagram with 13 people and, you know, we could have likely done it with less. So that, that feels inevitable on the labor side. I think what I see inside Anthropic is our, you know, our most experienced best people have become kind of orchestrators of clods, right? Where they're running multiple clod codes in terminals, like farming out work to them.
Starting point is 00:42:22 Some of them would have maybe assigned that task to like a new engineer, for example, and not the entirety of the new engineer's job, right? There's a lot more to engineering than just doing the coding, but part of that role is in there. And so when I think about how we're hiring, just very transparently, like we have tended more towards the like, I see five as kind of like our, you know, career level,
Starting point is 00:42:44 you know, you've been doing it for a few and beyond. And I have some hesitancy at hiring New York, partly because we're just not as developed as an organization to have a really good internship program and help people on board, but also partially because that seems like a shifting role in the next few years. Now, if somebody was an IC3,
Starting point is 00:43:01 IC4 and extremely good at using Claude to do their work and map out, of course, we would bring them on as well. So there is, IC4, and extremely good at using Cloud to do their work and map out. Of course, like we would bring them on as well. So there is, I think, a continued role for people that have embraced these tools to make themselves, in many ways, as productive as a senior engineer. And then their job is, how do you get mentored so you actually acquire the wisdom and experience
Starting point is 00:43:19 that you're not just doing seven hours of work to the wrong end, you know, or in a way that's going to be a spaghetti vibe-coded mess that you can't actually then maintain a year from now, or because it wasn't just a weekend project. The place where it's less known, and I think something that we'll have to study over the next several months to a year is for the jobs that are more, is it data entry, is it data processing,
Starting point is 00:43:42 where you can set up an agent to do it pretty reliably. You'll need people in the loop there still to validate the work, to even set up that agentic work in the first place. But I think it would be unrealistic that the exact same jobs look exactly the same even a year or two from now. So, you know, as somebody who runs a business,
Starting point is 00:43:59 I get the appeal of having a sort of, you know, digital CTO, salesperson, whatever else, you know, these APIs will soon be able to do that could create a lot of like value in my life. At the same time, most people do not run businesses. Most people are W2 employees. And they email us when we have conversations like this. And they want us to ask really hard questions of folks like yourself.
Starting point is 00:44:27 And I think it's because they're listening to all this and they're just like, why would I be rooting for this person, right? Like this person is telling me that he's coming to take my job away and he doesn't know what's gonna come after that. So I'm curious how you think about that. And like, what is the role that you're kind of playing
Starting point is 00:44:42 in this ecosystem right now? Yeah, I think for as long as possible, the things that I am trying to build from a product perspective are ways in which we augment and accelerate people's own work, right? I think there's been different players will take different approaches and I think there'll be like in marketplace of ideas here.
Starting point is 00:45:00 But when we think about things that we wanna build and from a first party perspective, it's all right, are you able to take somebody's existing, you know, application or their role and like be more of themselves, right? A useful thought partner, an extender of their work, a researcher, a augmenter of how they're doing. Will that be the role AI will have forever? Likely not, right?
Starting point is 00:45:21 Because it is going to get more powerful. And then, you know, if you spend time with people who are like really deep in the field, they're like, oh, you know, and eventually like, you know, as we'll be running companies, I'm not sure we're there yet. I think the eyes lack a lot of sort of like organizational and like long-term discernment to do that successfully.
Starting point is 00:45:38 I think, you know, it can do a seven hour refactor. It's not gonna get conceptualized and then operate a company. I think we are years away from something like that. So I think there's choices you can make around what you focus on. And I think that's where it starts. Whether that's the thing that makes it so that they're perfectly complementary forever, likely not.
Starting point is 00:45:58 But hopefully we're nudging things in the right way as we also figure out the broader societal question of how do we scaffold our way there? You know, what are the new jobs that do get created? How the roles change? Like, how does the economy and the safety net change in that new world? Like, I don't think we're six months through the year from solving those questions. I don't think we need to be just yet, but we should be having the conversation now. I think this is one place where I do find myself getting a little frustrated with the AI safety community in that I think they're very smart and well-intentioned when it comes
Starting point is 00:46:34 to analyzing the risks that AI poses if it were to go rogue or develop some malign goal and pursue that. I don't think the conversation about job loss and the conversation about AI safety are close enough together in people's minds. I don't think, for example, that a society where you did have 15 or 20 percent unemployment for early career college graduates is a safe society. I think we've seen over and over again
Starting point is 00:47:08 that when you have high unemployment, your society just becomes much less safe and stable in many ways. And so I would love if the people thinking about AI safety for a living at places like Anthropic also brought into that conversation the safety fallout from widespread job automation. Because I think that could be something that catches
Starting point is 00:47:30 a lot of people by surprise. Yeah, we have both our economic impact, kind of societal impacts team and our AI safety team. I think it's a useful nudge around how those two come together. Because there are second order implications on any kind of major labor changes. Are you guys in the conversations
Starting point is 00:47:46 with policymakers, regulators, sort of trying to ring alarm bells? Are you hearing anything back from them that makes you feel like they're taking you seriously? I'm not in the policy conversations as much being more on the products. I do think those conversations are happening, and there is more.
Starting point is 00:48:00 It's this interesting thing where the critique a year ago, maybe it's changed a bit, was, oh, you guys are talking your own book. You're like, this is not going to happen. It's all hype. And probably some of it was folks hyping it up, at least the kind of alarm bells or, you know, signals that I've seen at least coming out of an outbreak. Like, no, we think this is real. We think that we should start reckoning with it.
Starting point is 00:48:24 Believe it or not. Like, like, even if you assume it is like a low probability thing, shouldn't we at least have a story around what that looks like? You were one of the co-founders of Instagram, Instagram, very successful product used by many, many people. But social media in general has had a number of negative unintended consequences that you may not have envisioned back when you were first releasing Instagram. Are there lessons around the trajectory of social media and unintended harms that you take with you now into your work on AI? I think you have to reckon with these. I mean, AI is already globally deployed and has at least one billion issues or products. So it would be silly to say like it's early in the AI adoption curve,
Starting point is 00:49:05 but it actually is early in the AI adoption curve. I think with social media, when it was me and Kevin taking photos of really great meals in San Francisco, with our iPhone 3GS, like, you know. Kevin's sister, not me. I don't know, you were probably early on Instagram maybe. Yeah, Casey definitely was.
Starting point is 00:49:22 You were the hipstomatic guy. The important thing was you just would just never invite this Kevin to dinner. Yeah. But you were, okay, maybe. Yeah, Casey definitely was. You were the hipstomatic guy. The important thing was you would just never invite this guy to dinner. Yeah, exactly. Yeah, but you were, okay, yeah, so back in those days. Yeah, you could kind of maybe extrapolate and say, all right, if everybody uses this, what would happen? But I almost didn't feel like the right question to ask.
Starting point is 00:49:37 And the challenges that came at scale, I think as a platform grows that large, it just becomes much more a mirror of society when all of it's both, you know, positives and negatives. And it also enables new kind of unique behaviors that you then have to mitigate. But yes, you could have foreseen it at scale. I'm not sure you would have designed, maybe you would have designed different moderation systems along the way.
Starting point is 00:49:59 But at first you're just like, there's 10 people using this product. Like I don't, we just need to see if there's a there there, right? AI feels much different because one, on an individual basis, the reason we have the responsible scaling policy is that for biosecurity, that doesn't involve a billion people using cloud for, or an AI for something negative. It could just be one person that we want to actually make sure we address and mitigate.
Starting point is 00:50:21 So the sort of scale needed from a reach perspective is really different. That I think is very different from the social media perspective. And the second one, at least for Claude, which is primarily a single player experience, the issues are less relational, right? Like with Instagram, the harms at scale come, like if you only used Instagram in a private mode with zero followers, maybe you'd feel quite lonely. Maybe that's a whole separate thing. But the kinds of things that you might think about in terms of bullying among teenagers or body image,
Starting point is 00:50:49 like those wouldn't really come up really if you're not really looking at it, if you're using it as an Instagram diary, right? AI, you can have much more of that individual one-on-one experience and it is single player, which is why like, you know, there's a really thought provoking, you get an internal essay just recently around,
Starting point is 00:51:05 we shouldn't take thumbs up and thumbs down data from, you know, Anthropic or from cloud users and think of that as the North Star. Like we aren't out here to please people, right? And we should be, we should fix bugs and we should fix places where the, the, that didn't succeed. But we shouldn't just be out there telling people what they want to hear if it's not actually the right thing for them.
Starting point is 00:51:26 So this is something I've been thinking about a lot because, you know, there are many people today who have the experience of Instagram of like, I like this a certain amount, but I feel like I look at it more than I want to and I'm having trouble managing that experience and so maybe I'm just going to delete it from my phone. I look at where the state of the art is with chat bots and I feel like this stuff is already so much more compelling in some ways, right? Because it does generally agree with you. It does take your side.
Starting point is 00:51:51 It's trying to help you. It might be a better listener than any friend that you have in your life. And I think, you know, when I use Claude, I feel like the tuning is like pretty good. I do not feel like it is sycophantic or, you know, sort of being like, like very obsequious, but I can absolutely imagine someone taking the Claude API and just building that and is pretty good. I do not feel like it is sycophantic or being very obsequious. But I can absolutely imagine someone taking the Clot API and just building that and putting in the app store as fun teen chat bot 2000.
Starting point is 00:52:14 How do you think about what the experience is going to be, particularly for young people using those bots? And are there risks of whatever that relationship is going to turn out to be for them? Yeah, I think if you talk to like Alex Wang from Scale, he's like, in the future, most people's friends will be AI friends. And I don't necessarily like that conclusion, but I don't know that he's wrong also if you
Starting point is 00:52:37 think about like the availability of it. And I think it's really important to have relationships in your life around people that will disappoint you and be disappointed by you. That's this relationship you're looking at. Imagine, it was just pure AI. It would never be the same, right? And so, I think maybe two answers there. One, we should just confront it and be really vocal about it, not just pretend that it's not happening, right? It's like, what are the conversations that people are having with AI at scale? And what is, what do we want as a society?
Starting point is 00:53:07 Like, do we want AI to have like some sort of moderator process? It's like, hey, your conversation with this particular AI is getting a little too, you know, real, weird. Like maybe it's time to step back. Like it will Apple eventually build the equivalent of screen time. That's more like AI time. I don't know. It's like there's a bunch of interesting privacy questions around the world, but maybe that
Starting point is 00:53:28 is interesting even for parents. Like, how do you think about moderating the experiences that your kids have with AI? It's probably going to be at the platform level, right? It's going to get into your apps, for example, is an interesting one. That will be a really fascinating question. And then the second piece is, you know, as we think about moving up the, you know, safety levels thing, I mean, the Responsible Scaling Policy is also a living document. Like we've iterated on it and added to it
Starting point is 00:53:50 or refined the language. I think it will be interesting to think about, and manipulation is one of the things that's in there and something that we look for in deception, but also like over-friendliness, I'm not sure exactly the word I'm looking for, but that sort of like over-connectivity. Glazing, I believe, is the industry term of art. You know, that sort of like over-connectivity. Glazing, I believe, is the industry term of art.
Starting point is 00:54:05 You know, that sort of like over-reliance, I think, is also an AI risk that we should be thinking about. Yeah, so if you're a parent right now of like a teenager, and you find out that they're speaking with a chatbot a lot, what is your instinct to tell them? Is it you need to sort of supervise this closer, like read the chats, or maybe no, don't be too worried about it, or like unless you see this thing,
Starting point is 00:54:23 don't worry about it. It depends a little bit on the product. I mean, you have to, especially with Cloud, which currently has no memory, which mostly is a limitation of the product, but also makes it so that it's harder to have that kind of deep engagement with it. But as we think about adding memory,
Starting point is 00:54:37 what are the things? I've thought about one of the things that I'd like to do is introduce a family plan where you have child or teen accounts, but with parent visibility on there. Maybe we could even do it in a privacy preserving way where it's not like you can read all your teens chats so that maybe that's the right design. But maybe what you can do is have a conversation with Claude
Starting point is 00:54:54 that also can read the teens chats, but does it in a way where like, it might not tell you exactly what it felt, what your teen felt about you last night when you like told them no, but it will tell you like, hey, this behavior over time, I'm flagging something to you that you need to go and follow up.
Starting point is 00:55:08 Like you can't abscond responsibility from the parent though. Right, actually, I mean, that's really interesting if the bot could say something like, your teen is having a lot of conversations about disordered eating, you know, or something. Yeah, I wanna think more about that. My last question, earlier before you got here, Kevin and I had a huge fight
Starting point is 00:55:24 because I thought it was Claude for Opus, and then he was like, Kevin and I had a huge fight because I thought it was Claude for Opus and then he was like, no, it's Claude Opus for, and he turned out to be right. So why is it like that? We changed it partially because this was a vigorous internal debate. There was some we really spent our time on as well. I agree to one, aesthetically, I like it better, but it was tending towards it. Also like we think over time, we may choose to release more opuses and more sonnets, and having the major,
Starting point is 00:55:50 the big important thing in the thing be the version number, kind of created this thing where like, well, you had Claude 3.5 Sonnet, why didn't you have Claude 3.5 Opus? And it was like, well, we wanted to make the next Opus really worthy of the Opus name. And so maybe flipping the priority in there as well. But the team crazy because now our model page is like,
Starting point is 00:56:08 you have Claude 3.7 Sonnet and Claude Sonnet 4. Like, what are you doing? I feel like we can't go unreleased not doing at least something mildly controversial on naming. And as the person responsible for Claude 3.5 Sonnet V2, I hope we're getting better. And hopefully the AI can just name things in the future. Let us hope. Mike Kruger, thanks forigor. Thanks for coming. Thanks for having me.
Starting point is 00:56:30 When we come back, we're headed to court for Hartford Crimes Division. from time to time we like to check in on the miscreants, the mischief makers, and the hooligans in the world that we cover to see who out there is causing trouble. Yes, it is time for another installment of our Hard Fork Crimes Division. Let's open the case files. All right, Casey, first on the docket, Metta rest its case. After a six-week antitrust trial, the case of the Federal Trade Commission versus Metta platforms has wrapped up and is now in the hands of Judge James E. Boesberg,
Starting point is 00:57:34 who has said that he will work expeditiously to make a judgment in the case. Casey, how do you think Metta's antitrust trial went? Well, so if you're just catching up, Metta of course has been accused of illegally maintaining its monopoly in a market that the FTC calls personal social networking. And they did this by acquiring Instagram and WhatsApp in the early 2010s. And the government has said that prevented a lot of competition in the market, that introduced
Starting point is 00:58:02 a lot of harms to consumers, such as the fact that we have less privacy, because that's just kind of not an axis that there is any companies left to compete over. And the government spent a lot of time making that case, but Kevin, I'm not sure it went that well for them. Yeah, do you think Metta's gonna win this one? I think Metta has a really good chance. Your colleague, Cecilia Kong, noted in The Times
Starting point is 00:58:23 that Metta called only eight witnesses over four days to bat down the government's charges. When you consider how much revenue Instagram and WhatsApp generate for Metta and what a sort of existential threat to their business it would be to have to spin these things off. I thought it was pretty crazy that they felt like they had made their entire case in four days. Well maybe their case was so simple and straightforward that they didn't need to do anymore. Well, maybe they just wanted to frame it in terms of a real. Yeah, they did a short form antitrust trial. That's a huge right now. Um, well, look, I think the,
Starting point is 00:58:56 the real issue here is that Metta's argument is pretty simple. They're saying, we face tons of competition. Have you ever heard of TikTok? The way this case is built, if the judge considers TikTok to be a meaningful competitor to Meta today, it may be extremely difficult for him to say, we're going to unwind a merger that, in the case of Instagram, took place 13 years ago. I guess we will see, very shortly,
Starting point is 00:59:16 whether this is an actual crime that belongs in the hard fork crime division, or whether this was just a Tampa's thinner teapot. Well, you know, sometimes criminals get away with things, Kevin. Moving on! Case file number two, the crypto gangs of New York. This comes to us from Chelsea Rose Marcius
Starting point is 00:59:34 and Maya Coleman at the New York Times. And they write that another suspect has been arrested in a Bitcoin kidnapping and torture case. And let me say right up front, this story is not funny. It is extremely scary. Not funny at say right up front, this story is not funny. It is extremely scary. Not funny at all. In fact, quite tragic.
Starting point is 00:59:49 There has been a recent wave of Bitcoin and crypto related crimes, people attacking people to try to steal their Bitcoin passwords and steal their money. This has been happening over in Europe, in France, in just the last few months. There have been several attacks on crypto investors, people with lots of money in cryptocurrency. These have been called the wrench attacks because criminals are coming after these investors and executives violently in some cases with wrenches. This most recent case happened in New York in the Nolita neighborhood of Manhattan where an Italian man named Michael Valentino Teofrostro Carturon was allegedly kidnapped and tortured for nearly three weeks in a luxury townhouse by criminals who were apparently trying to get him to reveal his Bitcoin password.
Starting point is 01:00:43 Casey, what did you make of this? Well, to me, the important question here is why is this happening so much? And the reason is because if a criminal can get you to give up your Bitcoin password, that's the ballgame. In most cases, there is no getting your money back. It can be relatively trivial for this money to be laundered and for there to be no trace of what happened to your funds. That is not true if you're just a regular millionaire
Starting point is 01:01:08 walking around town, right? Obviously, you may be vulnerable to robberies or other sort of scams or theft, but if you give up your bank password, for example, in most cases, you would be able to get your money back if it had been illegally transferred. So this is just a classic case of Bitcoin and crypto continuing to be
Starting point is 01:01:25 a true Wild West where people can just run up to you off the street and hit you over the head with a wrench and it's really scary. Yeah, it's really scary. And I should say this is something that I think crypto people have been right about. Years ago, when I was covering crypto more intently, I remember people sort of telling me that they were hiring bodyguards and personal security guards and It seemed a little excessive to me These were not by and large famous people who would like get recognized on the street but their whole reasoning process was that they were uniquely vulnerable because crypto is very hard to
Starting point is 01:02:00 Reverse once you've stolen it. It's very hard to get your money back from a criminal who steals it. And that meant that they were more paranoid than like a CEO of a public company would be maybe walking around. You know, I read a blog post on Andreessen Horowitz's website recently. So, you know, I was having a great day
Starting point is 01:02:20 and they've hired a former secret service agent to among other things, help crypto founders prevent themselves from getting hit over the head with a wrench. And he has sort of an elaborate guide to like the things that you could do. But my main takeaway from it is, if you're a crypto millionaire, you have to spend the rest of your life in a state of mild to moderate anxiety about being attacked at any moment, particularly if you're out in public. Yeah, I do think it justifies the sort of lay low strategy
Starting point is 01:02:47 that a lot of crypto entrepreneurs had during the first big crypto boom, where they would sort of have these like anonymous accounts that were out there that were them, but no one really linked it to their real identity. I think we are going to start seeing more people, especially in crypto, using these sort of synonymous identities.
Starting point is 01:03:05 I mean, this is one of the reasons that, you know, people say that Satoshi Nakamoto has never wanted to reveal him or herself after all these years is because there would be a security risk associated with that. But I think this is really sad and criminals cut it out. And here's my message to all the criminals out there. I don't own any crypto
Starting point is 01:03:23 and I will continue to not own any crypto. You can keep your wrenches to yourself. All right. Last up on the docket for today, this one. Oh, I love this one, Casey. I've been dying to talk about this one with you. Tell me. Elizabeth Holmes' partner has a new blood testing startup.
Starting point is 01:03:38 So Casey, you may remember the tragic story of Elizabeth Holmes, who is currently serving an 11 plus year prison sentence for fraud that she committed in connection with her blood diagnostic company Theranos. Because God forbid a woman have hobbies. Well, Elizabeth Holmes has a partner named Billy Evans. They have two kids together. And Billy is out there raising money for a new startup called Hamanthus, which is, drum roll please, a blood diagnostics company that describes itself as a radically new approach to health testing. This is according to a story in the New York Times
Starting point is 01:04:14 by Rob Copeland, who says that Billy Evans' company is hoping to raise $50 million to build a prototype device that looks not all that dissimilar from the device that put Elizabeth Holmes in prison, the Theranos mini lab. And according to this story, the investor materials don't mention any connection between Billy Evans and Elizabeth Holmes. Well, I wonder why that is.
Starting point is 01:04:39 I have to say, she does have some experience that is relevant here, Kevin. Why not lean on that? Now, do we know what Hamanthus means? Is that like sort of a name taken from historical antiquity, and we'll look it up, and it turns out it's like an ogre that used to like stab people with a spear or something? I assumed it was like ancient Greek for like, we're serious this time. According to Wikipedia, Kevin, it's actually a genus of flowering plants that grows in southern Africa.
Starting point is 01:05:09 But members of the genus are known as the bloodlily. And I want to say, is it too late to change the name of the company to bloodlily? Yeah, that one, I like that one better. I did spend some time this morning because I was, I was on my commute, just trying to brainstorm some better titles for this startup that was run by Elizabeth Holmes' partner and does something very similar to Theranos. All right, let me run these by you.
Starting point is 01:05:34 Blood test two, electric boogaloo. No. Fake tricks reloaded. That's a matrix reloaded. I like that it was high concept. Okay, here's one. Okay. Th's one. Okay. TheraYes. That's good.
Starting point is 01:05:51 Let's go with that one. Okay. Well, good luck to Billy Evans with TheraYes. $50 million. Andreessen Horowitz will give that to them. You know, they love to be contrarians. Yeah. I think, here's my prediction.
Starting point is 01:06:04 This startup is going to get funded and love to be contrarians. Yeah. Here's my prediction. This startup is gonna get funded and they're gonna release something. Yeah. And you're gonna have to figure out how to keep your family safe from it. Listen, if they're doing another Fire Fest, they're gonna do another Theranos.
Starting point is 01:06:14 You better believe it. We have learned nothing. Theranos is back. Well, Casey, that brings to a conclusion this week's installment of Hard Fork Crimes Division. Mm-hmm. And all the criminals out there, keep your nose clean, stay low. And that brings to a conclusion this week's installment of Hard Fork Crimes Division. And all the criminals out there, keep your nose clean, stay low, try to stay out of the funny pages.
Starting point is 01:06:31 You're on notice. Hard Fork is produced by Rachel Cohn and Whitney Jones. We're edited this week by Matt Collette. We're fact-checked by Ana Alvarado. Today's show is engineered by Chris Wood. Original music by Diane Wong, Rowan Nemus-Dow, and Dan Powell. Our executive producer is Jen Poyant. Video production by Sawyer Rokay,
Starting point is 01:07:21 Pat Gunther, and Chris Schott. You can watch this full episode on YouTube at youtube.com slash hardfork. Special thanks to Paula Schumann, Queeving Tam, Dalia Haddad, and Jeffrey Miranda. You can email us, as always, at hardforknytimes.com. Send us your ideas for a blood testing startup. Music

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.