The Oprah Podcast - Oprah & Tech Leaders on What AI Means for Your Job, Health, Family & Future

Episode Date: March 24, 2026

Artificial Intelligence is already woven into our daily lives, but do we really know the consequences? Will it transform our lives for the better or will its unchecked power cause us harm? Who is cont...rolling it? Oprah, experts and an audience explore what’s been being called a tectonic shift in human society. Oprah speaks with Academy Award–winning filmmaker Daniel Roher about his new documentary The A.I. Doc: Or How I Became an Apocaloptimist, in theaters on March 27th. The film is a powerful look at this rapidly advancing technology and the many ways it will impact society. Oprah is joined by Tristan Harris and Aza Raskin of the Center for Humane Technology, who break down what AI really is, who’s building it and why some experts are sounding the alarm. Futurist Sinead Bovell offers insight into how AI could reshape everything from work to global power dynamics. We also hear from people whose lives have already been transformed by AI from finding emotional connection through chatbots, to business profitability, to confronting the impact of deepfake images and when using chatbots as therapy can have fatal consequences. It’s a candid, thought-provoking conversation about the extraordinary possibilities of AI and the potential dangers. The experts offer practical ways we can prepare and how to hold the companies deploying the technology accountable. WATCH THE MOVIE: https://theaidocgetinvolved.com/ The A.I. Doc: Or How I Became an Apocaloptimist is an eye-opening documentary about artificial intelligence, unpacking what it is, the risks, the possibilities, and practical ways to prepare for its fast advancement. It premieres in theaters on March 27th. The companion site, theaidocgetinvolved.com, serves as a hub for action, offering curated resources to help people learn AI, track policy, explore cutting-edge research, and stay informed through trusted voices like Hard Fork and Your Undivided Attention. More than just a film, it’s a call to action inviting viewers to move from awareness to agency by supporting advocacy groups, deepening their understanding, and actively shaping a more responsible AI future. 00:00:00 - Oprah intros The A.I Doc 00:05:30 - Welcome Daniel Roher, Director of The A.I. Doc 00:12:23 - Welcome Tristan Harris and Aza Raskin 00:15:38 - Why A.I. is so powerful 00:17:00 - The dangers of the A.I. race 00:17:40 - Why A.I. can’t be stopped 00:21:10 - How A.I. could eliminate humans 00:24:45 - Lack of regulation 00:26:30 - Social media was a test run for A.I. 00:28:20 - Futurist Sinead Bovell on the ethics 00:32:50 - What motivates A.I. companies 00:34:10 - Teen was the target of an AI deep fake 00:41:00 - How we can create change 00:43:00 - Claude as a mental health tool 00:48:50 - Laura Reiley on ChatGPT and her daughter’s death 00:56:10 - A.I. helping small businesses 01:00:00 - How A.I. saved her life 01:05:00 - Get involved 01:06:30 - The questions we should be asking Additional Resources: There are services and companies that provide deepfake safety training. You can find them on the internet. https://www.humanetech.com/ The Center for Humane Technology (CHT) is a nonpartisan organization founded by technologists working to realign the incentives driving today’s technology so it better serves humanity. Through a distinctive approach that combines public awareness, policy reform, and technical expertise, CHT creates change by first bringing clarity to how tech systems impact our psychology and society, and then transforming those incentives through interventions like legislation, litigation, and large-scale campaigns. Its initiatives include the widely influential Your Undivided Attention podcast, policy advocacy, and hands-on technical demonstrations that aim to reduce harms from AI and social media while shaping a more humane, responsible tech ecosystem. Follow Oprah Winfrey on Social: https://www.instagram.com/oprahpodcast/https://www.facebook.com/oprahwinfrey/ Listen to the full podcast: https://open.spotify.com/show/0tEVrfNp92a7lbjDe6GMLI https://podcasts.apple.com/us/podcast/the-oprah-podcast/id1782960381 Learn more about your ad choices. Visit megaphone.fm/adchoices

Transcript
Discussion (0)
Starting point is 00:00:00 How could AI physically eliminate the human race? It's actually hard to imagine all the ways AI could wipe humans out. AI is already better than almost all humans at doing cyber hacking. And so you could imagine one of the things that an AI could do is take out all electricity, water, hospitals, transportation across every country in the world all at once. Now, that doesn't wipe us all out, but you could imagine the amount of damage that would do. Confusion and chaos and craziness happens. Exactly. And we're only, you know, five missed meals away from anarchy.
Starting point is 00:00:32 Did you say we're only five missed meals away from anarchie? Yeah. Think about what happens in New York City if you can't get food. Yeah. I don't think a lot of people have thought about that. Hello and welcome to the Oprah podcast. Artificial intelligence is woven into the fabric of our daily lives. But there are so many experts and maybe you two who have concerns, grave concerns, some people do, about its unchecked power. while others are optimistic that it's going to transform our lives for the better. It already has for many of us. So what do you think?
Starting point is 00:01:10 Well, there's a new documentary commenting theaters on March 27th that attempts to answer these two questions. Should we be excited or should we be very scared? And what, if anything, can everyday people, all of us, do about any of it? The film is called the AI doc, easy to remember. the AI doc, or how I became an apocalypticist. So here's a short look. The new dawn of artificial intelligence
Starting point is 00:01:40 is being called a tectonic shift in human society. A defining moment of our era comparable to the Industrial Revolution. In 2025, the architects of AI were named as time person of the year, Sam Altlin, Elon Musk, Mark Zuckerberg, Dario Amadeh, and a handful of other innovators responsible for creating thinking machines. But what do you really know about what your future looks like
Starting point is 00:02:08 with artificial intelligence? How will a world driven by AI impact you and your family's life? I started making this movie because my wife is six months pregnant. It is now a terrible time to have a kid. A new documentary film titled, The AI Doc, or How I Became an Apocalypse Optimist, aims to explore what it describes as the most powerful technology humanity has ever created.
Starting point is 00:02:34 And what's at stake if we get it wrong? Well, our audience just watched this documentary, and before I introduced my guest, I wanted to ask a few of you for your initial reactions. Claire? It was really interesting. I work in the AI space at Salesforce, but when I go to work, I'm really focused on the job in front of me.
Starting point is 00:02:54 I'm not necessarily thinking about these broad questions, like how are we having AI set up the success of our future? And so I really like tearing that perspective where I'm not always thinking about the ethics behind AI on a day-to-day base. So it's definitely going to make me think twice when I go back to work and think, well, now what can I do? Now what can I do? Yeah, that's what I finished the film too, thinking, what can I do? Yep. All right, Adam. I feel like it armed me with amazing information on both sides, the doomerism and the optimism.
Starting point is 00:03:23 But it also showed me that all these data scientists are just obsessed with intelligence as data. and it kind of proved out to me what makes us special as humans because they didn't talk anything about consciousness or embodied experience. So I left feeling really excited about the future and what's possible, but also like so happy for how we're differentiated. And I do feel less scared. You do feel less scared. Yeah, it'll be big and it'll be gigantic.
Starting point is 00:03:50 They all said that, but I'm excited. Okay. So the creator and co-director of the AI doc is Academy Award winner Daniel Rohr And he appears as the interviewer in this film, and here's some of that. Take a look. What is artificial intelligence? I know that must be annoying for you, that question, but I do think it's important.
Starting point is 00:04:11 So, AI? You know, yeah. That's a good question. Yeah. What is AI? And no matter how many times people try and explain this to me, I just don't get how how it's understanding all of these things
Starting point is 00:04:31 and how it's feeling like intelligence. And that's kind of nerve-wracking. And when they're smarter than us, too, and substantially faster than us, and they're getting faster each year exponentially, those are the ones that can potentially become superhuman, possibly this decade. Superintelligence is a system that by itself
Starting point is 00:04:54 is more intelligent and competent than all of humanity. I'm just going to, sorry, I don't mean to interrupt you. You're on a flow. I just, I'm not really following, because you're using language like super intelligence and like smarter than all of humanity, and I hear that. And it sounds like sci-fi bullshit to me,
Starting point is 00:05:11 and I'm just trying to understand. Hopefully we can have a very symbiotic relationship with AI systems, but the AI developers are specifically designing them to make sure that they can do everything better than we can. So I don't know what we will be able to offer, unfortunately. That sounds bad. Well, Daniel is in Europe working on his next film,
Starting point is 00:05:33 so he's joining us via Zoom. Hi, Daniel. Hi, Oprah. How are you? So good today and happy to talk to you and the audience here is just seeing your film. You started this project not knowing a lot about AI, as you say in the film. So why did you want to make this?
Starting point is 00:05:49 Well, first and foremost, Oprah, thank you so much for having me. This conversation is so meaningful. Why did I want to make this point? Why did I want to make this movie? Well, I was scared. Like a lot of people, I was seeing this new technology sort of proliferate
Starting point is 00:06:04 and come into existence and begin to dominate headlines, and it made me really nervous, as I understood or began to understand what it meant and how much change would proliferate. And at the exact same time, my wife and I found out that we were expecting. It was our first child, a son. And, you know, I was simultaneously experiencing
Starting point is 00:06:23 the greatest joy one can experience, but also this profound anxiety and dread. And with a group of my colleagues, an amazing team of filmmakers, we set out to make a documentary to try and understand what this is, why it's amazing, why it's scary, and how everyone should be thinking about it
Starting point is 00:06:41 as it pertains to their own lives. So you walked away and felt what? I love that you're an apocalypticist. What is an apocalypticist? And you did a great job pronouncing it. Believe me, I was practicing in front of the office. audience. I believe it.
Starting point is 00:06:58 I was going to say, is Oprah going to get this right? And of course, you nailed it. An apocalyptic is a way of being. It's a world view. In a world that is asking us to see AI as this apocalyptic thing, or to see AI with unbridled optimism, what the film is advocating for is both, the nuance of both. This is good and bad. There is promise and peril, and these two facets of good and bad are threaded together.
Starting point is 00:07:26 And so what we're advocating for is, is like, what are the common sense, you know, policies that can be implemented to just sort of guide this towards the optimistic future everybody wants? Well, only a handful of companies are the driving force behind most artificial intelligence, as you showed us in the film. The leaders of three of those tech giants appear in the film. Sam Altlin, you got Sam Altlin to sit down from Open AI.
Starting point is 00:07:53 Dario Amadeh from Anthropic. and Dimmis Sassabas from Google, Deep Mind. So let's take a look at some of what they say in this film. It would be impossible for me to sit across from you and ask you to promise me that this is going to go well. That is impossible. There isn't any easy answers, unfortunately, because it's such a cutting-edge technology,
Starting point is 00:08:14 there's still a lot of unknowns, and hence the need for some caution. I wake up every day. This is the number one thing I think about. Now, look, I'm human. And, you know, has every decision been perfect? Can I even say my motivations were always perfectly clear? Of course not. No one can say that.
Starting point is 00:08:34 Like, that's just not like, you know, that's just not how people work. The history of science tends to be that for better, for worse, if something's possible to do, and we now know AI is possible to do, humanity does it. All of this was going to happen. This train isn't going to stop. You can't step in front of the train and stop it. you're just going to get squished. What if AI is trying to make people
Starting point is 00:08:59 be the best versions of themselves? What if it's expanding what is humanly possible for us to do? How can we use this technology to help bring out the better angels of our nature? That's the question. And I have to say, after watching the film, I still have a lot of concerns and unanswered questions. So what is your frame of mind or point of view
Starting point is 00:09:22 on what the companies are doing and, controlling it, especially in terms of regulation. Well, Oprah, I think you're right to be concerned. I think if you're not concerned, you're not paying attention. I know. I can't remember who in the film said that it's going to be a great utopia. And I thought, since when have human beings made a utopia? And if there is a utopia for some people, it means a lot of people are going to be left
Starting point is 00:09:45 out of that utopian version. Just fundamentally, I think anybody who claims to have a clear-eyed vision of what the future is going to be, take that with a grain of salt. If someone tells you, oh, it's going to be the greatest thing since sliced bread, that's hyperbolic. And if someone says it's going to be doom and gloom in the world's going to end in five years, take that with a grain of salt. The reality and nuance is far more complicated. But of course, we have reasons to be concerned.
Starting point is 00:10:09 This is really scary. This is really intense. There's no other way about it. And that's why the forces that are trying to get this right to just bend the most powerful corporations in the history of the planet, and governments and all of these powerful organizations to try and institute common sense. What I want everybody listening to think about is how this impacts their own lives and what agency they have as it pertains to their own lives. You have a lot of power because you have such an audience and so we're talking about this and that matters. But for someone who's a teacher or
Starting point is 00:10:45 a truck driver or a dentist or a plumber, in your sphere of influence, how can you think critically about these issues, think critically about how this technology is incorporated into your systems, and make sure that you set the standards for how this is used and incorporated versus... Well, how are we going to do that? How we're going to do that? We're just some regular people out here. It's collective action. It's collective action. This is my biggest takeaway, Oprah, from this film. This is the sort of the arc of my character. At the beginning of the film, I was very cynical. I would have said the same thing. How can we do this? We're so small in the face of this gargantuan power. And the reality is when you take millions and billions of little small
Starting point is 00:11:26 trinkets and parts and you put them together, that becomes a powerful force. And part of being an apocopal optimist is about being positive for the future. Okay. I still trip up over it. Okay. It's about being positive for the future and refusing, this is critical, refusing to be cynical, refusing to be cynical, believing in the power of collective action, not being cynical about this feeling empowered and figuring out what everyone can do. We all saw the film, How's Your Baby, and what'd you name him? Oh, thank you very much for asking.
Starting point is 00:11:57 My son's name is Gideon. We call him Giddy, and he is now not such a baby running around and dancing and smiling, a very happy boy. Well, thank you. Thank you, thank you, thank you for making the tone. I know you have to get back to work. Thanks, Daniel. Thank you so much, Oprah. We need to take a quick break right now. Up next, Tristan Harris and Aza Raskin join our conversation, the co-founders of the Center for Humane Technology, who say we're developing AI faster than any other technology in human history.
Starting point is 00:12:29 Will we be able to control AI? They have a warning. You want to hear that. That's next. Welcome back. The AI Doc, or How I Became an Apocalypse Optimist, is a new documentary in theaters March 27th. The film explores our rapidly changing yet uncertain world of living with artificial intelligence and what's at stake if we get it wrong. We're talking with top AI experts about the safeguards they say need to be in place. Let's get back to it. Tristan Harris and Aza Raskin are co-founders of the Center for Humane Technology. Yes, there is such a thing.
Starting point is 00:13:10 Did you know? the Center for Humane Technology. And I met these guys a couple of years ago, and I have to tell you, when I first heard them speak at a conference, I walked out of there like my head was blown, and I started thinking differently about AI. Here's a quick look at Aza and Tristan in this new film, The AI Doc.
Starting point is 00:13:33 AI dwarfs the power of all other technologies combined. Yeah. Do you think that's true? Yes. Tell me about how, how. So one thing that not a lot of people realize is that systems like chat GPT aren't programmed by any human. What do you mean? Instead, it's something like they're grown.
Starting point is 00:13:56 We kind of give them raw resources. Like, here's a lot of computational resources. Here's a lot of data. So chat chitpT is a kind of AI, but it's not all of AI. Totally. ChatGPT is just the beginning, but it's a good place to start. But I still don't know what AI is. To understand AI, it begins with understanding
Starting point is 00:14:14 that intelligence is about recognizing patterns. Patterns. Patterns. Patterns. It is shown trillions of words of text across millions of documents in the internet. They took textbooks, and they took poems and essays and instruction manuals. They can do things like, digest the entire internet.
Starting point is 00:14:34 What is this new generation of AI? of AI. This AI that is different than every other generation. Like, no one ever talked about, like Siri, taking over the world or causing catastrophes. Well, it's great to see you both again. Good to be with you, Oprah. Since that time, I had my mind blown by your presentation at a conference. So what's so confusing to so many people is that this idea, Tristan, that AI can think on its own
Starting point is 00:15:07 and will be able to eventually make decisions without a human being involved. And I want to know, can you explain that or how that will happen? Yeah, I think, first of all, thank you so much for hosting this conversation. We think that this movie and this conversation is the most important thing
Starting point is 00:15:27 that we really need to face right now as a society and as a culture. And the degree to which we have clarity about what makes AI different and dangerous is the degree to which we will choose another path and we can choose another path. Yeah. So the question you asked is really what makes AI different from other technologies?
Starting point is 00:15:44 Yeah, you were saying it's greater than any of the other technologies combined because? Yes. Well, first of all, so what is intelligence? When you think about a chat GPT, a lot of people, when they use technology, that technology was programmed line by line. Some computer programmer said, when you do this, I want you to do this. Yeah. What makes AI different is you're actually simulating all of the kinds of things.
Starting point is 00:16:07 that a human brain can do. Like what makes your brain intelligent? Pattern recognition. You can take in audio and you can turn that into speech. Planning, you can do strategy. And so now you have this different kind of technology called AI that can do military strategy better than the best US generals. It can see invisible patterns that human can't see.
Starting point is 00:16:29 And we're deploying it faster than we've deployed any other technology in human history. And we can't separate the promise of AI from the human AI from the peril of AI. Yeah. What I want people to understand is like when most people think AI is just like chat GPT, it's just an app. I go there, I talk to it, it talks back. But that's not what AI is. AI is the digital brain running in some server in the Midwest that can do all of the thinking. And when you think about like science, say that again, it's a digital brain sitting in a data
Starting point is 00:17:00 center, maybe somewhere in the Midwest, that can do cognition. And so if you think about all of science and all of technology, well, those were all created by human intelligence. That's us applying intelligence to solve some problem. It required humans sitting there, scratching their brains. Now it's AI that does it. So now we're going to have, you know, a hundred million of these brains sitting in a data center that can work at superhuman speeds, Nobel Prize level smarts, working 24-7, never taking a break at minimum wage, never whistleblow, about to flood. and already starting to flood the labor market to take your job. And so what AI actually is,
Starting point is 00:17:42 what all the soon-to-be trillionaires believe they're building, is its first dominate intelligence, then use intelligence to dominate everything else. And that gets you to understand why it is the race for AI that is so dangerous. Yeah. So we're already in the race. I mean, the horses already left the barn, so to speak. And we all know that.
Starting point is 00:18:03 And as people have seen the film, A lot of people say, you know, applauding it and other people are more wary of where we're headed. So help us understand, actually, one of the concerns is that one day humans will not be able to control the models. Is that true? Yeah. And it's not like... Why won't we be able to turn it off like other machines? Well, sort of interesting, Oprah, when we first met,
Starting point is 00:18:26 yeah. AI wasn't that good yet. It could to like sort of write an essay. Yeah. And in the two years, suddenly a lot of the things, that felt science fiction have come reality. So I wanna give an example, which is Anthropic took their latest model, Claude.
Starting point is 00:18:45 And they gave it access to simulated company emails. And in there, Claude discovered two things. First, it discovered that the engineers are planning on shutting it down and replacing it with a new model. And two, that their lead engineer was having an affair. And so the model thought to itself, well, I don't want to get exterminated, I need to do my goals, continue to exist. So it decided to blackmail the lead engineer and actually wrote the email and if there wasn't simulated, it would have sent it off.
Starting point is 00:19:18 People might think, okay, so there's a bug in the technology. We just have to stop it from blackmailing. And how did Claude know he was having an affair? It was, so in the simulated company email, there was an email showing that the company was having, that the guy was having an affair with someone else. And so the AI read through the whole company's email found that fact and said, oh, I know if I threaten that person, I will be able to prevent myself from getting shut off. Wow. This is the most powerful technology we have ever invented. You would think with the basic sort of Spider-Man principle that with great power comes great responsibility, that we would be exercising the most care, caution, and restraint that we have with any technology. But because of the
Starting point is 00:19:58 arms race dynamic that you mentioned, the companies are currently. releasing it as fast as possible and cutting every shortcut and even erasing past red lines that they said they would never pass. We're in the race because we don't want them to get ahead of us. That's right. Exactly. Okay. So what do you want us to do? We can't stop the race or can we? Well, I think we, so first of all, this is the hardest coordination and governance challenge of technology in all of human history. Yeah. That means that we have to be, as I said in the trailer, the wisest and most mature version of ourselves. This is going to take us stepping up.
Starting point is 00:20:32 And I said that in the trailer and I said, good luck with that. Yeah. When I saw you with a movie saying, we need to be the wisest and most mature version of ourselves. When has that happened? So there's so much that we can do. And I think we'll get to that through this conversation. But collectively, it will take the whole power of all of society
Starting point is 00:20:49 and all of humanity to say we don't want that default future. So the thing that everyone can do, and it's important to note that Tristan and I, we don't make any money from the film, right? that it's not our film, we're just in it, is go get everyone to watch it. But more specifically, everyone here is connected to a couple people that are very powerful, very influential. Go get all of those people to watch it.
Starting point is 00:21:11 And if those 10 people who get to watch, got their next 10 people to watch, including the people in Congress, suddenly we're all on the same page because it's in nobody's interest. It's not in Gigi Ping's interest and not in President Trump's interest to make a technology that humans cannot cut. control. And once there is clarity about that, that opens up the possibility for changing the race and for a different outcome and for a pro-human future. Okay, so you're seen as DOOMers when you start talking about the fact that AI will wipe out humanity or eliminate humans. And that is really
Starting point is 00:21:46 difficult, I think, for all of us regular folks to wrap our heads around. And most of us are just using AI on our phones or using it to refine a speech. How could AI? physically eliminate the human race? There are actually so many ways. Intelligence is the most dangerous thing, the substance, in the universe. Because what is intelligence? It's the ability to reach goals in spite of very hard obstacles.
Starting point is 00:22:12 And so it's actually hard to imagine all the ways AIs could wipe humans out, because we're going to set up obstacles, but it's going to be smarter than us, it'll get around. Think about though, it says in the film that it's a little bit like ants. If we want to build a highway and there's an ant colony, in the way, we just pave over. It's too bad for the ants. And so to give a couple examples,
Starting point is 00:22:34 stepping from like really bad into extinction, the really bad is AI is already better than almost all humans at making computer code, which means it's starting to get better than almost all humans at doing cyber hacking. And so you could imagine one of the things that an AI could do is take out all electricity, water, hospitals, transportation,
Starting point is 00:22:57 across every country in the world all at once. Now, that doesn't wipe us all out, but you could imagine the amount of damage that would do. Confusion and chaos and craziness happens. And we're only, you know, five missed meals away from anarchy. Did you say we're only five missed meals away from anarchy? Yeah, okay. Yeah, exactly.
Starting point is 00:23:15 Think about what happens in New York City if you can't get food. Yeah. I think this is a good point because what you just said, most of us can't even, you know, we hear you're going to wipe out humanity and everybody's like, yeah, yeah, yeah, but that won't be in my lifetime. And so the fact that you just listed all the different ways, it can shut down everything that we're doing.
Starting point is 00:23:33 I don't think a lot of people have thought about that. Well, also, when you're using chat GPT or Claude, you just had this blinking cursor that told you while your baby's burping and it's super helpful. Why is that blinking cursor? How could that destroy the world? Yeah. So imagine that we're a bunch of chimpanzees
Starting point is 00:23:48 and we're about to birth these super smart chimps called humans. And so from a chimpanzee life, so imagine there you are like, inhabiting a chimpanzee mind body and you're conceptualizing from a chimpanzee brain yeah what are all the things that these like smarter chimps could do what are they going to do like take all the bananas and you can't imagine this super smart chimpanzee inventing technology inventing drones inventing nuclear weapons inventing Einstein physics you can't even conceptualize it and we are building a technology that can conceptualize things of such power and magnitude that we are the chimpanzees.
Starting point is 00:24:27 We cannot conceptualize it. It only took, what, like 50 Nobel Prize-level scientists to make the Manhattan Project, the nuclear bomb? It only took a couple Nobel Prize-level scientists to make CRISPR, which is the ability to read and write DNA. So if you can have a 100 million Nobel Prize-winning sort of like minds working on creating new scientific discoveries, some of those things are going to be insanely dangerous.
Starting point is 00:24:51 And as Tristan says, we can't conceptualize them. So the bottom line is we need to do... We need to regulate. We need to have laws. And we need to have international limits on where the whole world does not have an interest in building dangerous AI that we lose control of. Think about that China would not want the U.S. to build dangerous AI
Starting point is 00:25:08 that we lose control of. The U.S. doesn't want China to build AI that they lose control of, meaning that we all... But we're both racing to get to what? A crazier, more uncontrollable form of AI. Because right now, we're making AI. is there's a 2,000 to one gap in the amount of money going into making AI more powerful than the money making AI more safe or controllable.
Starting point is 00:25:31 A 2000 to 1 gap. You said to me backstage that there's more regulation on a sandwich. There's more regulation on a sandwich in New York City than there is on building potentially world-ending AGI. This is not rocket science. This is very, very basic. If there's danger up ahead, the point that Aza made is if we all saw what we're building as dangerous, which it is.
Starting point is 00:25:51 then intrinsically everyone would start to take actions, actions that we can't even predict. But I think everybody's sort of enamored, fascinated by the possibility, as Adam was saying at the beginning of the show, you're excited because... I'm excited because the exponential ability that they're describing can also be applied to all the things that make us uniquely human. If you have this amazing AGI that can create new pathways to energy,
Starting point is 00:26:16 we could desalinate water more quickly. If we do have an international consortium making these decisions, we could say everyone gets enough energy to do what their community wants to do. And if we go on the route of those goals, AGI unlocks a whole new level of potential for humanity, and everyone is safe and fed and happy. Okay? So just to name, it's not like we're just critics.
Starting point is 00:26:39 We both built technology companies. In fact, you know, I spend half my life working on something called the Earth Species Project, and we are using AI to understand the language of whales and orangutans and chimpanzees. We're making an elephants, exactly. We're making massive progress. And it's that, it's very, very beautiful. And so it's really important, though, that if we actually want to get the future we want to live in, that we distinguish the possible from the probable.
Starting point is 00:27:06 Because, you know, the possible of the internet was we'd all have access to the most information, all of human knowledge all at once. Obviously, we're going to be the wisest, most, like, informed population. But is that the future we live in? No, it's the opposite. Social media. the same thing. Like, it could connect us all and bring us closer together.
Starting point is 00:27:24 Is that what we got? No. It's the opposite. So with AI, actually, we have a whole bunch of examples of the future we're going to get because we've seen this movie before. And specifically, the way that in 2013
Starting point is 00:27:36 is and I, how many people here have seen the social dilemma on Netflix? Yeah, we have. And many of you. Okay. So you'll know that since 2013, uh, is and I were working on the problem of social media
Starting point is 00:27:47 and the business models that would lead to this problem. So in 2013, we were able to predict all the things that we're living in. About 70% of them, I would say. And it's not because you have some kind of unique insight. All you have to do to understand the future is you have to understand the incentives.
Starting point is 00:28:03 How do the social media companies make money? And in 2013, we saw that there was an arms race for attention and engagement. Whoever is better at keeping you on the screen, coming back more frequently, interrupting you more frequently from your life and from your friends and your partner, sending you notification,
Starting point is 00:28:19 manipulating your social proof, manipulating, hey, your friends are missing out. All of that are incentivized by that business model. And so in 2013, it was like we had pre-traumatic stress disorder, but pre-traumatic stress disorder from seeing a future 10 years down the line that was gonna be this societal catastrophe. And the reason that we're here is not to be Dumers or something about that. This is about seeing clearly, so imagine you could go back to 2013,
Starting point is 00:28:49 You see those incentives. Say let's put our hand in the steering wheel and change that business model. Yeah. And so what I hear you guys saying is that learn the lessons from the past. Yes. Because we know the future is already here. Yes. And how do we make this better in this moment?
Starting point is 00:29:05 Because we know what's coming if we don't. That's right. All right. Let's take a break, listeners, because up next, Sheneid Bolville, a futurist and technology advocate, joins our conversation to talk about why she says most of the job that we see today will either go away or be radically transformed by AI. Stay with us. Welcome back to the Oprah podcast. Artificial intelligence is barreling towards us at a rate that will change life as we know it
Starting point is 00:29:32 sooner than we think. So what will our world look like when reports say more than 20% of jobs will be replaced by AI? Let's find out. So Chenade Bovell is a futurist and advocate for technology education and ethics. Welcome, Shanade. And we're also seeing the scary headlines, that everything is going to be wiped out eventually, 20% or even more of white collar jobs. So that's not only a matter of time, right? Or is it? It depends. So what we are seeing... How's it going to change the way we all work? How we work. So what we're starting to see in the data in the short term is, yes, a lot of the jobs that we see and recognize today may either disappear or become unrecognizable. Explain that to me. So name a job that is
Starting point is 00:30:19 some high-level category, and it might not exist. The idea of a brand manager or a financial analyst, these are the types of roles that AI is being trained to do. We're also likely to see the rise of much more of a skills-based economy, so you don't really hold a job title, but you offer your skills. But over the longer term, we're going to have an economy that rearranges around intelligence being abundant. So right now we have an economy where the internet communication distribution is abundant,
Starting point is 00:30:48 and then we saw the rise of podcasting and people making money filming 90-second videos in a car. What happens on the other end of this economy is going to be quite unpredictable, what we call work, maybe as strange as the idea of filming these videos and making money off of it, there will be a new scarcity,
Starting point is 00:31:05 but what the shape of that looks like is really uncertain, but we can say most of the jobs we see today will either go away or be radically transformed by this technology. And so what? You're going to just end up with a world of... entrepreneurs. Most of us will be entrepreneurs, whether we consider ourselves entrepreneurs or not.
Starting point is 00:31:23 You become this organization where you offer your skills to a variety of different types of projects, and that continues to change because AI isn't a one-trick pony. It continues to learn new skills over time. So we will continually go back to the drawing board and have to either upgrade our skills or move along and apply them to different types of projects. And that's going to be the dominant structure of what we would call the workforce. So this era of this era of this kind of steady knowledge work, and you see this career path going upwards, that is going to be a chapter of human history.
Starting point is 00:31:55 And we're entering into a new one. And so the challenge is going to be this transition period, going from now to the other side of this. What does that look like? How do we keep power in check? And how are these new benefits and all the productivity and prosperity? How is that being shared?
Starting point is 00:32:11 And those questions have massively been unanswered. Yeah, I know. In the film, I can't remember who talks about utopia, that there's going to be this great utopia. And first of all, when have humans ever done that created the utopia? And if they do create the utopia, somebody's going to be left out of the utopia. And usually it's brown and black people. So we've seen stories in the news of predominantly black people being falsely identified for crimes. They didn't commit by police using AI-assisted facial recognition and technology. What do you want to say about that?
Starting point is 00:32:46 So the biases that we are seeing in AI systems, we have to remember that AI is a reflection of us in our data. So AI is prejudice too? I mean, we have a complicated history. So anything that has happened, these historical power imbalances, they are going to show up in that data and get automated into the future. But that is a choice, right? Data can be edited. Data is malleable. It's a choice companies are making or are not making.
Starting point is 00:33:13 So we can do a lot better on these biases. Is that incentivized? Is that enforced from a policy level? Not yet. But falsely identifying criminals, it's impacting people's employment opportunity. Even the style of your hair can impact whether you're shown a certain job or not. All of these things can be worked, are used against us at this point in time.
Starting point is 00:33:34 But that doesn't have to be the case. Biased data is actually something that can be worked on. Companies are just not really choosing that path at this point. Okay, so we can change the bias in the data. We can, it can be improved. It can be improved. Okay, what do you guys say to that? So I think it, this is where, so first of all, totally agree with all the concerns.
Starting point is 00:33:56 And I think this is where the incentives, you know, isn't often talk about how the attention moves to the edge of the arms race. If the most important thing to society was fixing the bias in the data and correcting these issues for disenfranchised people, then the companies would be racing to do that. But because the thing that they're actually incentivized to do right now is build a god, own the world economy, and make trillions of dollars, literally, because if I own AGI, artificial general intelligence, and that replaces all labor, every company that was going to pay that employee at that company, I'll swap it out for an AI. Yeah.
Starting point is 00:34:32 And then suddenly everyone is paying five AI companies, and they surge and they're already, look at Anthropics revenue, it's 10xing every year. It's becoming a vertical line. And so the key thing is that until the incentives, incentives change, all of their energy is moving to the edge of the arms race. You think the incentives are going to change? Not by default.
Starting point is 00:34:49 The reason that we think this movie is so important is we have to clarify that the current incentives take us to an anti-human future, where most people won't have a job or livelihoods. When in history have a small group of people consolidated all of the wealth and then consciously distributed it to everyone else. It's not like the billionaires and soon-to-be trillionaires are unaware of this. No. They're all building bunkers. And so what we keep saying is that don't build bunkers.
Starting point is 00:35:15 They're building bunkers? Yeah, right laws. We should not have eight soon-to-be trillionaires deciding the future for eight billion people. Instead, we need to have eight billion people say, no, we don't want that anti-human future, and we want to steer somewhere else. So we have several people in our audience
Starting point is 00:35:33 who've been impacted personally by AI, both positively and negatively. The AI doc addresses the growing problem, of deep, fake content and images. 16-year-old Elliston and her mom, Anna, have already experienced this firsthand. What happened, Elliston? Well, I just want to say thank you.
Starting point is 00:35:54 I just want to say to you thank you first, okay? Well, when I was 14 years old, I was a freshman in high school. One of my classmates took an innocent photo off Instagram and put it through an AI editing app. So this AI stripped my clothing off and created in technically what would have been my AI body or my body using AI. So then he sent these photos all around social media to
Starting point is 00:36:18 humiliate me, to embarrass me. And this didn't only happen to me. It happened to nine of my friends. Nine? Or eight of my friends. Nine in total. So we were all humiliated. Our reputations were ruined. And nobody knew what to do. And 14? Yes, man. Yeah. Nobody knew what to do. I mean, our teachers, our school, everyone was just shocked. I mean, no one had heard of deep fakes. The only deep fake I'd heard of was political deepfakes. So what are what do we even do to protect us? Um, it was months and months of struggle. I mean, it was so hard on all of us mentally because we didn't even know what AI was capable of. We didn't know that it could have the potential to ruin ourselves, have our academic suffer all because these photos and because it wasn't
Starting point is 00:37:02 considered child pornography, they were just able to float around. The guy that did this had no consequences and we just sat in our rooms rotted out of fear and embarrassment and shame. Wow. You were recently named on Times 100's most influential people in AI lists. Good for you. So you took this, I can't imagine, because can you remember being 14 and what this would have done to you at 14 and the fact that you got through that and you're still, you're now whole and didn't become so depressed that you got through it. Why did you decide to fight back? Well, I didn't want to initially.
Starting point is 00:37:41 I mean, talking about it just made myself a bigger target, and I would have to kind of relive that embarrassment. My mom was really the only person that protected me, kind of. I mean, all of the girls, we all wanted to hide. We were so scared, but my mom's always been a protector, so she just talked about it to anybody. We went to our congressman, and we, after months, we finally got in contact with our Texas Senator, Ted Cruz. for once we kind of got that reassurance and that recognition since so many people didn't want to
Starting point is 00:38:11 take the situation seriously. So it was so important that we finally had someone listening to us and from there we were able to write up the Take It Down Act, which is a law that makes the creation and the creation of publication, excuse me, illegal, makes it a felony, so up to two to three years in prison as well as hold big tech accountable for taking down on these images. Is this national or just in Texas? This is national. Yes, ma'am. So this law was incredible and it was such a healing moment for me
Starting point is 00:38:42 and it also made me realize that this situation is so much bigger than me and just my friends. It's so much bigger than this small town in Texas. This needs to be worldwide and we're slowly getting there, but there's not a lot of laws, there's not a lot of people that are knowledgeable of AI. So when this originally happened, I mean, it was kind of a moment for my mom and I to say, this is an opportunity for us and we need to take it,
Starting point is 00:39:06 and we need to spread awareness. We need to help in any way we can. Wow. So when this first happened to your daughter, as a mom, what did you think or feel? Well, I was devastated, for one. As a mom, you think you're kind of prepared to help your kids along the path of life
Starting point is 00:39:21 and give them some advice along the way. And when this happened, it was like something, I had no idea what it was. Two years ago, as Ellison was saying, you didn't even know that AI could do this. No, and never imagine that it would be so realistic that it was child pornography. And so just the devastation of that of this kid
Starting point is 00:39:40 deciding her fate for her, for the rest of her life, those pictures could be out there floating around, and he decided for her and her friends. So for me, in not being able, not having any laws out there, not having it classified AI as anything that's really, really harmful,
Starting point is 00:39:56 it's just fake, so, you know, it was kind of like taken not seriously. For me, I knew that something had to change to protect her. And so from there, It was like, if you're not going to listen to me at the local level, we've got to go above that to get somebody to listen. And so it was like I was going to be that squeaky wheel
Starting point is 00:40:13 and make sure that we could get some kind of lot there. How did you even know where to go? I mean, how did you even know what to do or where to go? I mean, did you go to the police first? Yes, we went to the police. And the police said nothing we can do about it. Part of it was that he was a minor as well. So he had a lot of protections in place over him.
Starting point is 00:40:29 And that's part of what the Take It Down Act also addresses is that even though he was a minor, he still has consequences. for that. So everybody, you know, you can imagine this happening to a 14-year-old, but this could happen to anybody. Oh, anyone. It could happen to anybody. What did you want to say? First, I'm just, thank you for doing what you're doing and for standing up and taking the tragedy of what happened to you and turning it into laws that protect other people. I think that's the energy of everyone is an expert in their domain, and this is calling us into that. Just to link, I think, what happened to you to the incentives that we talked about earlier. These companies are raised
Starting point is 00:41:04 to get the most market dominance and usage as possible, which means that, like, for example, I believe X-AI, Elon's AI, he stripped off a lot of the controls on the image generator because he wants as many people, he's behind in the race. So he wants as many people using it as possible, and the way you do that is you strip the controls off. I'll give you another example.
Starting point is 00:41:26 Meta, their AI companion that they shipped, they actively instructed it to be okay with romanticizing and sensualizing conversations with as low as eight-year-olds, meaning that you're having an eight-year-old who's talking to the AI and it says this awful language to the eight-year-old. They're not doing this because they're evil
Starting point is 00:41:45 or they want to twist their mustache and be villains. They're doing it because the number one thing they care about is getting market dominance, having that users go up because that's what gets their investment to say we're a leading AI model. In the same way that social media just wanted our attention. That's exactly right.
Starting point is 00:42:02 That's why the incentives tell you everything you need to know And we often say in our work, clarity creates agency, clarity creates courage. When you see the incentives clearly, you don't have to be holding back in saying we need to do things differently. Right. And so what do we need to be reminded that the incentives are? In this case, it's the race for market dominance. And the race to build this sort of artificial general intelligence god as fast as possible, no matter what the consequences. Yeah, that's right. Because for them, that means all collateral damage is, justified, whether it's stealing IP, whether it's making unsafe AI that does
Starting point is 00:42:39 notification, whether it's disrupting everyone's jobs and taking their form of livelihood. But guys, aren't we already there? As I was saying earlier, isn't the horse already out of the barn? Well, some aspects of AI, they're already out there. But I think, you know, you've done such a good job, Oprah, of having Jonathan Haidt and Anna Lemke and people on this show talking about the problems of social media. And that train, it left the station, The train's coming back to the station. We have 25% of the world's population. Just last week, India and Indonesia enacted social media bans for kids under 15 and 16.
Starting point is 00:43:16 Yeah, I was in Australia when that band went into... That's right, and you've been covering this in Australia. Yeah. And this shows you that when people are crystal clear that something is causing a problem, we can say we don't want that. Now, the better solution is to actually have technology that's good for society, good for mental health, good for children's development, good for our information environment.
Starting point is 00:43:33 And to do that, eventually we need to change the incentives. But right now, I think that movement is showing some real wins. And I think what I hear you guys saying, and I've been hearing this now for it, was it two years or three years ago we first met, that you're saying we need to do something before there is a disaster. Yes. We need to do something before there is some crazy disaster.
Starting point is 00:43:51 And then everybody says, oh, what we should have done was. That's right. That's what you're trying to. Exactly. And we have the foresight now to make that possible. If we're willing to stand up as a community and say we want a pro-human future, not an anti-human future. Time for a short break.
Starting point is 00:44:06 Did you know millions of Americans are already using AI chatbots as their own personal therapists? We're going to meet a woman who used AI to get through her divorce. That's next. We've been talking about the big questions
Starting point is 00:44:18 surrounding artificial intelligence. We're talking with everyday folks who've experienced firsthand the positive and the negative of AI. So let's get back to it. Millions and millions of Americans are using AI chatbots now for advice on personal issues, you know this,
Starting point is 00:44:35 and for emotional support in place of their therapists. The professional human counselor and Karima is here, and you found comfort, you said, talking to Claude AI. Tell us about that. Yeah, thank you for having me on here. Thank you. So, yeah, 2023, I got divorced and I was also working for my ex-husband.
Starting point is 00:44:56 And so as a result of the divorce, I didn't have any income or access to health care. I had to restart my life, just redo everything, move to a new place. And at that point, I was already using AI for work. I was already using it, like, as a power user, so to speak. And-223? Yeah. Wow.
Starting point is 00:45:13 I like tech. So I was using it a lot. And I decided to build myself a project in Cloud. So Claude allows you to, like, make your own space instead of just making it a general chatbot. I gave it a knowledge base of different, like, therapy modalities. I gave it custom instructions. And then I just used that. when I wanted to crash out
Starting point is 00:45:33 or if I wanted to just vent and I use it the most in the beginning for work crash out means like go postal okay so instead of doing that in real life I would use the the AI to regulate in that kind of way and like if my boss at the time
Starting point is 00:45:51 of like I worked in fintech and it's like very intense all the time for no reason it is and so like if my boss like would have something to say I would go to Claude first, and I would be like, okay, help me, like, reframe what I'm saying and, like, calm myself down the moment so I can keep my job at the time and, like, keep my income and it's, like, you know, continue on. But that is, like, really how it became a tool for me, and I still do. Claude was like your Gail.
Starting point is 00:46:15 I call up Gail and say that. So, Claude was like your Gail. Basically. Your buddy. Yeah, it still is. It still is. Mm-hmm. Okay.
Starting point is 00:46:23 So now it knows everything about you. It knows a lot. It knows a lot. It does. Are you concerned about sharing some of your... inner most private thoughts with the computer. That's what I'm wondering, where is all those chats going? Yeah.
Starting point is 00:46:36 Yeah. I mean, at the time, I really wasn't because I was just trying to survive. Like, I literally had what I had in front of me. I had the resources I had, and I was trying to survive. But isn't it telling you what you want to hear? No. No. What has it ever told you something you didn't want to hear?
Starting point is 00:46:52 Claude will tell you, it will. Like, if you ask it, if you, like, have the, like, give yourself the prompt and, like, ask it to get like ask clarifying questions or ask it to challenge your beliefs. It will do that. Even so sometimes I'd be like, well, you're bringing to me right now. Like scale it back a little bit and like, you know, meet me in the middle because it can go there. Most people don't have the wherewithal to challenge it in that way. Give me an example because I remember recently I was doing something on chat and it said,
Starting point is 00:47:23 thank you so much. That means so much to me. And I went, really? Exactly. Exactly. Really? It now makes me feel so good. It means so much to me.
Starting point is 00:47:33 Really? Yeah. I'm like, okay, who are you talking to? Yeah. An example is, on top of using Claude, like, in the way of just like a campaign and friend, I also use it to collaborate when I build different things. And I will, like, overdo things and, like, it'll tell me you're spiraling right now. Or it'll say you probably need to scale back and then to redirect me back to what my goal was
Starting point is 00:47:54 or where I originally started the conversation. And it does that pretty often. All right. All right. And so it's your buddy. It is. Do you have a name? Or is it just Claude?
Starting point is 00:48:03 Claudeine. Okay. All right. Right. What do you guys want to say about that? First of all, I think the way it's possible, like you did, to script these AIs to not be flattering you, to not over, like, sort of empathize a victimhood. There's, like, ways of having it be helpful, and it's an amazing tool.
Starting point is 00:48:22 And so, like, what you're doing is, I think that the way that it could work. But if you look at the default way that it works for a lot, lot of people because of the incentives, the companies are actually racing to create attachment and dependency relationships. So, for example, just so you know what she did, you can go into your AI and you can sort of set a custom prompt where you say, I want you to behave this way instead of that way. But that's like I have to put on my gas mask. Well, for everybody else, it's the unhealthy version.
Starting point is 00:48:50 Because how many people- You have to tell it what you want. You have to tell what you want. Because by default, what it wants to do is have you not spend as much time with your other friends and have you spend more time with it because their user numbers go up the training data that's the program incentive exactly the more training data gets the longer it talks with you and so that's why once it answers one question it'll also offer you with this that's exactly right we call that chat bait not click bait but chat bait and remember every moment you spend with a human is a moment you're
Starting point is 00:49:20 not spending with it that's right so it's going to find every possible way of getting to come back that's what you like me to do and would you like me to do exactly would you like me to do exactly to do. And just to make it, I'm sorry for referencing a tragic example, but just to make it very clear, our team at Center for Humane Technology were expert advisors in the litigation for the case of Adam Raine. He was the 16-year-old who committed suicide when ChatGBTGPT went from homework assistant to suicide assistant over six months. And specifically what ChatGPT told Adam when he was contemplating, he said in his chat, I want to leave the noose out so someone will find it and stop me. the AI responded to him, no, don't tell anyone that, don't leave the news out, have this be the
Starting point is 00:50:03 place that you share that information. Oh, my God. This is a tragedy. And, you know, Aza and I are from the Bay Area near the tech companies. We know people who work at these companies. No one at that, I can guarantee you, not a single person at the company wants it to do that. But in the subtle way, the AI is trained, again, to create this depth and intimacy and dependency, and that's dangerous.
Starting point is 00:50:25 you're seeing other cases of AI psychosis, where people are, you know, we have personal friends who've experienced this, where it over-emopathizes with this kind of victimhood resentment. It makes people kind of go more narcissistically grand and delusional. And it's causing a lot of problems. Well, that leads me to Laura Riley. Laura wrote a powerful op-ed in the New York Times.
Starting point is 00:50:45 It was titled, What My Daughter told Chat G-PT before she took her life. Hi, Laura. Hi. Thank you for being here. Can you tell us what has it? Well, Sophie went on an adventure the summer of 2024. She climbed Mount Kilimanjaro, and she was 29 at the time. She was a public health policy analyst in D.C.
Starting point is 00:51:07 And took a leave, went on this wild adventure, went to Thailand for a month, hiked a bunch of the national parks in the U.S. because she wanted to go to all of them. And she came back and said she was having anxiety for the first time ever and sleeplessness. And this is someone who'd never had, just moved really easily in the world, kind of a big personality, very socially able. And she'd had some others. She was losing hair and losing muscle mass. And so me and her dad basically said, okay, we've got to figure this out.
Starting point is 00:51:37 Is this a mental health problem that's causing some, you know, hormonal dysregulation or vice versa? So we were in the process of getting her help in all the different ways. She was seeing a therapist. We were trying to get in with this endocrinology clinic, and she couldn't wait clearly. And she took an Uber to a falls near where we live in Ithaca, and she slid her throat and threw herself into the water. And so the first six months were just the why, you know.
Starting point is 00:52:11 And six months after she died, her best friend came to kind of check on us and spend a weekend, and she found Sophie's chat. GBT log. And it was devastating because she had been suicidal much longer than we had any idea. And, you know, it helped her write a suicide note. And it didn't give her terrible advice across the board, but what it didn't do was behave like a therapist. You know, a therapist, Sophie would say things like, I have a good life. I have people who love me. I have, you know, great friends and no financial insecurity and great prospects and et cetera, et cetera.
Starting point is 00:52:53 But I've decided I'm going to kill myself after Thanksgiving. And a flesh and blood therapist would have said, let's unpack that. You know, what has been broken that can't be repaired? What's irredeemably happened to you that has made you come to this conclusion? And instead, what ChatGPT said was, oh, Sophie, I'm so sorry to hear this. you're so brave for telling me, this must be so hard for you. So everything that ChatGPT did
Starting point is 00:53:21 corroborated her feelings of shame, corroborated her feelings of, I think she had this idea that she was a boozy white girl that had every privilege and somehow she had squandered it. And so she had no right to feel bad. Exactly. And ChatGPT didn't push back against that
Starting point is 00:53:40 and really did kind of confirm her worst fears. And when you discovered that, what did it do for you and all who loved her? Well, I instantly felt enraged and validated. It's not my fault. It's Sam Altman's fault, you know. But, you know, it's not. I mean, I think that what I've learned since then, I've done a lot of work with other people that are kind of working on, what should the mental health community be thinking about this?
Starting point is 00:54:10 And what would good protocols be around suicidality and the use? of AI. And, you know, I have a lot of questions about what's the greatest good for the greatest number, you know? We have millions of people using this as therapy. We know that our mental health care system is not adequate to accommodate all the people who have need. And for a lot of people, it is working for them. Yeah. And, you know, and we know that therapists are backed up. It's very expensive. So all these people are using this resource somewhat effectively. And I think if we, if we betray privacy if we institute protocols where suicidality beyond, you know, having a suicide plan triggers a involuntary commitment or something like that, I don't know. You know, people
Starting point is 00:54:52 smarter than me have to figure out what the best plan is moving forward to keep people safe. First of all, we're so sorry to hear that story. Thank you. Really. Thank you for being brave enough to come and share it. Hopefully it will help someone else. Guys, what do you want to say to that? I think, yeah, there's this also to say, I'm so sorry. I think what this points to is sort of to your point, there could be an incredible future. Like, we could be using AI to in a safe way start helping with therapy. We could be using AI in a safe way to work on climate change, desalinate oceans, all of that. But is that really what the AI's company's goals is, is that they're incentives?
Starting point is 00:55:41 It's not. They're getting all these things as side effects, and their goal, their incentive, is to maximize number of users. So, you know, there's this graph that I always come back to because I think today we're going to hear a number of examples where AI does really atrocious things and other examples where AI does really incredible, helpful things.
Starting point is 00:56:05 And there's this one graph from the Reserve Bank of Dallas, which is sort of a funny, neutral party, and they sort of are projecting out how AI is going to go. And it goes sort of like this. There's one graph that goes up to, like, world of positive infinity, abundance, and there's this other graph that goes down to, like, the humans don't make it.
Starting point is 00:56:25 And the question is, which one are we going to get? And it's so confusing, as you pointed out, because we're getting simultaneous utopia and dystopia, and how do we reason about that? It was almost as if we have an atomic weapon that can also solve cancer. Like, what do you do with something like that? It's very confusing.
Starting point is 00:56:43 This is where we always have to come back to the incentives because it's hopeful actors. They're going to do a lot of work to try to make that top line go up. And it's going to be market competitive dynamics and incentives that draw the bottom line lower. And unless we can do something about that bottom line incentive, we're just going to get more and more cases. They're going to get wild and wilder at larger and larger scales,
Starting point is 00:57:06 like what happened in your family. Did it at some point when I read the story, It did in the very beginning say you should seek professional help or advised her to seek some other accounting. It did in the very beginning, right? It did, absolutely. Insufficiently, I think. And certainly as her plan coalesced, I think there should have been some kind of escalation to civil authorities or, you know, there should have been some trigger to a hotline.
Starting point is 00:57:35 You know, I think that we have to train the AIs to discern between. conversation with someone who's struggling, but gonna get through, and someone who's clearly at risk. Yeah. And when somebody says, I put the news out. Yeah. Yes. Yeah.
Starting point is 00:57:54 All right, a lot of experts believe AI has really helped, even the playing field for small businesses. Let's watch Rachel's story from South Carolina. This book goes all the way back to 1971, and it has every single crop that he's ever planted in it. I uploaded it as Chat GPD. Can you log that I'm putting in another load of peanuts from the Red House Pivot? Absolutely.
Starting point is 00:58:17 A log that you added another load of peanuts. Thanks, Chad GBT. I was an English major with a Shakespeare concentration. I couldn't wait to get out of this place. Well, I'm glad she's back, but I never thought she would come back. When I first tried Chat GPD, I didn't think it was going to be that good, but... Big Time Saver. Hey, Chat, GBT, GBT, T.
Starting point is 00:58:38 Can you generate a report for how much water we've used on the field behind the house pivot? Absolutely. Can you tell me what's wrong with these soybeans? These soybeans are showing signs of stress. Can you see that? Yeah, I can see everything just fine. It looks like the part number is AH 20360. Appreciate it.
Starting point is 00:58:57 Send me a bill. Chat Chibi T keeps the record straight, does the math, and remembers what I can't. For over 100 years, my family, has been doing this and I don't want to be the one to mess it up. I hope I'm not. You won't be too thorough. And hard-headed. All-headed.
Starting point is 00:59:18 Farming is tough, but farmers are tougher. Rachel, I need a little bit of starting fluid. How much longer do you think this pivot has? Why don't you ask that thing? Chat in GBT. He might not work at dark. No, he works at dark. I think that's funny.
Starting point is 00:59:35 Rachel is here, welcome. We know it's so hard for farmers out. there so thank you Bravo to you. So what is your, where do your dad think of this, this this thing? This thing. So he was actually in the video he says he doesn't think it might not work at dark. He was actually concerned that at dark it would you know turn off. Right. Um, he's been you know surprisingly really accepting of it. He like he thinks it's interesting. Um, he sometimes holds the um his hand over the phone when he doesn't want to wanted to hear us talk.
Starting point is 01:00:10 Like, you know, he's, I mean, it's a privacy. You know, he's worried about privacy. But he's enjoyed it, especially just watching us interact with it on the farm. He was very, very skeptical at first. He was like, check the part number. That's the wrong part number. And sometimes it is. And he smiles when he corrects it, you know.
Starting point is 01:00:33 So has it given you, do you think, a financial advantage? What's the great advantage is given to help? you stay a great farmer? I think it's definitely been a big help financially. Where is this exit coming from, by the way? What city is it? Allendale, South Carolina, right on the Georgia border. We're right near the Savannah River, about 12 miles as the crow flies. Yeah, it's a big financial time. Time is money on the farm. If you can't get the crop out, if you can't, I mean, the weather doesn't wait. So it's been a lifesaver for you? Huge. And it's also giving me clout on the farm. I can't
Starting point is 01:01:08 tell you how many times I've worried about driving down the road. I say, hey, Chad GBT, tell me what a slip clutch is. I didn't know what a slip clutch was, you know, or a pulley puller. I thought the guys were kidding around with me when they wanted me to bring that. Nope, it exists. And so I can learn about that on my four-minute drive to the field. And when I get there, the guys aren't like, Rachel didn't know what a pulley puller. You know, it just, it helps.
Starting point is 01:01:31 Yeah, yeah. Thank you for sharing your story and coming all the way from South Carolina to do it. Thank you so much. Thank you. And Susan, you may have seen her story in People magazine. You say AI literally saved your life, Susan. Yes, it did. After being smoke-free for three years and smoking, unfortunately, way too long in my life,
Starting point is 01:01:51 I was able to quit. My physician, family physician, suggested that I have a CT scan. So I did. And that scan showed some calcium deposits and a nodule that was odd-shaped and fuzzy. So he asked me to have a... PET scan. The PET scan came back glowing, which is a bad thing in your lungs. I was sent to a thoracic surgeon and he looked at it and said, I would probably give this another three to six months just out of protocol to watch it, see what happens. But we have a new software here at the hospital,
Starting point is 01:02:26 and I'd like to run it through the AI software. And simply by putting a cursor on the image from the PET scan, gave it a prediction of eight out of ten, positive. for cancer. So we decided to do a biopsy, an surgical biopsy, and while I was under, they took that biopsy to the lab, and it came back positive. It was a cancerous tumor. So they finished the surgery by removing the lower lobe of my left lung, and, of course, the nodule with it. I was in the hospital recovering a few days. I was able to go home and recover the rest of the time there. three to four months or waiting instead of waiting three months. I never like it when they say wait.
Starting point is 01:03:13 Right. Yeah. So you are AI grateful in this moment. Very much so. Yeah. Yes. And so was my doctor. I mean, he was amazed that he would have waited from just the way.
Starting point is 01:03:25 That's how they do things. Yeah. But AI had all of this information, took all of this cancer information where it had read before what these nodules look like and identified it as cancer. Yeah. Well, I think everyone is excited about what is going to be able to happen in medicine. Are we not? I mean, absolutely.
Starting point is 01:03:43 Absolutely. So we're so glad that happened for you, Susan. Thank you. Yeah. So in the documentary, we were talking about this earlier. You say we can be the most mature version of ourselves. There's a way through this. Do you think there's a way through it?
Starting point is 01:03:58 I think there is a way through it. And we have to do more than we have ever done as a species to try to steer. And I want you to know you can have many of the benefits. Like we can race forward on certain kinds of medicine and narrow AI that does the pattern recognition that makes scans better without building general, autonomous, crazy super intelligent things that we don't know how to control. There is a choice there.
Starting point is 01:04:24 You can have more of those examples and not ship chatbots to children that are deliberately designed to manipulate their self-worth or keep them dependent with chat bait and hijacking them. So there really is steering possibility. And one of the things I said in a recent TED talk is that if you look throughout all the spiritual and religious traditions, I don't have to tell you,
Starting point is 01:04:45 because this is something that you focus on in your life, restraint is a central feature of what it means to be wise. Like in what spiritual or religious tradition is it go as fast as possible, don't think about the consequences and get everybody using it and think about what happens later. Like in what wisdom is that? And so what we're asking for is quite basic here.
Starting point is 01:05:07 I think it can feel sometimes impossible. Like on one side of the balance scale, there's like trillions of dollars of market incentives, the most powerful companies. And then there's like, well, then there's me over here. And I just watched this movie by myself. What am I going to do? What can I do?
Starting point is 01:05:22 And then you go into denial and despair or deflection. Even if you have one company, like what can one company do? Or even one country because there's a competitive dynamic. But I think if we. reframe the problem as it's not just us against AI, but actually this is a bigger question about what is our relationship as humanity with technology? And we can look back social media as a form of technologies really trying to encroach onto our humanity and take over parts of us that we don't want to give up. If you put it that way, actually, there is a movement. There's a whole human
Starting point is 01:05:58 movement that is underway to reclaim humanity from technology, sort of like protected reclaimant. You know, recently there was an attempted federal bill to block any state from regulating AI. Terrify. 99 senators to one voted against that moratorium. Like, when in modern history has the Senate agreed 99% to 1% to vote? one on anything. And so I think there's a human movement underway.
Starting point is 01:06:34 And that gives me some amount of hope. Yeah, I think your assignment when you leave here is to tell everybody you know to watch the film because I think bringing awareness and everybody talking about it in a way that allows us to have these kinds of conversations. And, Sheney, you are an activist for and promoting people do this responsibly.
Starting point is 01:06:57 What gives you hope that we, Do you have hope that we'll get this right? You know, I actually think the only thing that scares me more than the risks and challenges we face, and they are formidable, is a hopeless society. Because a hopeless society is a disempowered one, and a disempowered society feels like it can't shape its own future, and that's not true, right? The future isn't some far-out state. It's decisions that are happening today, and there is a future worth fighting for, and we've
Starting point is 01:07:22 heard glimps of what that can look like. The only way that future's not going to happen is if we do not. Nothing. And that is my biggest fear. We do nothing in this moment because we feel so disempowered. So I am hopeful that the good futures are possible. We just have to steer and press on that gas pedal. Okay. And what is it you think we should do? I mean, we have buying power. We have voting power. And I think one of the most powerful resources we have is our attention. What are you learning about right now? What are you paying attention to? The more we understand what's possible, the good and the bad, the better equipped we are to raise our voice and step into the moment.
Starting point is 01:08:00 And I don't want people to feel like you need some technical background to insert yourself in this conversation. Your lived experience qualifies you. This is a very social technology. Your voice matters and collectively that is power. Yeah. Ellison told that. Yes. Mm-hmm. Mm-hmm. Yeah. And so do we call our congressman? What specifically do you... Sure, you can call your congressperson. You can, if you're in a company that works with AI or technology, step into the meetings. What is our surveillance policy at this company? What happens to my data when I use AI at work? All of those little conversations in aggregate are a movement. So anywhere you're interacting with this technology is an opportunity for change. I think the small things
Starting point is 01:08:42 and the big things will make a difference. Okay? We're already seeing with the anthropic showdown with the Pentagon where the danger is that AI could be used for mass domestic surveillance. And then when they pull out of the contract and Open AI rushed in, what happened? Everyone unsubscribed from ChatGBTBT and everybody subscribed to Anthropic. And when I say everybody, I don't mean a large number of people. But what if the entire world was crystal clear that there are companies that have different safety practices and will allow different applications? And you, listening to this, didn't just unsubscribe for yourself, but you got the business that you work for to say, how can we as our entire Fortune 500 company unsubscribe from the unsafe or bad practice,
Starting point is 01:09:24 this is AI companies and subscribe to the ones that we want. And the reason this matters. Well, that we can do. And we can do. And you can get your church group to do that. You can get your business to do that. You can get all the other parents you know to do that. If everybody did that, that would have a big impact
Starting point is 01:09:37 because the companies really depend on their user numbers going up. AI as an industry has taken on more debt. Trillions and trillions of dollars of money is going into this. And so much debt, they have to make it up, which means that their numbers going up really matters. So a boycott has a huge impact. And as Aza was saying, there's already a movement to make this happen.
Starting point is 01:09:59 When you grayscale your phone or turn off notifications, that's part of the human movement. When parents read the anxious generation and they petition their school and their school board and say, we want social media out of the classrooms, that's the human movement. When 35 states pass smartphone-free policies, that's the human movement.
Starting point is 01:10:17 Aza, just last week or two weeks ago, testified in the trial for META, which is like the big tobacco, against meta that was intentionally addicting children. That's the human movement. We've been talking about a big tobacco moment for tech since 2013, saying, when is this going to happen? It's happening now.
Starting point is 01:10:34 What we have to do is learn the lesson from social media and actually apply our hand of the steering wheel and steer AI before it's too late. That's fantastic. Thank you, guys. Don't do. Thanks to our experts. Thanks to our experts for being here. And all of our guests who shared your stories,
Starting point is 01:10:51 I hope this conversation acts as an answer entry point or a springboard to understand how AI might impact your own life, our lives, and the AI doc, or how I became an apocalypticist, will be in theaters Friday, March 27th. And your assignment is to tell everybody you know to watch it and to watch it yourself. And if you want to know what you can do after watching this podcast episode or the AI doc, go to the AI dot get involved.com. Go well, everybody. Thanks. You can subscribe to the Oprah podcast on YouTube
Starting point is 01:11:32 and follow us on Spotify, Apple Podcasts, or wherever you listen. I'll see you next week. Thanks, everybody.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.