Pivot - AI Basics: Why AI Is Not a Job Killer

Episode Date: October 30, 2024

Kara and Scott are back with part two of Pivot's special series on the basics of artificial intelligence. How is AI revolutionizing the workplace? And what are the skills workers need to learn to stay... ahead? Andrew McAfee, MIT research scientist and co-founder of Workhelix, explains why he doesn't think AI is a job killer, and shares the advice he gives to business leaders about integrating AI into workflows. Follow us on Instagram and Threads at @pivotpodcastofficial. Follow us on TikTok at @pivotpodcast. Send us your questions by calling us at 855-51-PIVOT, or at nymag.com/pivot. Learn more about your ad choices. Visit podcastchoices.com/adchoices

Transcript
Discussion (0)
Starting point is 00:00:00 Support for this episode comes from AWS. AWS Generative AI gives you the tools to power your business forward with the security and speed of the world's most experienced cloud. Hi, everyone. This is Pivot from New York Magazine and the Vox Media Podcast Network. I'm Kara Swisher. And I'm Scott Galloway. And you're listening to our special series on AI. AI has changed our personal and professional lives in many ways, with plenty of more changes to come. Here to talk with us about how AI is impacting work and what people need to do to stay ahead is Andrew McAfee. Andrew is a research scientist at MIT and author of The
Starting point is 00:00:41 Geek Way. Welcome, Andrew. Hi, Cara. Hi, Scott. Thanks for having me. So you've been studying AI for over 10 years. You've also co-founded WorkHelex, a startup that helps companies figure out where to deploy generative AI. Scott and I talk about it a lot. Scott says if you're not using AI, you're like a person who says they're not using computers and you're fired.
Starting point is 00:01:00 Can you talk about how AI has changed the workplace and where people are with it right now? I'm completely with Scott, and I realize this makes for a boring podcast if he and I just agree with each other all the time. But it's pretty clear that AI and specifically generative AI are already changing the world of work. And there are lots of different ways to think about that. Here is my favorite. What we have as of the dawn of AI is an army of clerks, colleagues, and coaches who are available on demand to just about any professional. So a clerk, for example, is somebody who will listen
Starting point is 00:01:32 to the meeting and transcribe the notes and give you the highlights of it or transcribe a patient meeting. That's kind of low-level grunt work. And we now have an armada of technologies that are good at it. A colleague is somebody who will translate a foreign language document for you or help you tighten up a thing that you've written by 20% or change the writing level of it. Or for me, because I'm a terrible artist, I've now got a colleague called Midjourney who figures out the artwork for me. And a coach, I'll give you a silly example. I love crossword puzzles. And there are a bunch of words that show up over and over again in crossword puzzles, but have silly definitions. And I use AI to quiz me on those,
Starting point is 00:02:11 so I stay sharp on them. A much better example is something that my co-founder, Eric Brynjolfsson, discovered when he did research. It turns out that when they deployed generative AI in a call center, performance went up by about 15% very, very quickly. That's great.
Starting point is 00:02:26 The cooler part is that the biggest performance gains were found among the newest and the least skilled employees. So what was going on was that generative AI was kind of whispering in their ear, prompting them to say what they should say in real time to a customer on a chat. I think all three of us have had careers in tech long enough that we've heard about knowledge management over and over and over again,
Starting point is 00:02:49 and it's been a bitter disappointment overall. I think we finally have incredibly powerful knowledge management technologies, so just about any of us can have a coach now. So I think of clerks, colleagues, and coaches available on demand to all of us, I get pretty excited. All right, Scott, since this guy agrees with you violently, it seems. Yeah, I like the terms used more. I would say it's having a thought partner shadowing you. But so there's, well, I'll put forward a thesis, a two-part thesis on what you'd respond to it.
Starting point is 00:03:17 One, the catastrophizing by people after they've vested their options, I just find really obnoxious and overdone, quite frankly. I don't believe it's going to go sentient and kill us. I mean, and this comment might age. Very poorly. Yeah. If it ages poorly, we have bigger problems, right? Amen.
Starting point is 00:03:37 One, do you believe that this imposes the kind of threat that a lot of, and the thing I can't stand about is, now that I have my $10 million in Google stock, I'm going to talk about how worried I am about it, that I created this monster. Anyway, one, do you think the catastrophizing is warranted, overdone, underdone? And two, given that every technology has externalities, what are you most worried about? Can I just add some facts around this? This is a job killer question too. And a 2023 Goldman Sachs study estimates 300 million jobs could be lost or diminished by AI,
Starting point is 00:04:10 though the cost saving from the technology companies will ultimately help increase the annual global GDP by 7%, according to the same study. Elon Musk recently predicted AI will get humans to the point where no job is needed, which is probably overstating it, but go ahead. Yeah. And the robo-tax, it, which is probably overstating it, but go ahead. Yeah, and the robot taxi was here in 2019. Yeah. I'm sorry, go ahead. Yep. So there are about three different kinds of risk
Starting point is 00:04:31 that I keep hearing about. And again, I'm sorry to be boring in advance, Scott. I'm again going to violently agree from you. There's the existential risk. This is the super intelligence killer AI. Skynet. Skynet thing. I think it's put out by people who A, have vested ha-ha options,
Starting point is 00:04:48 and or people who read way, way too much sci-fi when they were sophomores and didn't really get past it. I always fall back on what Andrew Ng says about this question. He's one of the founders of modern machine learning. And I love his response. He was at Google. He was a bunch of people. He was at Google. He's founded a bunch of companies. He's an investor in Work Helix. And to me, he's nailed this issue. He said worrying about Skynet and superintelligence is like worrying about overpopulation on Mars.
Starting point is 00:05:15 I guess it's theoretically possible. Boy, do we have more important things to worry about. Now, there are other real harms that are going to come along with any technology as powerful as I. We know it's going to be used for misinformation and disinformation. We know the bad actors are going to grab this technology and do things with it. This is a real problem.
Starting point is 00:05:33 We need to deal with it. But in a sense, it's not new. All powerful technologies, especially ones that are pretty democratic, are going to be used by the bad actors. Kara, you brought up the economic risks, and in particular, the risk about massive technological unemployment. We have been worrying about this since the Luddites were smashing looms 200 years ago. Every powerful technology brings along another wave of-
Starting point is 00:05:55 Well, they did have a point, to be fair, but go ahead, sorry, I'm teasing. Well, they did, but let's get to what the point is. Every powerful technology is part of the cycle of creative destruction. And with 200 years of history, so far, the pattern is incredibly clear. The creation outweighs the destruction, and that has to do with jobs as well. The problem we face in America right now, and in most of the rich world, is absolutely not technological unemployment.
Starting point is 00:06:23 We have chronically low levels of unemployment. We don't have people at every levels of skill and education to do the work that needs to get done. The idea that that's going to change in five or 10 years because of a technology, even a technology as powerful as AI, just doesn't make any sense to me. Now, Cara, to your point, the Luddites did have a point. The skilled weavers did see their employment decline. We used to point to bank tellers as an example of how even when there's a new technology called the ATM, demand for bank tellers still goes up. That was true until about 10 years ago, and now bank teller unemployment has been declining.
Starting point is 00:06:59 So we can see technology displace workers in specific jobs. We can see technology displace or reduce employment in overall industries like agriculture and manufacturing, even though output from those industries continues to go up. But the creation part of creative destruction is really, really easy to underestimate. And what's been happening is that the innovators and the entrepreneurs have kept on coming up with new things that need to be done that can only be done by people. So in that vein, what are some of the most important AI skills and tools that employees need to be learning right now to stay relevant and related? What are the best and worst cases
Starting point is 00:07:32 you've seen of AI in various professions? In the short to medium term, the skills are keep doing your job, but realize that you have this very powerful clerk, colleague, coach alongside you that is going to take over sometimes really big chunks of your job, but realize that you have this very powerful clerk, colleague, coach alongside you that is going to take over sometimes really big chunks of your job. Translator is a great example. And I was shocked to learn earlier this year that as far as we can tell, the number of translators employed in the United States continues to go up. Now, it's not millions and millions of people, but that job is still increasing in employment. And that's weird to me until I thought about it a bit more deeply. And I realized that translating the document is actually only one of the things a translator
Starting point is 00:08:12 does. And now they don't start with a blank sheet of paper. They start with a machine-generated translation. But that translation is going to have mistakes. It's not going to be very good at some kinds of idioms. It's going to be imperfect. So the proofreading, the copy reading becomes really, really important. And the importance of negotiating with the client, making sure you've got the gig right, making sure you deliver a clean product that's free of flaws,
Starting point is 00:08:35 those skills become more valuable and translators see demand for their services go up. That's the pattern we're going to see a lot of. So instead of trying to predict any massive changes in how a particular job is going to be done, the one thing I can say with huge confidence is that technology is available to help you with any knowledge work job that you have right now. And the sooner you start exploring the toolkit and putting it to work and learning, the better off you'll be. So this is what I do advise, just use it and you'll see. Best and worst cases, you seem very briefly. I'll give you the absolute worst case because it's me. When I was finishing my last book, I deployed generative AI as a clerk to help me do the
Starting point is 00:09:14 citations for the book, you know, the messy end note stuff in the back. And it engaged in what's called hallucitation. It made stuff up. This is a terrible, terrible use case, and I should have known a lot better for it. But at WorkHelix, we've looked around and we found that if you add additional technologies on to generative AI, something close to 50% of the jobs in the economy can see a huge portion of their tasks substantially improved with no quality loss via AI. So the mantra that we have inside the company
Starting point is 00:09:46 is plan a little and iterate a lot to figure out how this is going to change your work. Okay, let's take a quick break. When we come back, we're going to talk about how companies can integrate AI into their workflows, and we'll ask Andrew for some predictions. Support for this episode comes from AWS. With the power of AWS Generative AI, teams can get relevant, fast answers to pressing questions and use data to drive real results. Power your business and generate real impact with the most experienced cloud.
Starting point is 00:10:18 Support for this episode comes from AWS. AWS Generative AI gives you the tools to power your business forward with the security and speed of the world's most experienced cloud. Scott, we're back with our special series on AI. We're talking to MIT research scientist
Starting point is 00:10:40 Andrew McAfee. Scott, what would you ask him? Yeah, so Andrew, one threat or one externality I don't think gets enough attention is that I worry that a lot of people are struggling with loneliness and AI will only create, replace more and more relationships and decrease the motivation to go out and make friends, pursue romantic relationships because these AI bots, the search volume for AI girlfriend has skyrocketed. Have you thought at all about the role AI might play in loneliness, specifically amongst young men who are kind of increasingly sequestering from society? Yeah. And you bring up this loneliness epidemic, which I'm hearing a lot about.
Starting point is 00:11:29 My read of the evidence is a little bit more ambiguous, but I'm not saying that there is no loneliness epidemic. There clearly are people who are dropping out of society and technology, especially modern technologies, are really good if they want to do that at helping them do that. It feels to me like it's a bit related to the anxiety epidemic among adolescents, which I do think is a real thing. And social media is a part of it. So again, this bundle of extraordinarily powerful technologies that we have is going to bring some harms and some downsides along with it. And I think you've nailed probably an important one. What I believe is that we need societal community policy kinds of responses to the harms that
Starting point is 00:12:10 technology brings up, as opposed to trying to make sure that technology can never, ever do that harm. So I would not want to call a halt or put guardrails on conversational AI just because I'm worried about a loneliness epidemic. In other words, I would try to think of other things that we can do to get people out of the room, get them to engage with each other in this pretty atomized world that we live in. And Kara, you bring up the workplace
Starting point is 00:12:32 and there's a massive controversy. Because if you're getting rid of coaches in it, these are all people you interact with. Amen. And I think this is a real challenge. Like I say, we have very, very powerful knowledge management technologies now. However, we humans learn from each powerful knowledge management technologies now. However,
Starting point is 00:12:50 we humans learn from each other. We are the most social species on the planet. And I think we need physical proximity at least sometime to pick up and accelerate that learning. So along with the return to work wave that's happening now, we also have to figure out the how do we use AI and how do we coach and advance our skills with AI. So it's interesting. I just interviewed Seth Meyers, and he said the things he misses is around the table, stuff with young comics. They're online. They don't need to go to Chicago and meet a group. It was really interesting because it had the same implications. But if you're advising a CEO or business leader about integrating AI into their company right now, what would you tell them? And given these capabilities are growing so quickly, how can you stay ahead? You feel like
Starting point is 00:13:28 you're perpetually, and it probably does remind me of early internet days, like use this browser, use this, use that. A lot of them were bullshit, but I tried them all. It was an enormous waste of time, but it wasn't in a weird way. Yeah, it wasn't, it wasn't, right? And what you, Kara, what I think you and I learned back in 1994 or whatever was how to surf the web and the value of surfing the web. So the fact that, you know, Firefox might not have won out or what was the Norwegian called? There were so many. There were so many. I forgot. Yeah, there were so many that the particular one you were using didn't pan out, was not that relevant, it turns out. They weren't crazy expensive. And the broad skill about how to use the web effectively, the quicker you started with that, the better off you
Starting point is 00:14:09 were. And in particular, when a technology this powerful comes along, it's called a general purpose technology on a par with the steam engine, I think, sitting on the sidelines and waiting for the environment to calm down, I believe is a terrible, terrible idea. Because instead of the first mover or the fast follower, I think the race here goes to the fast learner. And the environment to calm down, I believe is a terrible, terrible idea. Because instead of the first mover or the fast follower, I think the race here goes to the fast learner. And the way to learn is to do a little bit of planning and a whole lot of iterating. Scott?
Starting point is 00:14:33 I worry, Andrew, that there's already, there just is time to mass adoption. The cycle time is decreasing across products that are truly breakthrough, including AI. That you also see the emergence of a dangerous concentration of power. And I see a new kind of Wintel-like duopoly forming, and that is what I affectionately refer to as OpenVIDIA, that OpenAI and NVIDIA with 70% and 93% share, respectively. Do we have a new duopoly that we should be thinking about breaking up
Starting point is 00:15:04 already, or at least keep our eyes on it? Oh, great. We get to disagree. So, my answer to your great question is yes and no. We do have a new duopoly. I don't think we need to be getting out the antitrust hammer and whacking on OpenAI and NVIDIA, because, Scott, your career like mine is long enough to remember when we were terribly worried about IBM's dominance and then terribly worried about Microsoft's dominance. And your duopoly did not include Alphabet slash Google, which is now the subject of antitrust scrutiny. So the pattern that I think all three of us have seen over and over in high tech
Starting point is 00:15:36 is what I would call dominance and then disruption. There are gigantic, really powerful tech companies and they seem unassailable. And then they get a sale from some quarter that we weren't expecting. It's bizarre to me, Scott, that the duopoly you bring up features a startup, features how old is OpenAI, founded out of a weird nonprofit, and now it's the one that we're worried about. I do not have a crystal ball. If and when the three of us come together in five years, let alone 10, we're going to be worried about another duopoly or crazy concentration of power, not the one we have today.
Starting point is 00:16:09 And that's not because the Justice Department or the FTC broke things up with the hammer. It's because tech is a very, very hard industry to stay on top of. Yeah, that's absolutely true, although it is the same people. You know, now I'm remembering a thing I forgot, the Pointcast Network. Remember them? Oh, man. This is a walk-down memory. Push technology. ICQ? Yeah, we were doing all that, man. ICQ. Yes, you've already. He's smart. He hit the bed. Oh, there's so many. Oh, my God. There's so many. And there will be in this, too. So, last question. Anthropic CEO Dario Amodi recently wrote an essay predicting AI's utopian future. He's one of the safety people. So that was interesting, this world where AI has achieved superintelligence and where the progress is achieved at a faster pace.
Starting point is 00:16:54 Underhyped, overhyped, I guess that's kind of too broad a question. And can you give us a prediction for how you think we'll be using AI in five years and 10 years? So that's a two-part question. Overhyped, underhyped, what did you think of the Amodi essay? I think hyping the superintelligence, singularity, Skynet stuff is really, really overdone and premature and actually harmful to the industry.
Starting point is 00:17:16 I heard a guy at one of these superintelligence conferences a while back say to the assembled people, you all need to fire your PR department. This is attracting scrutiny from regulators, which only incumbents want, and it's turning the public against these extraordinarily powerful technologies that we need to cure disease, get through the energy transition, and do our 21st century homework. So I agree with Dario that we are creating crazy powerful
Starting point is 00:17:44 technologies. We're not done yet. But my prediction for the next five or 10 years is, first of all, we'll be talking about different, very powerful technology companies on top. And second of all, that the companies that get off the sidelines, that plan a little and iterate a lot, we're going to see the performance gaps that already exist and are real across the economy. We're going to see the performance gaps that already exist and are real across the economy. We're going to see those gaps get bigger. And I'm actually really worried about the
Starting point is 00:18:09 incumbents of the 20th century, the really successful companies that got built up over the course of the 20th century, because as far as I've seen, they're getting their butts kicked by a bunch of kids from Silicon Valley. And I want the incumbents to make it a less lopsided fight. That's what I'm keeping my eye on for the next five or 10 years. And the incumbents to make it a less lopsided fight. That's what I'm keeping my eye on for the next five or 10 years. And the incumbents being? Think of the incumbents in the auto industry. Think of the incumbents in the space industry. Think of the defense industry. Think of entertainment. Hollywood was a very, very stable place for a century. And then a DVD by mail company comes along and stands it on its
Starting point is 00:18:44 head in less than 15 years. This is bizarre, right? We don't see this very often in businesses. Fine, they deserved it, but no one was going to do it for them until the DVD by mail company comes along. Yeah, they deserved every bit of it. Sorry, I was there. No, we're better served today. That's great. Right. Very last question, Scott. If you're a kid coming out of college, or maybe not even college, but you have some aptitude for Right. Very last question, Scott. at 22. I just knew I wanted to be successful at some point. And I was trying to build a base. What is that in your view, if you were coaching someone who really was open to almost anything in terms of where to live, what industry to be in, what skills to acquire, what advice would you have for them? I love this question because my answer to it is both vague and adamant. The adamant part is get yourself to a concentration of alpha geeks, apprentice yourself to them,
Starting point is 00:19:45 and learn to build stuff and make stuff and think about business in the world the way that they do. So, you know, go knock on the door of the people pushing the envelope the hardest in the area that you think you're interested in, and just go put yourself by their side
Starting point is 00:20:00 and learn how they're doing stuff. This is different than go work for one of the great advisory companies of the 20th century, a bank, one of the industrial giants of the 20th century. The future is already here. It's just not evenly distributed. Go learn from the alpha geeks. Define the way that you want to define it. Okay. Thank you, Andrew McAfee. A really interesting time we're going through going forward. It's a really interesting podcast. As we all say, just start using it. We keep saying that to you. Stop not using it. Use it. It's like saying, I don't want to use a telephone. I don't want to use a telegram. I don't want to use television. Just use it. You'll see. You'll see
Starting point is 00:20:39 and figure it out yourself. Okay, Scott, that's it for our AI series. Please read us out. Today's show is produced by Lara Naiman, Zoe Marcus, and Taylor Griffin. Ernie and her Todd engineered this episode. Thanks also to Drew Burrows and Ms. Severio. Nishat Kerouaz, Vox Media's executive producer of audio. Make sure you subscribe to the show wherever you listen to podcasts. Thanks for listening to Pivot from New York Magazine and Vox Media. You can subscribe to the magazine at nymag.com slash pod.
Starting point is 00:21:04 We'll be back next week for another breakdown of all things tech and business. of AWS Generative AI, teams can get relevant, fast answers to pressing questions and use data to drive real results. Power your business and generate real impact with the most experienced cloud.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.