In Good Company with Nicolai Tangen - Ethan Mollick: AI Urgency, Leadership Responsibility, and Why Companies Must Act Now

Episode Date: June 11, 2025

Which companies will lead and which will be left behind as AI transforms the way we work? Nicolai Tangen connects with Ethan Mollick, Wharton professor and author of 'Co-Intelligence: Living and ...Working with AI,' to explore how organizations can harness AI's revolutionary potential. They discuss the growing adoption of AI tools across workforces, proven tactics for driving company-wide implementation, the rise of autonomous AI agents, and why traditional training approaches may be missing the mark. Ethan reveals insights from his research showing that AI works best as a collaborative teammate rather than a replacement. With AI capabilities advancing faster than expected, organizations face increasing urgency to act. Tune in for an insightful conversation! In Good Company is hosted by Nicolai Tangen, CEO of Norges Bank Investment Management. New full episodes every Wednesday, and don't miss our Highlight episodes every Friday.  The production team for this episode includes Isabelle Karlsson and PLAN-B's Niklas Figenschau Johansen, Sebastian Langvik-Hansen and Pål Huuse. Background research was conducted by David Høysæter and Yohanna Akladious. Watch the episode on YouTube: Norges Bank Investment Management - YouTubeWant to learn more about the fund? The fund | Norges Bank Investment Management (nbim.no)Follow Nicolai Tangen on LinkedIn: Nicolai Tangen | LinkedInFollow NBIM on LinkedIn: Norges Bank Investment Management: Administrator for bedriftsside | LinkedInFollow NBIM on Instagram: Explore Norges Bank Investment Management on Instagram Hosted on Acast. See acast.com/privacy for more information.

Transcript
Discussion (0)
Starting point is 00:00:00 Hi everybody, Nicola Tangen from the Norwegian Sovereign Wealth Fund. And today I am here with Ethan Molyk, one of my favorite professors, a professor at Wharton and who was out not long ago with a book called Co-Intelligence, Living and Working with AI. And actually you can see it behind Ethan there, down to the right. If you haven't got it, run and buy it. Ethan, if you were a chief AI officer in a company for the next three months, what kind of top actions would you take straight away?
Starting point is 00:00:44 So I think that the most important thing is to get people actually aware of where the state of the art in AI is. I talk to companies all the time, and I think that a lot of executive level people may have tried AI a while ago or didn't use it personally and don't realize how potentially transformative it is. And I think that I have a sort of general idea that you need to involve your team leadership,
Starting point is 00:01:07 you need to involve a set up a lab that's doing research and you need to think about how to roll this out to the crowd, to everybody in the organization. So you've got to kind of bring the whole company with you, which is not always an easy thing to do. And how do you maximize the uptake in the organization? So I think that's a really good question. And I think we're still figuring out the answers. But it's kind of like any other thing you want to do. You have to think about incentives.
Starting point is 00:01:30 And you have to think about leadership, right? So why are people set device to use this? Now, the thing that makes AI interesting is everybody's already using it. So there's a new study that just came out that showed and represented a sample of American workers that usage went from 30% in February to 40% of people using AI at work,
Starting point is 00:01:46 a little over 40% as of April. So it's used everywhere. The thing is people aren't just showing you they're using it because they're hiding it, because they're not incentivized to show you. They're worried they'll be fired if they use it either because they'll be told that they're using it improperly or because people realize they're not necessary or there's less workers needed. And so there's less workers needed. And so there's incentive problems.
Starting point is 00:02:07 And then there's also, for everyone else who isn't using it, there's role modeling problems. What do I do to get started? How do I use this? Why is it important? So you've got to both deal with the skeptics and the people already adopting it who may not be showing you they're using it yet. Why would people not want to show that they use it? I mean, here, if people don't use it, I'm like,
Starting point is 00:02:33 you know, that's bad, right? You know, so if you think about the incentives of the average worker who's using AI, first of all, they look like geniuses when they're using AI. And one problem that we see is people don't want to show you they're using it because it makes them look like they're not geniuses, like people get the AI credit. The second reason is IT in many companies is viewed as a thing that is a cost cutting measure. So if I could show you that the AI does some of my work, do I get a reward for that or do we fire employees? Do I feel safe to reveal it? And a lot of people are just using it to work less. So you're working 90% less. Am I going to show you I'm using AI so I have
Starting point is 00:03:08 to do more work as a result? Am I going to get a credit for doing this? So there's a lot of incentives to not share, probably a lot more than to share. Wow. How do you measure AI usage in a company? So the metric thing is really interesting, right? So what I see a lot of people drive towards, we're in a world of sort of new innovation, right? So it can be a little challenging to have a direct, like easy answer to that question of like, what does AI do?
Starting point is 00:03:34 How should we be using it immediately? So chief AI officer, you're doing experimentation. That said, I think raw usage actually matters. I mean, the typical KPIs you see are how many people are using our AI app internally. The downside of that is in most companies I talk to, that maxes out at 20% or 30% of the population. And that's because an additional third
Starting point is 00:03:53 is probably secretly using AI or doesn't want to use your apps, your apps aren't good enough. And another third is kind of waiting for instructions. Like they're happy to use it, but they're the opposite of the people who were early adopters. They want clearer rules about how I should use it and why.
Starting point is 00:04:07 So I think you can measure adoption by use of your app. I think you can measure adoption by other internal measures of how much you're using AI at work. But I think you have to be aware that there's still this secret cyborg problem of people not revealing their AI use. So how are the best companies going about this? Seeing some really interesting examples of how this works, I can tell a few of the stories I can tell all of them.
Starting point is 00:04:31 One example is radically changing incentives. I've seen companies that offer $10,000 bonuses at the end of every week to every employee that uses, the employee who best uses AI to automate their job. They think they're saving money versus other approaches. I've seen people build this into their hiring process. So before you hire somebody, you have to, your team has to try and use AI to do their job. And then you adjust your request for hiring
Starting point is 00:04:55 based on that experience. Or before you request money, you need to show how you're using AI to do it. Moderna has this really great example that they put together. They use AI for everything. What they did was build around the process of annual reviews. Basically, they built a whole series of GPTs that help people uncover their own performances,
Starting point is 00:05:17 improvement needs, and what they've done over the year, and talk to the right people about their jobs and things so they can write a really good yearly update about themselves. And they said, well, if you don't use these GPTs, you're probably not gonna do as well on your performance reviews and that will hurt your annual salary. And everybody ended up using this series of things which introduced them to AI.
Starting point is 00:05:35 So putting these bottlenecks in place where people have to use it, thinking about building into internal processes in a way that encourages positive use rather than negative use. Those tend to be really effective methods. So you need a combination of the stick and the character. I think you do and I think you also need role modeling, right? A leader who uses AI will make sure AI seems critical. Someone who doesn't use AI and says use it is kind of a problem. I mean,
Starting point is 00:06:01 you mentioned the chief AI officer at the beginning. One of the things that worries me a little bit about that is there is no ability to have a chief AI officer really at this point. I mean, generative AI has been around three years. Everybody who's on their very first project. So if you hire chief AI officer, that is often somebody who is actually doing machine learning beforehand, which is great.
Starting point is 00:06:19 Like AI writ large, but not generative AI. But they're not gonna have any idea of how to make an organization transform with AI. It has to be at the leadership level, I think. It has to be at the C level of the company. But does it make sense to have a central unit which kind of disperses the best use cases and make sure that everybody is at it?
Starting point is 00:06:37 So that's the concept of the lab that I was talking about earlier. So you do want a centralized unit for experimentation, but that lab has to be staffed partially by the crowd, by the people who are the best users inside your company. And they're going to be sort of ambidextrous. They're going to be shipping out like, here's a great use case, and also building a philagetic system that doesn't quite work yet, that automates the entire job, right? But I think the traditional view of having IT produce things, I think
Starting point is 00:07:04 that actually is on its way out. Some of the most interesting experiments I'm seeing in companies are dispersing engineers from the IT department out into work with subject matter experts. Because Vibe coding has strong limitations, but it's absolute best when you have a senior coder working with a subject matter expert. You can do incredible things in a couple of days that used to be whole processes. So I think there's gonna be
Starting point is 00:07:28 centralization of some AI functions, but it has to disperse other parts of the organization in a more decentralized way. Yeah, yeah. We do that. But what about ambassadors and people across the firm being trained regularly? Yeah, and I think one of the things we've been struggling with a lot as we think about AI use and observant companies has been what training means. So I think one of the things I know is the approach that you take that I think is smartest about training is using the opportunity for contact hours with AI. I think one of the smartest things I said in the book, I didn't know it was the smartest
Starting point is 00:08:01 thing at the time, but it turned out to be quite useful, has been that you just need to use AI to get it, right? You need to bring it to every work task and you'll figure out what it's good or bad at 10 plus hours. And people are very resistant, even very smart, very well-meaning, very self-motivated people are often very resistant to using this for reasons they can't fully explain, right? It's a weird technology. It's not that easy to use.
Starting point is 00:08:24 It's uncomfortable in some ways to sort of confront, I talk about the need for an existential crisis. So I think that it's hard to see, adoption is strange that way, right? So training can kind of get you into adoption, but we don't actually have that much we can teach people about AI, like a lot of the property techniques you learn don't really work that well or aren't that important
Starting point is 00:08:42 as the model gets better. So it's not that experience. And I think you will develop this idea of these champions. Some people will just get it and it's really important, like you said, to get those people out there. They're the cutting edge. They're the representatives of the lab in the world. Will the people who are top performers without AI be the top performers with AI? This is one of the biggest questions we're facing. If you think about it, there's four possibilities for what happens in an AI world on skills.
Starting point is 00:09:21 The first effect that we saw, and we saw this in our study, the Boston Consulting Group study I did with my friends at Harvard and MIT in the University of Warwick, where we found big performance gains, and this came out like a year and a half ago, kind of made a big stir, 40% improvement in quality for people who use GPT-4 versus not, big speed improvements. And a lot of other studies like that have shown a leveling effect. So bottom performers get the biggest boost. When you really look at what's happening,
Starting point is 00:09:45 it's actually the AI doing the work of the bottom performers, right? So the AI is pretty good. So it moves everybody up to the eighth percentile. So one option is it boosts the bottom performers. A second option that could exist simultaneously is the idea that top performers get some sort of massive returns.
Starting point is 00:10:00 We have a couple of studies show that, but not that many. It's hard to study. There's actually a, one of the best pieces of evidence for that was actually turned out to be a fraudulent paper from MIT that didn't exist, right? But I think there's a lot of suspicion that top performers using AI can get a huge boost, just harder to measure.
Starting point is 00:10:17 So there's a possibility that maybe there's a hundred times return if you're already a good coder. And we'll know more about that in the near future. There's also a possibility that, you a possibility that AI lifts everybody up. So everybody's performance goes up by a similar amount. And then there is this sort of other possibility that there's AI whisperers who are just good at AI, and they're the ones who get all the returns.
Starting point is 00:10:36 So we don't know whether it's contrary in the lower end, on the top end, whether it lifts everybody up, or whether there's just sort of magical AI whisperers who are just built to do this. And then agents are coming to replace everybody if you listen to the AI labs. And what do you think? Do you think that will happen? So I, you know, I think... I mean, first of all, just tell us, an agent, what exactly does an agent do? Yeah, so nobody has a good definition of anything in AI.
Starting point is 00:11:03 So people will, if you are a leader in a company, you will have vendors selling you things with every possible label and they'll say it's agentic and everything else because there's no clear definition. But the simplest version, and this sort of overlaps with what the labs think, is imagine any AI system that can be given a goal and can autonomously go and try and accomplish that goal without further human intervention using its own judgment and tools. That's what an agent does. So an agent, I would say, prepare me for this podcast and it would do all the research necessary. We can even demo a little agent right now if we wanted to do that, a little bit of agentic
Starting point is 00:11:38 work. But it's a tool that goes out and kind of does things in the world. And the idea with an agent is if an agent can go out and do work for us, then I get to skip the whole problem of trying to figure out how to integrate AI in with my workers because the AI will basically be a worker. I'll say, write this code for me and then deploy it and then test it and come back. I think we're further from that than the labs think, but I do think narrow agents are already very possible.
Starting point is 00:12:04 Deepak Chakravartty Have you got a good one available? But I do think narrow agents are already very possible. So deep. Have you got a good one available? I mean, I think, yeah, let's pick an agent for fun here. I will actually share my screen and I'm going to log into an agent while we talk. Okay. So let's look at an early version of an AI agent called Manus. Manus is a small Chinese company. Actually, I don't know how big they are. But they use
Starting point is 00:12:32 Claude as their underlying data set, so Anthropics data. And I just want to show you an agent would do, right? So in this case, I asked it to come up with 20 ideas for marketing slogans for new mail or cheese shop, select the best one, build a financial marketing plan, build a website, carrying cheeses. And so what you'll see is what the model actually does here is it comes up with a do-do list and then it starts coming up with a plan.
Starting point is 00:12:54 So here, it comes with a bunch of slogans and ranks those slogans and then decides on the best one. Then it goes out and actually does market research on the cheese industry. Then it goes ahead and puts together a whole financial plan for us that's many pages long. Then it goes ahead and figures out, and here's a to list that goes through to do this.
Starting point is 00:13:14 It does market research. It comes up with a color scheme for the site. And then ultimately what it does is without any further intervention from me, launches a website. And this is 100% created by the AI. Right? I can go shop cheeses, there's a little built a shopping cart functionality. I've got a subscription model I can do that has forms built in. And this is without any further
Starting point is 00:13:35 intervention from me at all. Right? So the idea is that this is- So how big was your prompt? How big was your prompt here on this? You saw it. The prompt was literally this one, right? That's not that complicated. It figured out all my intent and everything else from this. So this is an example of an agent at work, right? And when will these be widely available? I mean, this is available right now.
Starting point is 00:14:01 And I would also argue that 03, which is the model that everyone can use right now, is basically an agentic model already. So if we just go to ChatUBTs03, and I can just ask a question saying, let's ask it to do something. Let's say something like, give me 10 ideas for a new on trend shoe design based on market research and develop pricing, Proforma financials for the best one and show me a photo shoot of it. And I'm not so interested in the prompt being amazing as to show you what I mean by this.
Starting point is 00:14:57 So this is, 03 is an example of both a reasoning model, a model that goes through this process of thinking. We can talk more about why that's important and interesting if you want to. But what you'll see, what I want to show you about this is it's come up with a plan. Right. I'm going to look up latest trends in footwear, and then I'll come up with 10 designs, then I'll pick the best one. Then I'll create a pricing strategy.
Starting point is 00:15:16 So it's come with a plan. And what you'll see is that it's going to go through this step wise. So just on that command, right, it's going to, you can see it's doing web searches already about to do market research. Now it's choosing a design based on the research that it did. So it's a stepwise, agentic approach, right, that the AI takes even for just a simple command.
Starting point is 00:15:37 This is kind of how an agent works. So you can see it's doing multiple web searches. It's thinking about what this all means for the financials. It's figuring out, that it made a mistake in doing a search, so it's looking for different things that it did before. That's an agent at work. So this agent could have done these whole poll calls, right? And pretended to be you and me. So almost, right? There's a weird disconnect. Video and audio is a little bit different than other approaches to AI, like large language models themselves, but increasingly so. Yes,
Starting point is 00:16:15 you can now bring a live agent onto a call and it has a pretty good conversation with you and live video is getting much better. So if not now within the next months, that's a pretty plausible thing to be able to do. Do you think you would do better than you and me? No. I mean, not just because I want to keep our jobs, right? And I'm not just trying to flatter you, because we still have interview left
Starting point is 00:16:38 and I want you to make me look good. But aside from those issues, right? I think the thing about AI is that mostly it's not superhuman. So you're a very good podcaster, and I would say that anyway. But at the top of your field, whatever you're top at, you're definitely probably better than AI. And by the way, we got our shoe here. And so I think that it's not as good as us yet.
Starting point is 00:17:01 That's the real question is, does it get to be? And does it get to be good across, as good as across every field, which we'd call artificial general intelligence, AGI, a machine smarter than a human expert across every field. And that's the biggest question in AI right now. Absolutely. When will we be there? So that is a thing I don't know the answer to. If you ask the AI labs, they think two years. You ask AI skeptics, they think 10 years, which is a weird place to be for skeptics. To be like, yeah, this is definitely happening. It's just not happening right now.
Starting point is 00:17:32 Yeah. So what aspects of human work do you think AI will complement and what will it replace? So I think that is a really interesting question. The whole premise of the book is about co-intelligence, the idea that the machine works best with a human. That's the same way, right? And so right now, because the AI has these,
Starting point is 00:17:55 has what we call a jagged frontier. It's really good at some stuff you'd expect and not good at other stuff. And it's also missing functions, right? Like it's hard to have it kind of connect the world together. If I ask you to prepare for the podcast, I'm gonna get only so far with it. It will probably give me great preparatory advice, but then I still need to go on the
Starting point is 00:18:09 podcast and have this conversation with you. And it doesn't use my email well yet, so I can't let it interact completely with your team. So there's all these jagged edges that make it hard to use as a universal person replacement. Again, the goal of the AI Labs is agents will solve this problem by doing the work for us. Again, I'm a little reluctant to believe that we're going to be there as fast as they think. But I think compliments is pretty wide. I think compliments across a wide range of tasks.
Starting point is 00:18:41 I think where you're weakest is where it will substitute, but almost at the individual level. So if you're not good at idea generation, the AI is probably better at idea generation than you. If you're good at idea generation, you will definitely get value out of using AI, but you should probably be using your own idea generation as well. If you are terrible at email communication,
Starting point is 00:18:59 the AI is probably better at that than you, but that doesn't mean that you don't have a role to play in making sure the email is shaped properly. Yeah, I use it for a lot of my emails just to help because my English isn't so good. So it just really improves my emails a lot. Now, do you think to which extent is it now being used for CFOs trying to cut costs and to which extent is just amplifying power and helping us to do things better. So this is where companies get to make choices and one of the things I worry about with AI is if the leadership isn't well informed in companies about how they work, they view this as another
Starting point is 00:19:37 normal technology in the sense of like this is a cost-cutting measure. So I can increase productivity by 20% so I can fire 20% of my staff. I think there's two things that worry me about that approach outside of any sort of moral or other kinds of concerns you might have, which is that first of all, no one knows how to use this, right? There is no off the shelf product
Starting point is 00:19:56 that just does things for you with AI yet, they'll come, but you have to figure out how to use it inside your own company. And doing that requires you to actually have experts figure out how it's used, and the experts are your own organization. Your HR department is your R&D. So if you start firing people for using it, because AI makes it more efficient,
Starting point is 00:20:15 everyone just stops showing they're using AI, and you're going to be in trouble. So I think there's some danger in making a cost-cutting move right away. That doesn't mean people aren't doing it. The second big danger for making cost-cutting is, if you believe we're on the edge of a real revolution in how work gets done, which I do, then the idea that you're going to slim yourself down. So if I get 20% performance improvement, I'll cut 20% of people, feels like a really bad solution
Starting point is 00:20:38 in a world where everybody else is going to have 20% performance gain overnight. And so I think that organizations that are in growth mode will tend to outperform those who are using this as a cost cutting technology. But we don't have all the models yet. People are still figuring this out. When we had a breakfast recently, we talked about the role of compliance or general counsel. How are you seeing that?
Starting point is 00:21:02 So the opinion that will probably get in the most trouble is the two most, not universally, but the two most risky places to assign all of your AI responsibility to is often IT and legal. That's not true in every case, right? But legal compliance, the issue is, this is a weird technology. A lot of people know about it is based on rumor. The number of companies I talked to that will refer, where the legal office will refer to an incident where Samsung's data was stolen by ChatGPT,
Starting point is 00:21:30 which never happened, right? What actually happened was that people in Samsung were worried about ChatGPT using their data, so they banned use in the very early days. So it's all rumor based. Right now, these AI models are, I can tell you they're being used at Moderna and they're being used at JP Morgan, like companies that are very worried about data use with legal restrictions are using it, right? You guys are using AI. There
Starting point is 00:21:53 are ways to get around at this point the legal issue. So if the legal team is holding you back, it's because they don't fully understand the problem. And on the IT side, there are some incredibly brilliant and innovative IT people out there who will run with this. But the traditional way that IT handles a project, right, is they'll build a product for you around this and check out vendors and do this approach. And they'll make AI and IT technology as opposed to everybody technology.
Starting point is 00:22:17 And that's another danger is if you just, like we need to build an application for AI, well, we actually need to figure out use cases. Everybody needs to be using this to get there. So those are two different danger spots. They're not universal. I've seen some incredible compliance officers who lead directions with AI, but I've also seen resistance happen from there.
Starting point is 00:22:32 Yeah. No, I think we got a tremendous compliance here. Really, really good. But it's interesting because there are very few cases where a compliance officer can kill a company. I mean, here, if you hold back the usage, you kill your company. Because competition is just pulling apart by 20% a year. And within two years, you're dead.
Starting point is 00:22:55 I think that that urgency you feel is really interesting. I talked to lots of executives and you see this light switch go off for them. A lot of them are treating this as like they put this down seven levels of their organization or they've hired a consultancy who's going to produce a report on their AI readiness. And then you see the executives who kind of get it and there's just night and day because once you get what's happening here, it's very hard to not feel urgency and to not be anxious about resistance everywhere. And that's in our previous conversation,
Starting point is 00:23:25 that's one of the things that struck me was that feeling like, oh, this is the big one and we need to figure this out. And organizations that haven't put that in the list aren't gonna be in trouble. What proportion of companies have got it now, you think? I am surprised by how quickly the religion is spreading, but not as many as you think.
Starting point is 00:23:41 I mean, I talked to a lot of top executives. I would say, it's gone from like two or 3% of people, you know, getting it to where 20% of executives in a lot of the firms that should feel urgency feeling it. But that's a pretty big increase in a short time. So, I mean, this technology is remarkably rapidly adopted. Slight change of topic. How do you stay updated through these extraordinarily fast changing times? So, I think I'm probably one of the most kind of, you know, current people on using this. I've got this virtuous cycle going, right, which is as somebody who does a lot of work
Starting point is 00:24:23 in AI and is influential in the space, all the AI labs come to me. I don't take money from them, but they all give me early access to stuff, so I know what's coming. I'm on all these weird private discords and conversations. I'm on X and blue sky. So, but you know, I'm a professor who is on sabbatical and spends a lot of time thinking about AI stuff, and I'm having trouble keeping up. I think keeping up is challenging.
Starting point is 00:24:45 At the other hand, I don't think you need to that much, right? I actually think if you go with like a chat GPT or Gemini and just use it a bunch, those tend to have the really up-to-date models, but it's hard to kind of keep up otherwise. I mean, I've got a newsletter, so people can subscribe to that. But there isn't one sort of great source on this. And AI is still, it's sort of like how it's treating organizations. For a lot of publications, it's like one,
Starting point is 00:25:09 it's spread across multiple parts of their organization. So it's reported a little bit in politics, a little bit in technology, a little bit in business. So nobody really has a full picture, I think, including people in the AI labs. People move from one company to the other, right? And it seems like these models are not ahead for a long period of time. They are being overtaken all the time by other things, right? Is this something that will continue? So a lot of questions of the future are unclear. I think, you know, so the frontier models, the best models are in one point. There's only a few companies that can afford to make them at this point. And so generally you want to stay close to one of the model makers, right?
Starting point is 00:25:46 So the people who make frontier closed source models are, you know, open AI and anthropic and Google by and large, right? There are some other options out there, but those are sort of the three big closed source ones. Generally, if you go with one of those, they're going to stay in the frontier for the foreseeable future. There's not a reason to suspect that they might fall four months behind for a little while.
Starting point is 00:26:06 If that matters to you, that's what your lab is supposed to be doing in your company is like, how good is the new model should we switch over? Somebody else has to be doing that testing 24 seven. Another thing always surprised me in companies is how few of them have people assigned 24 seven to just working with AI. Like it just, there's lots of other departments
Starting point is 00:26:24 that work on things, but there's very few people whose job is to stay on top of these things. So you're in the lead, you get all this stuff, you're invited to pre-releases. When you now look into the future here, what's been the biggest surprise for you lately? So I think the biggest surprise for a lot of us has been this idea of reasoner models that you kind of see here, right? So I showed you a little bit of this as an example earlier, but it turns out that models that sort of think out loud outperform those that don't. And this very kind of simple trick has
Starting point is 00:26:58 increased the ability of AI by a tremendous amount. So I think the capability curve is coming faster than I expected it to. And that's been a big surprise. And then the other side of it that's been a big surprise is how fast adoption has occurred. So this is a very fast adoptive technology, according to any historical precedent. We're probably up to a billion people using, you know, chat GPT at this point. The last numbers they released were somewhere between 500 million and a billion people. There's another few hundred million using other models. This is an insanely high adoption rate for technology that sometimes doesn't work or is weird to use and where we don't quite know what it's good or bad for yet. And so I think that the speed of adoption and the speed of capability gain are both faster than I thought.
Starting point is 00:27:43 By the end of the year, what is it that it can do that it can't do now? So I think the arbitrary end of year deadline comes right in the middle of where agentic systems may actually be useful. We're starting to see these narrow purpose agents work quite well. So if you use AI for coding work, for example, you're starting to see that be very
Starting point is 00:28:07 useful, but the agents where you just say, fix the code for me, still kind of flawed. A narrow agent like Deep Research is a very good research agent already and replaces a lot of what you'd have an analyst do and increasingly will go into territory that was lawyer territory and other places of doing research and analysis and pulling stuff together. We're not quite there yet, but it's getting there. So I think the question is how quickly you could just assign an agent, do this job for me and it does a reasonably good pass that you still need a human interaction with. And I think we might be there for many jobs before the end of the year.
Starting point is 00:28:39 When can I say, hey, I'm going to be in New York first week of November, please look up the 10 best Mexican restaurants and book a table 730. I mean, I, I'm going to be in New York first week of November. Please look up the 10 best Mexican restaurants and book a table, 730. I mean, I think we're there already. That's not a hard problem. The only question is the book a restaurant. And if you use Manus, it would probably be able to email that person or do that.
Starting point is 00:28:57 That's an easy one. And I think your current AI will do everything to book the table for you. And that's an example of a connector that we just need to connect. It does voice. In fact, Google has been, even before generative AI took off, has a model where you try and book a table at a restaurant
Starting point is 00:29:14 that doesn't take reservations. It will call and have a voice call them, and an AI try and book the reservation for you. So that's already there. But one of the interesting things about AI is it's already there almost. You can't just chat to BT to do it. There's a little bit of hoops to jump through. So it's not a capability problem. It's a user experience problem. It's a UI problem. It's a communication problem. And that's, I think, AI ability, Fireout strips our ability to use it.
Starting point is 00:29:39 Now you wrote about Cybernetic teammate. When have you seen teammates most strikingly or before my human colleague? So this is again, a case of co-intelligence, right? Of working with the AI. So the cybernetic teammate paper, we again, with my colleagues at Harvard and MIT, University of Warwick and at Penn, we went out and we did an experiment Procter and Gamble, where they gave us 776 of their employees and we did real work that they actually do. So, you know, whether that's product analysis work or
Starting point is 00:30:17 product development or marketing, and we had them either work alone or in cross-functional teams of two with one marketing person, one business person, and one technical person. What we found was that individuals working alone with AI had the same performance statistically as teams of two. They also produced more diverse ideas than if they were working alone, and they were happier.
Starting point is 00:30:42 That's the cybernetic teammate idea of working with the AI to do things. I think a lot of people are already doing cybernetic teamwork, right? I know you are. You consult AI about all sorts of things. Now you have a small team working for you, you know, a panel of experts. So I think that people who use this, this is the natural way to use AI actually, is as a teammate where it's helping you and you're filling on goals on it. So I think we're seeing pretty big gains across the board from that approach. How can we best develop the people in the usage of this? What's the best way to really drive? I mean, we talked a bit about the institution and how you drive the institution forward.
Starting point is 00:31:19 What's the best way to drive the individuals? So I think that we are still learning about how to do that. This is not like traditional training that we give you four or five rules for using AI. I mean, I do think there's some things to learn, but a lot of training focuses on like prompting techniques, and prompting techniques turn out to be less and less important as time goes on. AI models are fairly resistant. Why is that becoming this important?
Starting point is 00:31:43 Well, two things I think are happening. One is the models are getting better, and larger, better models get your intent better. And the second is it turns out we just don't know a lot about prompting. It's very contingent. We have actually a little study that we did at the General AI Lives at Wharton,
Starting point is 00:31:56 where we measured the accuracy of AI on answering questions if you were polite to it versus if you were mean to it. And it turns out that politeness matters a huge amount on certain answers. If you do 100 tests on a particular math problem, it turns out in that math problem, if you say please, you get more accurate answers than if you don't say please, right? But it turns out on a separate math problem, saying please makes it do worse and yelling at it does better.
Starting point is 00:32:22 And we don't know why it works in one case, but not the other. So the average effect is zero, basically, but the individual question effect is quite large. So if we told people, always be polite or always use this approach, it's sometimes going to work, sometimes it's going to backfire. And the best- So I normally say, please, I should just not do it. It doesn't seem to help.
Starting point is 00:32:42 But then again, I think there's value in the mental model of treating AI like a person, and if please helps you do that, even though it's not a person, then I think it's no problem. And it's just kind of hard otherwise, right? It feels weird to just order your computer around. And then if you talk to people who really believe super intelligence is coming soon, you better be polite now because they'll know. How do you test for AI literacy when you interview people in a job interview? I want to avoid over-indexing on today's literacy because we don't have a definition of AI literacy.
Starting point is 00:33:15 I mean, I teach AI and we don't have a definition of being AI literate. It tends to be you're an AI user and then the question is, are you a sophisticated user or not? Does that matter to you? Because it might turn out that they're not a sophisticated user, but a small amount of experience with a sophisticated user will make them good. If we're judging AI literacy today, you're really judging people's independent ability
Starting point is 00:33:35 to go out and figure out ways to use AI in their job. And that can be valuable, but that doesn't always tie in with subject matter expertise and other issues. So I would just ask people at this point, show me, you know, when we were hiring people for the generative AI lab, we both asked them how they used AI, we would give them a task and say, use AI to accomplish the impossible task and get as far as you can in an hour, right? And we would also be asking people about, you know, actually interviewing them about what's working, what isn't for their AI use and having them show us some of that that. So I mean, I think you would judge AI,
Starting point is 00:34:06 but that's kind of an example of usage and creativity with AI, and we're an AI lab, so we kind of need those people. I worry at this stage about companies picking AI literacy as the major issue. One of the things we found from the BCG study was that more junior people were actually worse at using AI in the organization than more senior people.
Starting point is 00:34:24 They may have gotten AI, but they didn't understand the organization. So they would produce a document and say, this is a great memo. And then someone with 20 years of experience would look at the memo and say, no, it needs to be better in these following six ways. They would have been better using AI than the junior person. So I worry, it kind of goes back to that chief AI officer example you gave earlier. Companies are trying to hire their way out of this problem. And I don't think there's a way of hiring your way out of the problem without the executives
Starting point is 00:34:46 also engaging deeply and making these decisions about what AI means in their organization. What personality type do you think will adopt the best? So the model, we talk about agency a lot. We don't fully know what it means. I teach entrepreneurship, so there's an entrepreneurial action or a tendency to entrepreneurship that I think is the same thing as agency, which is a feeling that you have control over your environment, a locus of control, that you are in control of your own destiny and you should seek out opportunities to do things that are different or better improved.
Starting point is 00:35:22 That is not something everybody has. That seems to play a role in AI, but we don't have measurements of it. So I'm a social scientist. I'm making stuff up here, which always makes me feel bad because we don't have a measure. I don't have an AI measurement
Starting point is 00:35:33 that makes people, that can tell you, but there is this sort of agentic sense. On the other hand, some of the most entrepreneurial interesting people I know are late adopters of AI. For whatever reason, they hit a barrier when they were using it. And I'm not sure what that is.
Starting point is 00:35:45 So again, we don't have easy answers to almost any question about AI at this point. Two slightly different ones. How do you see the regulation in Europe versus the US? So I'm not a regulation expert, but I will tell you that one of the interesting things that happened with the release of DeepSeq, which is a Chinese open source model, R1, DeepSeq R1 specifically, was it was the first time a Chinese model was on the frontier of capabilities. It did not beat American models, they're better again at this point, but it's a typical closed source open source.
Starting point is 00:36:25 There are a few months behind, but a very good model. It caused a bit of a panic in the US that I don't think was warranted necessarily of even from a great powers competition standpoint. But that sort of, that plus the new administration, I think has created this opportunity where there's very little regulation. I mean, in fact, one of the bills in front of Congress right now in the US is that this
Starting point is 00:36:49 would ban AI regulation at the state level for the next 10 years. I don't know whether that passes or not, but I think there is a desire to put in the accelerator and less regulation and more sort of letting AI rip. I think that my impression from talking to people in Europe has been there has been a move to less regulation from a very regulated viewpoint. I think part of this came down to the idea that a lot of the initial regulation was based around existential risk, which I think is important, but we don't know how to measure. And now I think we have to start moving towards a place where we're regulating harm. So even though there was this there's a bill saying you can't regulate AI, another bill went through Congress banning deepfakes. So I think we're going to have to regulate the outcomes of AI, but I think that Europe is still a much tighter regulatory
Starting point is 00:37:35 environment, but is probably looser than it was a couple of years ago on the AI front. You mentioned existential risk. Is there an existential risk? the AI front. You mentioned existential risk. Is there an existential risk? I have no idea, but a lot of very smart people think there is. For me, the idea that AI gets smart enough that it takes over the world doesn't feel real to me, but that doesn't mean that it isn't. As a good social scientist, I have to realize my own views are kind of secondary
Starting point is 00:38:05 to what does smart people say if we have a forecasting contest. And there's a lot of the founders of AI think that it's an existential danger to humanity. A lot of very smart people think that way. So we have to view it as an existential risk as well. The problem is we don't know what form that takes and we don't know what the level is. Originally a GPT-4 or GPT-03 class model was considered an existential risk level before we had these things. Now we have them, we're like, these aren't going to let anyone take over the world or spontaneously go out and destroy the economy. Is there a magical point to where that happens?
Starting point is 00:38:39 I just don't know. Well, Ethan, this has been 40 incredible minutes. You are at the very forefront of the very forefront. It's been so interesting talking to you. So keep up the good work. Look forward to staying in touch and see what the future brings. Thank you. It's a pleasure being here. And I will say it's exciting to talk to people who get the AI thing, because I think it is important and spreading that news is important. We can't go with eyes closed. We have to make decisions about the future
Starting point is 00:39:09 and you can only do that if you're using these things. Absolutely, fantastic. Yeah.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.