Big Technology Podcast - Could LLMs Be The Route To Superintelligence? — With Mustafa Suleyman

Episode Date: November 12, 2025

Mustafa Suleyman is the CEO of Microsoft AI and the head of the company’s new superintelligence team. Suleyman joins Big Technology to discuss Microsoft’s push toward “humanist superintelligence...” and what changes after its latest OpenAI deal. Tune in to hear whether LLMs can get us there, how self-improving systems might work safely, and what power, data, and memory advancements mean for progress. We also cover Microsoft’s strategy shift to AI self-sufficiency, the economics of frontier models (including price pressure and commoditization), world-model and robotics questions, and the rise of personalized AI companions. Hit play for a candid, technical, and forward-looking conversation about where Microsoft—and AI—are headed next. --- Enjoying Big Technology Podcast? Please rate us five stars ⭐⭐⭐⭐⭐ in your podcast app of choice. Want a discount for Big Technology on Substack + Discord? Here’s 25% off for the first year: https://www.bigtechnology.com/subscribe?coupon=0843016b Questions? Feedback? Write to: bigtechnologypodcast@gmail.com Learn more about your ad choices. Visit megaphone.fm/adchoices Learn more about your ad choices. Visit megaphone.fm/adchoices

Transcript
Discussion (0)
Starting point is 00:00:00 Microsoft's AI CEO returns to explain why the company is now pushing for superintelligence, what that means, and how Microsoft is moving forward after its latest open AI deal. That's coming up right after this. Industrial X Unleashed is bringing together leaders from IFS, Anthropic, Boston Dynamics, Microsoft, Siemens, and the world's most progressive industrial companies at the frontier of industrial AI applied in the real world. There's a clear market shift happening. The world's largest industrial enterprises are done experimenting with AI. They're deploying it at scale, and they're choosing IFS to co-innovate with them. IFS is purpose-built for asset and service-intensive industries, manufacturing, energy, aerospace,
Starting point is 00:00:40 construction, where downtime costs millions and safety is non-negotiable. Industrial X Unleashed will feature live demos of AI products embedded in real-world operations, customers sharing measurable outcomes, and learnings from companies deploying industrial AI at scale today. Learn more at industrialx.a.i. The truth is AI security is identity. security. An AI agent isn't just a piece of code. It's a first-class citizen in your digital ecosystem and it needs to be treated like one. That's why ACTA is taking the lead to secure these AI agents. The key to unlocking this new layer of protection and identity security fabric.
Starting point is 00:01:15 Organizations need a unified, comprehensive approach that protects every identity, human or machine with consistent policies and oversight. Don't wait for a security incident to realize your AI agents are a massive blind spot. Learn how ACTA's identity security fabric can help you secure the next generation of identities, including your AI agents. Visit octa.com. That's okayta.com. Welcome to Big Technology Podcast, a show for cool-headed and nuanced conversation of the tech world and beyond. Today we're joined once again by Mustafa Suleiman, the CEO of Microsoft AI, and also the head of the company's new superintelligence team, who is here to speak with
Starting point is 00:01:54 us about what that means, what superintelligence is, but more broadly what the future of this technology is going to look like, and whether we're at the end of the curve or the beginning or somewhere in the middle. Anyway, we'll get into it all. Mustafa, great to see you again. Welcome to the show. Hey, Alex. Great to see you again. Thanks for having me. It's always a pleasure. And so recently you wrote this post about a new push towards what you call humanist super intelligence at Microsoft. You say you're working towards it, what you call incredibly advanced AI capabilities that always work for in service of people and humanity more generally. Let me ask you a question about this.
Starting point is 00:02:37 It's so interesting to me to see so many labs running towards what they call superintelligence, which I guess is sort of like a cooler version of AGI, as the research is mixed about whether we're going to see a lot more progress with the current paradigm. A lot of people are talking about diminishing marginal returns. I think we've talked about that. There are some questions about the viability of LLMs in terms of pushing the state of the art and AI forward. And yet we're also seeing this push towards superintelligence. So just explain as we begin sort of the discrepancy there.
Starting point is 00:03:12 Why are we hearing so much about superintelligence where we're not even sure if the current methods are going to get us to the step before, which is AGI? Yeah, I mean, superintelligence and AGI are really goals rather than methods. And I think that the ambition is to create superhuman performance at most or all human tasks. Like we want to have medical superintelligence. We want to have the best expertise in medical diagnosis be cheap and abundant and available to billions of people around the world. We also want to have world-class legal advice on tap that costs almost nothing, a few bucks a month. We want to have financial advice, we want to have emotional support, we want to have software engineers available on tap, and I think that the project of superintelligence is about saying
Starting point is 00:04:09 what type of very, very powerful intelligence systems are we actually going to build? And what I'm trying to propose is that we subject each of these new technologies to a very simple test. Like, does it in practice actually improve the prospects of human civilization? And does it always keep humanity at the top of the food chain? It sounds like a kind of simplistic or obvious thing to have to declare. But the goal of science and technology, science and technology, in my opinion, is like to advance human civilization, to keep humans in control and to create benefits for all humans. And I think in some of the rhetoric in the last few years, you can feel that there's a little bit of like, you know, a kind of creeping assumption that it is inevitable
Starting point is 00:05:06 that these kinds of systems exceed our control and our capability and move beyond us as a species, as a human species. And I'm pushing back on that idea with the framing around humanist superintelligence. I think it's quite different. But then is your view that super intelligence won't be one broad intelligence, that it will be, you can maybe achieve superintelligence in one discipline when it's smarter than, let's say, the best doctors in medicine, but maybe it's just not there in accounting, for example?
Starting point is 00:05:42 One way of thinking about it is that how we train these models at the moment is that we work through verticals and we make sure that we have training data, knowledge, expertise, reasoning traces, chains of thought that reflect the kinds of activities that people do in each one of these disciplines to build their expertise overall. So we're already training generalist models from a verticalized position. We're starting off by saying what specific tasks are we trying to optimize. And, you know, the project of humanist superintelligence, is first trying to say, what good will this technology do
Starting point is 00:06:22 and how will it be safe and controllable and aligned to human interests? And one of those dimensions of safety is verticalization. If a model has been designed explicitly to achieve medical superintelligence, then by definition, it isn't going to be the best software engineer in the world. It isn't going to be the best mathematician or physicist. And so narrowing the domain, not too much, not entirely, because you can't collapse it,
Starting point is 00:06:53 but narrowing it and reducing the generality is one of the ways that I think is likely to help create more control. It's not the only solution. There are many other aspects of like how we achieve containment and alignment, but domain specific models are one part of it. Is it possible that something can be super intelligent but not generally intelligent? Like, is it possible that maybe superintelligence happens without AGI? Because AGI is all about generality and what you're talking about is not.
Starting point is 00:07:24 It's not possible, I don't think. I think they need to be general. They need to transfer knowledge from one domain to another. They need to, you know, have generalist reasoning capabilities. But when you apply it and you put it into production and you let it have more autonomy to make decisions, or you let it generate arbitrary code to solve a particular problem, or you let it write its own e-vows so that it can modify its own code and generate new prompts to generate new training data to write new evows
Starting point is 00:07:58 to then iterate on its own performance. These capabilities, autonomy, goal setting, writing code, modifying itself, you know, if you add to that, then also are perfectly, generalized model or sort of general purpose model, that's a very, very, very powerful system, which today I don't think anybody really knows how we would contain or align something like that. And so it's not to say that we should not do any one of those dimensions. It's just to outline a roadmap of capabilities which we're all working on, which add more risk, especially when they compound with one another and you combine them all together. And so, you know, my claim is
Starting point is 00:08:41 that we should just approach this with caution, remembering that we don't want to bundle together all these capabilities so that there's a higher risk of a, you know, recursively self-improving exponential takeoff that then replaces our species. And I think that that is very low probability from what I see today, but it's one that we have to take seriously in the next like 10 years or so. Okay, I do want to get to that in a bit, but let me tell you what I find odd about these conversations. And I want to go back to the first question that I asked you, which is researchers are talking about how the current methods are leveling off. Give you one example. Data is not plentiful. Synthetic data, not very useful yet. Power might be
Starting point is 00:09:28 running out. And you need that scale, a lot of people say, in order to make these models better, or at least even to run the basic capabilities. So given the limitation, of LOMs, are you seeing something that we're not that will sort of pave the way to superintelligence? I mean, how do you get from here to there? Look, I think we're power limited but not fundamentally power constrained. Clearly, there's like huge appetite to build bigger data centers and train in larger, more contiguous, more fully connected clusters. So clusters are where all the chips are connected to each other. But that's not the bottleneck at the moment. That's not holding back progress. Obviously, if we had more right now, it would definitely
Starting point is 00:10:13 help. But there's many, many other things in the stack that are slowing down progress. If we are not data constrained right now, we're generating vast amounts of high quality synthetic data, which is proving to be useful. Obviously, again, the same is also true. Like, more high quality data would be great. But I don't see a slowing down in progress because of either of those two things. if anything, the rate of progress has been insane over the last five years and to expect us to continue to make doublesings every three months in the size of clusters that are trained for the largest models, you know, given the base that we're now starting from when training runs are often, you know, 50 megawatt or 100 megawatt or soon 500 megawatt.
Starting point is 00:10:58 You know, you can't just double on that every six months. There's like the laws of physics kind of do create restrictions. And we're talking about tens of billions of dollars of, you know, cluster. so pace might slow a little maybe but it's also clear that pace is still going to be unbelievably fast like you know um sort of sort of objectively speaking so i i don't see or fear or currently feel any sense that things are slowing down or that we're losing momentum and it's just quite the opposite well then let me ask it this way do you think lLMs are the way there um Look, I think one thing to consider is that every year for the last few years, there's been a major new contribution to the field, still principally based around the transformer architecture, but we're bending the transformer architecture into new shapes all the time. Fine-tuning emerged three years ago on top of our pre-trained models to adapt them to specific
Starting point is 00:12:03 use cases. They're now fully multimodal, which requires further changes and the introduction of diffusion models. Then we had reasoning models in the last 12 months, which again are still fundamentally based on the same core architecture. Things are just rearranged slightly differently. So, you know, even though the scaling laws weren't able to continue exponentially in the way they had from such a low base, new methods. appear on top of those like reasoning and new methods will come you know even newer methods will come soon too so for example um i expect that there's going to be quite a lot of progress in recurrency soon right the moment you know the models don't kind of attend to their working memory
Starting point is 00:12:52 very well um you know at the moment when they're training right and so you know i think people are experimenting with lots of different types of loss function and lots of um training objectives The other one is memory. I think memory is getting better and better and I think is going to totally change what's possible. And the other one is the sort of length of a task horizon that can be predicted. So at the moment it's like a few steps, but soon it will be tens of thousands, hundreds of thousands of steps accurately. And that will mean that a model can use APIs or query a human or check another database or call on another AI. And so that will be another like sort of exponential lift when something.
Starting point is 00:13:33 like any one of those three things work, you'll get another kind of, you know, rapid acceleration in progress. So I don't think there's anything fundamentally wrong with the LLM architecture, and I don't think we're fundamentally compute or data constrained. I think that there are so many people focused on this problem now. There are just going to be more and more, you know, breakthroughs coming. Okay, that's very interesting. So your perspective, basically, is that LMs are the path.
Starting point is 00:13:56 Yeah. That we don't need another breakthrough that's a different model format to get toward superintelligence. Well, I mean, so far, no, I don't think so. I mean, so far deep learning and the transformer model has been the workhorse for, I guess, like, 12 years, you know, since Kraschewski and Alexnet. And, you know, there's been variations on a theme, but it's been delivering. And I don't think it's fair to say that it's like not delivering at the moment. I think it's, I think it's really making a lot of progress. Yeah, it's definitely delivering.
Starting point is 00:14:33 And it's so funny because whenever I'm like bringing up these criticisms, it's like some way I'm saying to myself, what do you want? The computer is talking to you. But the question is sort of, right? I feel silly being like, well, where's the more improvement? But I think when we hear words like super intelligence, then we see the gap between where we are today and where you want to head. And those questions naturally come up. And just to go back to the power thing, I was sort of struck by Satya's comments in the podcast with Brad Gersner, where he said he has GPUs or chips that aren't plugged in yet, but he needs warm shelves for them. So I'm curious to hear your perspective.
Starting point is 00:15:20 If we're not power constrained right now, how does that square up with the inability to plug these chips in right now? Well, I think what he was referring to is that we have. have so much inference demand that we're power constrained on inference. We're not power constrained, at least from the Microsoft AI perspective, on training chips. And obviously, my team is mostly focused on training right now. So obviously, copilot is inference constrained and desk really needs more chips to scale. And so does M365 and our other products. One more thing I want to talk to you about on this superintelligence push is the world model. Yeah, a lot of people have talked about how these are models are trained on text and some video.
Starting point is 00:16:07 I mean, it's actually been amazing to watch them be able to create video that has some understanding of physics and liquids and lighting. It's not really supposed to happen that way, but it's doing it. But there's been questions about whether models understand gravity and what happens in the real world. And L.M. can't drive a car right now. So how's it going to be super intelligence? So I am curious to hear your perspective on what's needed to or whether it's really a priority to figure out like the physical world and if so, how you get there. Yeah, that's a good question. I mean, right now, you know, it's actually amazing, as you say, that models can learn from a compressed representation of reality and then produce a version of reality which looks like.
Starting point is 00:16:57 the thing that has been compressed from. I mean, it's just like text and the description, text describes the physical world and the properties of the physical world. The model has never seen that. And then actually is able to produce very compelling stories, code, business plans, videos, and so on. So it's surprising that we've come so far with that structure. I'm kind of open-minded about like, you know,
Starting point is 00:17:27 sort of robotics and streams of input from the real world. I mean, I think that you, my instinct is that you can't just like crudely pile this data into existing pre-training runs because, you know, those runs have tokenized or they've sort of described text data in a certain way. And that, you know, meshing that with other like telemetry data from a robotic art. for example, you'd have to think about, like, at what level of abstraction to do that. And obviously, there's good specialist models that have become pretty good at that. But I don't think right now, at least, that, like, that is holding us back. I think in general, more data is always better,
Starting point is 00:18:17 but, you know, I don't think in the next few years it's going to be the big differentiator. I think that more synthetic data, more human feedback and high quality data is going to to be the differentiator. Okay, so you brought up recursively self-improving AI models, and maybe that is where this path towards superintelligence goes. Open AI has said they want to build an automated AI researcher by 2028, and I think every lab, I'm curious if this is your interest as well, is just trying to build AI that improves itself.
Starting point is 00:18:55 Is that realistic? I think that in some ways the RL loop is already doing that. And at the moment, there are human engineers who are in the loop who are generating data and writing e-vows and deciding what other data goes into training runs and running ablations on that data. You can well imagine different parts of that stack being automated by sub-components of AIs. It doesn't necessarily mean that one single system does it. Today we have, you know, RLHF, the human feedback, grew into RLAIF, where we have AI judges or AI raters to judge the quality and the usefulness of data that was also AI generated. And in many cases, prompts that are used to generate diverse training data were also AI generated.
Starting point is 00:19:48 So, like, you know, today we're at a point where data, the core complex, commodity, which is sort of driving the progress of these models, is, you know, albeit not completely automatically in a closed loop way at large scale, you know, individual parts of that pipeline have been, you know, developed by, you know, LLMs. So it doesn't seem very far-fetched to say that in a few years' time, at significant scale, that will get closed loop. And, you know, it'll be interesting to see on, you know, what happens and whether the quality bar can be maintained and where the performance does increase. I think it will. But it's
Starting point is 00:20:29 definitely something to be very cautious about because, you know, a system like that could end up being very, very powerful. Yeah. I definitely want to talk to you about the downsides of it. But we had a debate on the show recently about whether that is an ambitious thing. It even seems funny to say. But to me, that's the ultimate ambition, right? If you're able to do that, then you get into a situation where, you know, potentially you have fast takeoff of intelligence. But I guess it's hard to really imagine the, and maybe my imagination isn't there. The AI's finding the, you know, the next new method, like discovering reasoning on their own. So talk about both of those, the ambition and then whether, whether I'm just, my imagination is too small on this front.
Starting point is 00:21:16 I mean, I think the self-play work that we did a deep mind, you know, back sort of six or seven years ago now with Alpha Zero, you know, that obviously paved the way to the first large-scale, you know, sort of self-improvement effort, frankly. And I think everybody in the field is aware that it can be done in a certain domain where there's verifiable rewards and where you're in a kind of closed loop gaming type environment or simulation. environment and I think people are thinking hard about how it might be possible to recreate some of the components of that in this setting and you know I do think that's going to drive a lot of progress in the next few years I think it's a big area that everybody's focused on you know because fundamentally scale always ends up trumping you know you know anything else and so if if you can have models explore the space of all possible you know sort of combinations in a compute efficient way then it may well discover reasoning
Starting point is 00:22:29 by itself it may discover um you know new knowledge that we hadn't even you know thought about ourselves or even like found in in in any training data to represent that knowledge so But it is highly inefficient, right? I mean, learning from supervised examples with SFT and stuff like that, like imitation learning is very efficient and clearly works very well because these models learn from, you know, just as we've talked about, an incredible amount from, you know, from web text, which is really just an artifact or a record of human interactions. But both are going to be true. I think the RL paradigm that involves more online learning from streams of experience is also quite promising. And I think is kind of adjacent to, if not orthogonal, to imitation learning. So both of those experiments will sort of accelerate in the next few years.
Starting point is 00:23:32 Now where could this go wrong? Well, I think being in the loop as a human developer adds a certain amount of friction. And that oversight is quite important, I think. You know, if a system like that had an unbounded amount of compute, it would end up being incredibly powerful. And I think we have to sort of figure out how to force these models to communicate in a language that is understandable to us humans. You know, and that's like a very obvious safety thing, to be able to regulate the language. that it uses so that it, you know, we're already seeing examples of what some people are calling deception, but it's really just like kind of reward hacking. Hacking kind of implies too
Starting point is 00:24:24 much intentionality. So it's just, it's, it's an accidental exploit, is found a path like, you know, to satisfying the reward or achieving the reward, you know, in unintended ways. And so we shouldn't anthropomorphize it. It didn't deceive us. It didn't intentionally. tried to hack us, it just found an exploit. And that's a problem with poor specification of the training objective and of the reward function. And so, you know, the way that we make that safer is that we get sharper in our articulation of like, what is it that we're actually trying to train for? What are we, you know, what are we trying to achieve? What are we trying to prevent? And then monitor, like, you know, monitor outputs during training time rather than, you know,
Starting point is 00:25:13 reasoning traces, chains of thought, and so on, rather than just at, like, the final stage. So as we grant these models more capacity to self-improve, we're going to have to change the framework with which that we use to kind of provide oversight to them during training. We're here with Mustafa Suleiman. He is the CEO of Microsoft AI on the other side of this break. We are going to talk about, well, it seems like there's a little bit of a strategy shift here. It's gone. Microsoft AI has gone from wanting to work on.
Starting point is 00:25:43 and the frontier of the best models but not building them themselves to trying to build superintelligence. So why now? And what does it mean now that Microsoft AI and Open AI have a new agreement? We will cover that right after this. Capital One's tech team isn't just talking about multi-agentic AI. They already deployed one. It's called chat concierge and it's simplifying car shopping. Using self-reflection and layered reasoning with live API checks, It doesn't just help buyers find a car they love.
Starting point is 00:26:15 It helps schedule a test drive, get pre-approved for financing, and estimate trade and value. Advanced, intuitive, and deployed. That's how they stack. That's technology at Capital One. And we're back here on Big Technology podcast with Mustafa Suleiman. He's the CEO of Microsoft AI. Mustafa, is it a coincidence that Microsoft just came to this agreement with OpenAI that you could go ahead and attempt to build AGI on your own that you've now decided,
Starting point is 00:26:46 hmm, let's go ahead and start a super intelligence team or is that directly related like I think it is? No, I think it's directly related. You know, I think that the Microsoft Open AI partnership is going to go down as one of the most successful partnerships in technology history for both sides. You know, Satya did this deal, you know, at a certain time when there was a of risk and huge amount of upside. And I think, you know, the last five years have turned out amazingly well for Microsoft. But then Satya made a call that, like, you know, we've, we've also got to make sure that, you know, we're self-sufficient in AI. For a company of our size, it's inconceivable that we could just be dependent on a, you know, on a startup, on a third-party
Starting point is 00:27:30 company to provide us with such important IP. And so, you know, we basically took the view that we should extend the IP license through to 2032. We'll continue to get model drops from OpenAI and all their IP. We'll continue to be their primary compute provider, a huge scale to the tune of, you know, billions and billions of dollars. And also, we would remove the clause in the contract that says that we couldn't build superintelligence or AGI. And that was actually expressed as a flops threshold, a flops per second threshold for a size of a certain trade. running run. So there's a big limitation on what we were able to do. Now that that is no longer there, you know, our team is reforming around this idea of humanist superintelligence. We're pursuing
Starting point is 00:28:22 the absolute frontier, training omnimodels of all sizes, all weight classes to the absolute max capability. And over the next two or three years, you'll see us really try to build out one of the top labs in the world. We want to train the absolute best AI models on the planet. it. And, you know, we're a very young lab. We've barely been going for a year. But, you know, we've got some good models on the leaderboards, text and image and audio now. And, you know, over the next few years, we'll be striving to be the absolute best we can. I was just speaking with the chief technology officer of a pretty big technology company. And this company has decided not to build their own large language models. And it sounds a little bit wild. But I think it makes
Starting point is 00:29:06 sense in a way that there's going to be obviously like to build these models it's extremely expensive uh resource intensive you don't always get a payoff like we saw that with meta and lama i'm not saying that's what's going to happen with you um and maybe it makes sense just to you know buy off the shelf or uh use open source and in fact that seemed like that was the strategy that you had for a long time it seems logical um and so i'm curious like why you would disagree with that Why is it so important to build your own models? I mean, we're going through a foundational platform shift, you know, in software. From the operating system to apps, from browsers, search engines, mobile, social.
Starting point is 00:29:53 This is the next major platform, and it's going to be bigger than all of the other platforms put together. So the idea that a $3 trillion company with $300 billion of revenue and 80% of the S&P 500 on our Azure, stack and M365 stack, you know, could depend on a third party. It's, you know, just in perpetuity. It doesn't make sense. So we, you know, this is a company that's been around for 50 years and navigated many of the past platform shifts incredibly well. And that's the, that's the journey that we're on. We have to be AI self-sufficient. There's an important mission that Satya set last year. And I think that we're, we're now on a path to be able to do that. And so, hence the formation of the super intelligence team.
Starting point is 00:30:35 Exactly. So we're launching the super intelligence team. We're going to be focused on, you know, SOTA at all levels, but also pushing the frontier of research. I mean, there are many hard problems in machine learning which, you know, a few months ago we weren't really focused on. Continual learning being one. Like how do we store representations of knowledge in a way that they're modifiable by different networks and they kind of accumulate knowledge over time just as humans do rather than having to retrain them from scratch? So that's just like one of many examples of more fundamental research questions that our superintelligence team is now going to spend time on. Okay. Now, let's go back to the business side of it. Your episode is going to air back to back with an episode that will run with Nick Clegg, the former president of global affairs at Meta. And, you know, meta, of course, they also have a superintelligence lab. And we were talking about the economics of it. And Nick's point was very interesting. He said, I don't. don't see how you can hoard super intelligence if you build it. I think his idea is if meta builds it, the Microsoft will build it and OpenAI will build it. And we've seen very fast follows in many of these labs after they come up with a state of the art model.
Starting point is 00:31:51 And so the question is, will it commoditize? Will it be economically viable once two companies build it? What do you think? Well, it's definitely commoditizing. You know, the cost per token has come down 1,000 X in the last two years. It's just a crazy, crazy thought, right? So things are getting massively cheaper and more efficient.
Starting point is 00:32:13 And, you know, the top four or five models are within, you know, a few tiny percentage points of each other in terms of performance. But that doesn't mean that one can afford to leave that to the market and just hope that somebody open-sources it that we can use their open-source models. For a company of our scale, we have to be able to do that. And I think, you know, Microsoft is a platform of platforms, you know, like our API is critical.
Starting point is 00:32:40 Many, many people depend on it. And I think if you're a, you know, a smaller software company or a technology company of any kind, I think you can depend on the market, right, which is very different. So Amazon, Google, us, Anthropic, I guess OpenAI, are all providing, you know, APIs to the very best language models in the world. world and that means that you know you as a buyer even if you're a large public company can feel pretty assured that for the long term there's going to be healthy competitive forces driving down prices and improving quality for you to be able to use you know models via the API right and so I understand why you'd want to build it but again going back to this question it's like okay if it just doesn't seem like there won't be a price war
Starting point is 00:33:34 I mean, if you have a couple of companies that do this. Yeah, yeah. Well, I mean, I think a price war is a great thing for consumers and for businesses. I mean, we're bringing down the cost of intelligence. I mean, I think that's an amazing story for humanity. The ability to access knowledge, the ability to use that knowledge to get stuff done, to write new programs, to do new scientific discovery, to get access to AI companions and emotional support. These things are going to be zero marginal cost in a decade.
Starting point is 00:34:04 that's a form of abundance. That's the aspiration of society and civilization, in my opinion. That's the great, that's why I work on AI is to make intelligence cheap and abundant. And it will be market forces that drive down the cost of that. So I think it's very cool. But I agree, super cool. And again, I don't, this is, these conversations often put me in a place where I don't want to be, which is like, now I'm going to butt the idea that there's going to be super intelligence at zero cost.
Starting point is 00:34:34 Which is, again, from a business standpoint, how does that make sense if the marginal cost is zero? Well, I mean, look, we're still going to, you know, charge a significant amount for it. I mean, we have $300 billion of revenue, like I said. I mean, this is a huge company providing great value. But the point is, where we provide value to our customers, our customers will be happy to pay us for it, right? And that means that, you know, good integrations inside of M365, great models inside of, you know, GitHub and VES code. We have co-pilot deploying on LinkedIn, co-pilot in gaming. Our consumer product is growing from strength to strength.
Starting point is 00:35:13 You know, we just crossed 100 million, wow, across all our co-pilot surfaces. So all the products are growing great. And, you know, there's plenty of revenue to be had in this transition, no question. Okay. Wow means week over week? Oh, sorry, yeah. Weekly active user. Oh, yeah.
Starting point is 00:35:31 W-A-U. I guess I'm used to Mao and Dow. But wow. Wow. Why has wow become the term of art in AI? No, sure. I think we're all using wow. Yeah. Yeah. No guess as to what happened? We've been using wow since forever. I think it shows like a more sustained engagement. Yeah. Okay. But not the daily wouldn't.
Starting point is 00:35:53 Could be mad as well. Yeah. Yeah. I don't know. It's always fun for me to figure out why the acronyms are the way they are. It'll remain a mystery. So you actually do list, so you got, and you wrote about this, you list a couple forms of intelligence that you want to pursue. And one of them was a personal companion or an AI companion for everyone. A couple questions for you to start on that one. Let me just start with one. You told me about a year ago that you think that AI will differentiate on the basis of personality. Do you still believe that? Definitely, yeah.
Starting point is 00:36:31 I mean, we are right at the very beginning of the emergence of these very differentiated personalities because all these models are going to have great expertise. They're going to have great capabilities and they'll be able to take similar actions like we've just said. But people like different personalities. They like different brands. They like different celebrities. They have different values.
Starting point is 00:36:52 And those things are very controllable now. Like we just released in co-pilot something called Real Talk last week. and it's really cool. It is truly a different experience compared to any other model. It's more philosophical, it's sassy, it's cheeky, it's got real personality. And the usage is way, way higher than the average, you know, session of a regular co-pilot. And it's built in a very, very different way, actually. So, you know, I think that's just the first foray into proper personalization.
Starting point is 00:37:28 I think we'll be able to see a lot more of that coming down. the pipe do you think it's good that people will have a new friend if you want to call it a friend that they can sort of customize in the way that they want there's been worries that you know people are like what does it mean to for real friendship then and are you going to have not normal expectations for your friends in real life yeah I think it does raise the bar and I think we have to be cautious about that because um AIs provide high quality accurate information immediately on demand. They provide high quality emotional support increasingly. And naturally as we get more used to that, that's going to put us under pressure as humans to
Starting point is 00:38:16 provide that support to other humans and provide that knowledge to other humans and be available to them to get things done. And that's going to be an interesting effect. It's going to change what it means to be human in quite a fundamental way. Like being human is going to be more about our flaws than our capabilities right but it also i mean thinking of the expectation it sets i had one entrepreneur talked to me about how well there's things you would never go to a human with right now because of norms like if you're working on a project you wouldn't like go to a colleague every five seconds and say how about this how about this how about this or what if i tweaked it this way uh but you could do that with a bot and the bot will be like oh yeah i'm happy to help you so is there any worry that that will
Starting point is 00:38:57 spill over into human relationships? What would that mean? I think that's a very interesting point. I mean, in some ways, AIs provide us with a safe space to be wrong. And, you know, it's kind of embarrassing, but we can ask the same question over and over again. And in 10 different ways. And that's how we get smarter. So I think it's a, I think, yeah, it's a, it's a, it's a, it's a good philosophical question to reflect on these kind of things because it is going to really change what it means to be human. All right, Mustafa, one final question for you. You say technology's purpose is to help advance human civilization. It should help everyone live happier, healthier lives. It should help us invent a future where humanity and our environment truly prosper.
Starting point is 00:39:51 So my question for you is, has it lived up to that promise? I think science and technology has lived up to that promise. Yeah, I think so. I think we're in an incredible place. I mean, you know, we've doubled life expectancy in 250 years. We're curing all kinds of diseases. We can communicate with one another on these devices. I think it's incredible.
Starting point is 00:40:15 There's every reason to be optimistic about technology and science and the project of progress. And I just genuinely think AIs are going to. going to provide us all with access to abundant intelligence, which is going to make us more productive and more creative. And I think we're already starting to see it. So yeah, I feel optimistic about that. All right, Mustafa. Great to see you. Thanks so much for coming on the show. Great to see you, man. Thanks for your time. See you soon. Thank you. Thank you.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.