Big Technology Podcast - Microsoft AI CEO Mustafa Suleyman: Building AI Personality, OpenAI Relationship, Data Center Demand, AGI Timeline

Episode Date: April 4, 2025

Mustafa Suleyman is the CEO of Microsoft AI and a co-founder of DeepMind. He joins Big Technology to discuss Microsoft's strategy to build more personalized and emotionally intelligent AI companions. ...Tune in to hear how Microsoft is differentiating its AI offerings through personality design, memory features, and action capabilities that could transform our digital interactions. We also cover Microsoft's data center plans, its relationship with OpenAI, and predictions about when we might reach AGI. Hit play for a fascinating look at the future of human-AI relationships and what it means for work, technology, and society. --- Enjoying Big Technology Podcast? Please rate us five stars ⭐⭐⭐⭐⭐ in your podcast app of choice. For weekly updates on the show, sign up for the pod newsletter on LinkedIn: https://www.linkedin.com/newsletters/6901970121829801984/ Want a discount for Big Technology on Substack? Here’s 40% off for the first year: https://tinyurl.com/bigtechnology Questions? Feedback? Write to: bigtechnologypodcast@gmail.com

Transcript
Discussion (0)
Starting point is 00:00:00 Microsoft has an upgraded, more personable AI bot that will remember you, help you organize your thoughts, and may even appear as an avatar. Why is it building it? And how will it get you to use it? Microsoft AI CEO Mustafa Suleiman is here and has some answers. That's coming up right after this. Welcome to Big Technology Podcast, a show for cool-headed and nuanced conversation of the tech world and beyond. We're joined today by Mustafa Suleiman.
Starting point is 00:00:27 He's the CEO of Microsoft AI and a co-founder of DeepMind. And I'm so excited for this conversation because today we're going to talk about the company's upgraded AI bot. How much AI has left to improve Microsoft's relationship with OpenAI and when we might expect to get to AGI. So no lack of topics to cover. Mustafa, it's great to see you. Welcome to the show. All the great questions. I'm super excited.
Starting point is 00:00:51 Thanks for having me on the show. And it's great to be here. Great. So let's get right into the product news. right off the bat. You have a number of product announcements you're making today as this show is coming out. And basically what this amounts to is building a more personalized companion. This is a vision you've had for a long time. But this is starting to roll out in co-pilot. This, the upgrades that we're talking about is better memory, which I think is really interesting.
Starting point is 00:01:17 So the bot is going to remember you. Actions like booking flight tickets or making a table reservation, a shopping assistant, and then, of course, you're teasing some sort of avatar play. So talk a little bit about how your vision for this more personalized co-pilot is starting to play out. Yeah, you know, the amazing thing about the time that we're in is that we're actually transitioning from the end of the first phase of this new era of intelligence into the very beginning of the next phase. And what I mean by that is that, like, over the last couple of years, we've all been blown away by the basic factual, succinct Q&A style responses that these chat bots give us. And I think that that's awesome and has been incredible. You can think of that
Starting point is 00:02:04 as I do as it's IQ. It's basic smarts. And obviously that's totally magical. And obviously, early adopters tend to be really, really focused on, is it good at math? And, you know, can it do coding really well and stuff like that. But the majority of consumers, I think, really care about its tone. They care. Is it like polite and respectful? Is it occasionally funny in the right moments? Does it remember, you know, not just my name, but how to pronounce my name? And when I correct it, does it remember that correction? And that's actually a really hard problem. And so I think these subtle details make up its emotional intelligence. And, And I think that's what we're taking small steps towards today as we launch a bunch of new
Starting point is 00:02:54 features around memory, personalization, and actions. So how long will the memory go back? Because to me, one of the more annoying things about using these bots is having to kind of tell it who I am at each time. And we know that Open AI, for instance, has some memory baked in. It will remember things and bring it into new conversations. Open AI, by the way, which we're talking at a moment where it just announced a $40 billion fundraise with Microsoft included as one of the funders. So we'll get to that in a bit. But we're
Starting point is 00:03:23 talking about these bots having memory. And how far back will your bot now go back? Well, I have to tell it like every couple months who I am. Like, I feel like I'm living in the notebook every time I'm trying to talk to one of these things. Unfortunately, it's not going to be perfect, but it is a big, big step forward. So it's going to remember all the big facts about your life. You know, you may have told it that you're married, you have kids, that you grew up in a certain place, you went to school at a certain place. And so over time, it's going to start to build this kind of richer understanding of who you are, what you care about, what kind of style you like, you know, what sort of answers you like, longer, shorter, bullets, conversational, you know, more humorous.
Starting point is 00:04:05 And so although it won't be absolutely perfect, it really will be quite a different experience. And I think it's the number one feature that I think is going to unlock a really different type of use, because every time you go to it, you'll know that the investment that you made in the last session isn't wasted, and you're actually building on it, you know, time after time. And along with memory, you're also releasing actions, things like booking a flight, going a ticket, I think ticket master, open table, reserving restaurants, space at a restaurant. And I'm curious if you think this goes hand in hand. Like, if you think the AI bot knows you well, then you're saying, okay, you can maybe
Starting point is 00:04:45 take my credit card and go book that flight. Is that the idea? Exactly. It's basically saying you know, getting access to knowledge in a succinct way is all well and good. Doing it with a tone and a style that is friendly, fun and interactive also cool. But really what we want these things to be able to do is to, like you say, buy things, book things, plan ahead, just take care of the administrative burden of life. That's like always been the dream. of certainly why I've been motivated to build these personal AIs, going back as far as I can remember 2010 when I first started DeepMind, that's really what we're going after,
Starting point is 00:05:25 is like take time and energy, you know, off your plate and give you back moments where you can do exactly what you want with more efficient action. So things like it will now be able to take control of your mouse on windows, navigate around, show you, for example, where to turn on a particular setting or to fill out a form or like you may not know how to edit a photo and it'll point out where you need to adjust the slider or where to click on a drop down menu. So it's just going to make things feel a little bit less frictionful and a little bit
Starting point is 00:06:03 easier to get through your digital life. And you're also going to release avatars at some point where we'll be able to kind of look at these things as sort of digital people? You know, I think that this is definitely going to be one of those that, you know, we would say in the UK is like Marmite, you know? Marmite is like a... You like it or you don't. You like it or you don't. And for some people, they absolutely love it. In testing, it completely transforms the experience. You know, some people love a text-based experience. They like the facts. They like to get in and out. They want to know what's what, and they're done. Some people like an image-based experience or a video-based experience. Other people really resonate when their co-pilot shows up with its own name, with
Starting point is 00:06:49 its own visual appearance, with its own expressions and style. And it feels much more like talking, you know, to you or I now, you know, its eyebrows adjust, its eyes open or close, you know, its smile changes. And so we're really just experimenting. We're actually not launching anything today, but we are showing a little bit of a hint of where we're headed. And I think it's super exciting. I genuinely think this is going to be the next platform of computing. Just as we had desktops and laptops and smartphones and wearables. And I think over time, we're going to have deep and meaningful lasting relationships with our personal AI companions. Yeah, I agree completely. It's clearly that's where this is heading. But Mustafa, everybody that's listening is going to ask
Starting point is 00:07:34 the same question, probably about this point. They're going to say, all right, Mustafa is building this at Microsoft. Amazon, we just said Panos Panay on the show, they're building it with their new AI bot. OpenAI, who you're a big supporter of, is doing the same thing. The second Sam Altman tweeted her, I think Chat Chip-T goes from 100 million users, where it's stagnated for a year to about 500 million today. And then, of course, you mentioned you're from, you started at DeepMine. Well, we know that that's what they're interested in as well. So, everyone seems to be building this. How is Microsoft going to be different? And is it that you just differentiate by the basis of your personality? Do you carve off a certain area? What's the
Starting point is 00:08:22 plan? Great question. I mean, the way I think that we're going to be different is by leaning into the personality and the tone very, very fast. We really want it to feel like you're talking to someone who you know really well, that is really friendly, that is kind and supportive, but also reflects your values, right? So if you have a certain type of expression that you prefer or, you know, a certain kind of value system, it should reflect that over time. So it feels familiar to you and friendly. At the same time, we also want it to be boundaryed and safe. We care a lot about it being, you know, just the kind of straight up simple individual. We don't really want to engage in any of the chaos here. It's really trying to keep it as simple as possible. And so the way to do that,
Starting point is 00:09:09 found is that it just stays, you know, reasonably polite and respectful, super even-handed. It helps you see both sides of an argument. It's not afraid to get into a disagreement. So we're really starting to experiment at the edges of that side of it. So is it really just making it more personal than the others? Like that's the way to differentiate? Yeah, I think so. I think like at the end of the day, we are like at the very beginning of a new era where there are going to be as many co-pilots or AI companions as there are people. There are going to be agents in the workplace that are doing work on our behalf.
Starting point is 00:09:46 And so everyone is going to be trying to build these things. And what is going to differentiate is real attention to detail, like true attention to the personality design. I've been saying for many years now, we are actually personality engineers. We're no longer just engineering pixels. We're engineering tokens that create feelings
Starting point is 00:10:06 that create lasting, meaningful relationships. And that's why we've been obsessed with the memory, the personal adaptation, the style, and really just declaring that it is an AI companion, you know, not a tool, right? A tool is something that does exactly, you know, what you intend, what you direct it to. Whereas, you know, an AI companion is going to have a much richer, more kind of emergent, dynamic, interactive style. It would change, you know, every time you interact with it, it will give a slightly different response. So I think it's going to feel quite different to pass waves of technology. It's kind of wild to think, and we're already starting to see the differentiation between the bots, but it's wild to think that you might just go shopping for your flavor of companion. I mean, the open table integration is something that we've seen across every single bot, and we've seen it for a while. I think now it's actually starting to become possible to do that and trust that your table is going to be there after you instruct the bot to do it,
Starting point is 00:11:04 and it will be a normal conversation. But it is interesting that it, is that the right way to look at it? you're picking your flavor of AI companion? Yeah, I think I think you are. You're going to pick one that has its own kind of values and style and one that kind of suits your needs and, you know, one that really adapts to you over time. And as it gets used to you, it'll start to feel, you know, like a great companion, just like your dog feels like, you know, a part of the family often. I think over time it's going to feel like a real connection. And I can kind of already see that in, you know, hearing from users. We do a lot of user research and I actually do a user interview every week with someone who
Starting point is 00:11:42 uses the product, one of our power users, and just listening to them tell stories about how it makes you feel more confident, less anxious, more supported, more able to go out and do stuff. I mean, I was chatting to a user last week who is 67 and she was out there, you know, fixing her front door which the hinge had broken and it needed repainting and every time she repainted it it was coming up with bubbles and so she phoned co-pilot had a long conversation
Starting point is 00:12:16 about how to sand it down, coat it in the right way, she ended up going to Home Depot, forgetting what paint to get called co-pilot again, had a chat about it. I mean it sounds mundane but it's actually quite profound it's actually incredible that like people are relying on co-pilot every day
Starting point is 00:12:32 to you know help them feel unblocked in her words And so I just thought that was an amazing story, and it kind of gives an insight into like how this is already happening. It's already transforming people's lives every day. Oh, it doesn't sound mundane to me at all. And in fact, who are you having those type of conversations with? If you call your friends up, it only is your best friend who you're going to call and ask about the Home Depot stuff. Maybe it's your spouse. I have a list of maybe, you know, five people I could call with those type of questions. So instantly what happens is that co-pilot, if, this is built right and we know they're getting more personable becomes, you know, in your inner circle right away. And it just reminds me, I knew, I said we're going to get a little weird when we logged on. So I think we need to talk about this. This reminds me very much of a conversation I had with the CEO of Replica who mentioned that her, she also wants to build an AI assistant. And the path to being an AI assistant is to build a companion. And a lot of people have developed feelings
Starting point is 00:13:34 for their replicas. In fact, she told me she's been invited to multiple weddings between people and their AI assistance. Now, to me, it just seems like if you're building this, you have to be ready for the fact that people are going to fall in love
Starting point is 00:13:48 literally with your product. Not just I love my iPhone. I love co-pilot. And maybe you'll get invited to weddings. Are you prepared for that? I think that's a question of how we design it. I know the replica people and I met Eugenio and I respect what they've done.
Starting point is 00:14:07 But at the same time, it's really about how you design the AI to draw boundaries around certain types of conversations. And if you don't draw those boundaries, then you essentially enable the user of the technology to, you know, let those feelings grow and really kind of go down that rabbit hole. And that's actually not something that we do and it's not something we're going to do. And in fact, you know, we have classifiers that detect for any of that kind of interaction in real time and will very respectfully and but very clearly and very firmly push back before anything like that develops. So we have a very, very low instance rate of that. And you can try it yourself when you chat to co-pilot. You know, if you try to flirt or even if you just say, oh, I love you, you'll see it tries to pivot the conversation in a really polite way without making you feel judged or anything. And I think that, you know, to your earlier question of like, what is going to differentiate the different chatbots, well, some companies are going to choose to go down different rabbit holes and, you know, others won't.
Starting point is 00:15:12 And so the craft that I'm engaged with now is to design personalities that are genuinely useful, that are super supportive, but are really disciplined and boundaryed. Yeah, I do have to say that this isn't how I anticipated to spend my weekends as a team. tech journalists trying to push the boundaries of these bots and see how much they would respond to flirtation. But it is becoming a thing. And I'm curious, like, if it comes to the point where people want to build that deeper relationship. And maybe it's not like a person-to-person relationship. Maybe it's a third type of relationship. But they really do have these deep feelings for a bot. Like, where do you draw the line? Like, are you willing to, if this is how people are going to differentiate? Are you willing to lose because you wouldn't go that route? Yeah. I mean,
Starting point is 00:15:58 I like your empathy there. I think it's important to keep it open mind and be, you know, respectful of how people want to live their life. All I can tell you is that here at Microsoft AI, we're not going to build that and will actually be quite strict about the boundaries that we do impose there. And I think you can still get the vast majority of the value out of these experiences by being, you know, just a really supportive hype man, just being there for the mundane questions of life, being there to talk to you about. that lame, boring day that you had or that frustration that you had at work, like, that is already a kind of detoxification of yourself, you know? It's like an outlet, you know, a way to kind
Starting point is 00:16:40 of vent and then show up better in the real world as a result. And I see that a lot in the user conversations that I have as well. Like people feel like they've got out what they needed to get out and they can show up as their best self with their friends and their family in the real world. Yeah, competency matters as well. Like it has to actually be able to do the things. But I guess I anticipate that every company will be able to get there eventually because the technology is improving. Now, one more question for you about this. It is interesting right now, a theme that I'm hearing in the AI world is just that the bots have been refusing too much. And you see Open AI recently with their image reveal, they refuse a little
Starting point is 00:17:22 less. They allow you to do it in the style, make an image in the style of Studio Ghibli, allow you to make images of celebrities and public figures. And it's become, I mean, it's, it is literally, it seems like it's melting their servers. They added a million users in an hour in the day that we're speaking. We're speaking on Monday this show is going out Friday. Is this going to be a race between the labs to just kind of limit their refusals? I know that Microsoft had that moment where Bing tried to take Kevin Ruse's wife away from him. And then, you know, Microsoft put the clamps down a little bit on that. But how do you find the middle ground between wanting something to be robust and personable, but also holding true to your values? Yeah, it's a great question. It's something I think
Starting point is 00:18:06 about a lot. I think that it's not a bad thing that there are refusals in the beginning. And over time, we can look at those refusals and decide, are we being too excessive? Are we going overboard? Or actually, have we got it in the right spot? Going the other way around too early on, on, you know, I think has its own challenges. And so I kind of like the fact that we've taken a pretty balanced approach because the next sort of question that we're going to be asking is, you know, how much autonomy should we give it in terms of the actions that it can take in your browser? I mean, as we're showing today, it is unbelievable to see copilot actions operate inside
Starting point is 00:18:49 of a virtual machine, browse the web, essentially independently, with a few key. check-ins where it's like, you know, gets your permission to go a step further. But the interesting question is like how many of those degrees of freedom should it be granted, right? How long could it go off and work for you independently and stuff? So, you know, I think it's healthy to be a little bit cautious here and take sensible steps rather than be sort of too, you know, gung-ho about it. At the same time, the technology is really magical. This is actually working. you know, I think that, like, in that kind of environment, we should be trying to get it out there to as many people as possible, as fast as possible. So that's the balancing act that we've got to strike. Okay, let me read one more bit of product news that are a bunch of different product announcements and then see if I get your quick reaction to this, because I definitely want to cover all the product news. Okay, you're allowing people to check their memories and interact with their memories once the bot has built this memory database, it seems like. You're also doing AI podcast.
Starting point is 00:19:54 You're launching deep research, your own version of deep research. You're doing pages to organize your notes, and you have co-pilot search. So what is this? I mean, is there a conclusive, like a comprehensive strategy here, or are these disparate updates, or is it, again, all about building that AI personality? The way to think about it is that all of those things that you mention enable you to get stuff done, right? The IQ and the EQ are really about its intelligence and its kindness. But really what people care about is like, can it edit my documents? Can it rewrite my paragraphs when I want
Starting point is 00:20:32 it to? Can it generate me a personalized podcast so the first thing in the morning it plays it exactly how I want it? Can I ask a question about, you know, my search result and interact in a conversational way based on search? All of those things sum up to bringing your basically your computer and your digital, you know, experience to life so that you can actually interact with it and it can interact proactively. I think that's the big shift that's about to happen. So far, your computer really only ever does stuff when you click a button or, you know, you type something in your keyboard. Now it's going to be proactive. It is going to offer suggestions to you. It'll proactively publish podcasts to you. It'll generate new personalized user interfaces
Starting point is 00:21:18 is that no one else has entirely unique to you. It'll show you a memory of what it knows. All those things are about it switching from reactive mode to proactive mode. And to me, that's companion mode. A companion is thoughtful. It, you know, tries to kind of pave the way for you ahead of time to smooth things over. It knows that you're, you know, taking the kids out on Saturday afternoon. You've been too busy at work.
Starting point is 00:21:44 You haven't booked anything. It suggests that you could go to the Science Museum. But then it second guesses it for itself because it knows that Science Museum's going to be jam-packed. So then it's like suggests, you know, it's just like this constant ongoing interaction that's trying to help you out. And that's why I always say, like it's on your side, in your corner, it's got your back looking out for you. On that, I mean, this is a vision that we've heard again from Microsoft, from Amazon, from Apple, for sure, from Google. No one's fully delivered it. What makes it so difficult to build?
Starting point is 00:22:15 It's hard. I mean, the world is full of open. ended edge cases, as people have found for the last 15 years in self-driving cars, you know, we're really at the very first stages of that. That's why I said to you, we haven't nailed memory. It's not perfect. We certainly haven't nailed actions. But you can start to see the first glimmers of the magic.
Starting point is 00:22:36 You remember back in the day when we first launched, when OpenAI first launched GPT3 and when at Google, we had Lambda, which I worked on when I was at Google, you know, most of the time. It was kind of garbage and it was crazy. But occasionally it produced something that was really magical. And I think that's what great product creation is all about is like locking in to the moments when it works and really focusing on increasing those moments, addressing all the errors. And I can see now, having been through this cycle a few times, that we're nearly there with memory personalisation and actions. It's really at the GBT3 stage. So it's really buggy and stuff. But when it works, it's breathtaking. It reaches out at just the right time. It shows that
Starting point is 00:23:22 it's already taking care of a bunch of things in the background. And that is just a very, very exciting step forward. Yeah, I guess if every single company is saying that this is where they're going to, they see the technology, I guess I'm willing to be patient to see it, to see it come to fruition. And we have this debate on this show all the time. Is it the models that are important or the products built on top of the existing models that are important? I believe that it's the, if you get better models, you'll get better products. We have Ron John Roy who comes on on Friday. His, well, actually, he was on Wednesday because we're flipping him this week. His belief is, it's all about the product at this point. The models are good enough.
Starting point is 00:24:06 My question to you is, you know, is this kind of at the point where the models are going to be saturated and now you're going out to build the products? You put a tweet out recently. You said something along the lines of, it's a myth that LLMs are running out against. Yet it does seem like the conventional wisdom is that they're at the point of diminishing returns at least. So take us into this model versus product debate. No way, Jose. No, we have got so much further to go.
Starting point is 00:24:37 I mean, look at, for example, you know, people sort of, what happens is people get so excited, they jump onto the next thing and they gloss over all of the hard-fought gains that happen when you're trying to optimize something which already exists. Let's take, for example, hallucinations and citations, right? You know, clearly that's got a lot better over the last two or three years, but it's not a solved problem. It's got a long way to go. And with each new model iteration, all the tricks that we're finding to improve the index of the web, the corpus that it is retrieving from, the quality of the citations, the quality of the websites we're using, the length of the document that we're using to source from, you know, there's so many details that go
Starting point is 00:25:20 into increasing the accuracy from, you know, 95% to 98% to 99 to 99.9. You know, and I think that is just a long march. People forget that that last mile is a, is a real battle. And often, um, a lot of the mass adoption comes when you actually move the needle from 99 point cent accuracy to 99.9%. I think that's kind of happened in the background in the last two or three years with dictation and voice. I've really noticed that across all the platforms, voice dictation has got so, so good. And yet that technology has been around for 15 years, right? It's just, you know, some of us used it when it was like 80% accuracy. I certainly did. But now I'm seeing like my mum was using it the other day. And I'm like, how did you learn how to do that? And she was
Starting point is 00:26:13 just like, oh, you can just press this button, da-da-da. And I was like, oh, that's kind of incredible. And I think that's just on the dictation side. On the voice conversation side, I mean, we see much, much longer, much more interesting, much deeper conversations taking place when somebody phones co-pilot. It's super fast. It feels like you're having a real world conversation. You can interrupt it almost perfectly.
Starting point is 00:26:39 And it's got real-time information in the voice as well. so it's aware of like the latest sports result or the traffic in the area or the weather and stuff like that. And, you know, a lot of people use it in their car on the way home or on the way to work or when they're washing up and they're in a hands-free moment and they just have a question. It's kind of a weird thing because it sort of lowers the barrier to entry to getting an idea out of your head.
Starting point is 00:27:02 You know, like, you, everyone, weird things occur to us during the day. We're all like, oh, I wonder about this. I wonder about that. And then you go to kind of look it up on your phone, you search it or whatever. Whereas now I think that there is a modality that I'm increasingly seeing where people just turn to their AI and be like, hey, what was the answer to that thing or how does that work? And it might be a shorter interaction, could turn into a long conversation, but the modality is enabling a different type of conversation, a different type of thought to be expressed. So I think it's like a super interesting time like that. We're really just figuring it out as we go along.
Starting point is 00:27:37 All right. So we're definitely seeing these new modalities come out voice, of course. We've, obviously, we're in the middle of a firestorm with images. But I am, okay, I guess let me ask the previous question a little bit differently. Do you think that there are diminishing returns on pre-training right now, basically scaling up the biggest possible model and then building from there? You're shaking your head, no. Maybe specifically on pre-training.
Starting point is 00:28:00 It's been a little slower than it was in the previous four orders of magnitude. But the same computation, the same flaw, or the units of calculation that go into turning data and compute into some insight into the model. That is just a different application of the compute. We're using compute at a different stage. We're either using it at post-training or we're using it at inference time where we generate lots of synthetic data to sample from. So net net, we're still spending as much on computation.
Starting point is 00:28:36 It's just that we're using it in a different part of the process. But for as far as everyone else should be concerned, aside from the technical details, we're definitely still seeing massive improvements in capabilities. And I think that's for sure going to continue. Okay, Mustafa then, can you help me understand some headlines I've been seeing about Microsoft? This is from Reuters. Probably not. I doubt it.
Starting point is 00:28:58 Well, I'm going to ask anyway, and you tell me what you think. I mean, Reuters says Microsoft pulls back from more data center leases in the U.S. and Europe. and it says Microsoft has abandoned data setter projects to use two gigawatts of electricity in the U.S. and Europe in the last six months due to an oversupply relative to its current demand. I mean, how does that make sense in context of what you just said that you are still seeing results with scaling up? So it's funny, I actually did ask our finance guy who's responsible for all these contracts on Friday morning. And I was like, dude, I read this thing in the news. Like, what's going on? I could use the extra power for our training runs. And he pointed out that, in fact, we have optioned many, many different contracts, many of which we haven't even signed. So a lot of these are actually just explorations where we're in conversations, nothing's been signed. Some of them where we've optioned, where, you know, we're taking it, you know, just to keep our options open. And we've actually made bets in other areas, you know, other parts of the world. But I can tell you, we are still consuming at an unbelievable. rate. I think that we've, like, something like 32 or 34 gigawatts of renewable energy since
Starting point is 00:30:13 2020, we've contracted and consumed. So I think we're one of the largest buyers in the world. So I don't expect that to change any time soon. So I guess the headlines that are saying that Microsoft pulled back, you would get those headlines unless you picked up every single one of your options. Is that what you're saying? That's right. That's right. Yeah. And in fact, many of them are not even options that we signed there, just conversations that we were in with certain suppliers. Okay. I mean, I guess another explanation that we've heard is that because Open AI is now working
Starting point is 00:30:44 with others with, you know, for data center capacity like Oracle, this was a sign that Microsoft basically had allocated data center capacity to Open AI. It doesn't need as much anymore. Any truth of that? No. So like, I mean, all of their inference comes through us. And so we, there's no slowdown in our, you know, relationship with them. We sell them as much as we want to offer them. And then if there's any
Starting point is 00:31:10 extra demand that they have, particularly on the, you know, on the Oracle side, they go off and consume that. But there's really no slowdown from our perspective, at least. Okay. That is clarifying. So it's always good to have these conversations. Throw the headlines out there and see what the truth is. Let's talk about your efforts. I mean, you're building your own models, but you've decided to not try to build, I guess, the biggest possible models. You're working on smaller models. I want to ask you again if there's going to be endless value to endless scale. Why not try to, you know, throw in with the big models,
Starting point is 00:31:47 especially because others are building those big models with your aforementioned scale? Totally. I mean, you know, we have a lasting long-term relationship with Open AI, which is amazing. They've been incredible partners to us, and they'll continue to supply us. us with, you know, the best IP and models in the world for many years to come. So we can rely on them to do the absolute frontier. But I think what we always see in technology is that it always costs like 10x more to build the absolute frontier. And once that has been built, all of the engineers and, you know, developers find much more efficient ways to essentially build the same
Starting point is 00:32:27 thing that's been out there, but, you know, six months later. And that's what we refer to as our kind of Pareto optimal strategy or off frontier. And we've actually seen it, you know, across the whole field over the last three years. I mean, there are folks who have trained models that perform as well as GPT3 that are a hundred times more inference efficient that cost an order of magnitude less to train. And yet they can still deliver the same predictive capability. So I expect that to happen for GPT4, GPD40, and all of the other models down the road. So, you know, we have our own internal team of developers and, you know, world experts working on building our own MAI models. And I'm very, very proud of what they're doing. They're doing a
Starting point is 00:33:13 great job. And you mentioned in terms of where this computer is going, that inference is going to be one of the places that it's going, which is basically when the model is answering versus training up the models. I want to ask you two questions about inference and they're both related to reasoning. To build these new personalized products that you're building, how important is reasoning versus just a better model? And then in terms of compute that reasoning uses, is it true that reasoning uses 100 times more compute than training? Yeah, I mean, it's a good question. I mean, the exciting thing about reasoning models is that in a way, they've learnt how to learn. They have a method largely by looking at the logical structure of code and
Starting point is 00:34:02 math and puzzles. They've sort of learned the abstract idea of logic. They can follow a path of reasoning in its most abstract way and then apply that to other settings, even if they don't obviously appear to be, you know, logical settings. So it could be like planning or booking or learning in some other setting. And that has turned out to be a very, very valuable skill. It's kind of like a meta skill or, you know, or in some sense like a metacognition because it actually now can think out loud in its own head or, you know, talk about in its own mind what it's planning to do before it goes off and does it. And just taking a beat, you know, giving it a moment to think behind the scenes.
Starting point is 00:34:52 It might take a few minutes or 10 minutes at most. Allows it to like draw on other sources. So it can look up things on the web. It can sort of follow a path of logic down one path. Realize that doesn't, you know, turn out in the best way possible. Go back up the tree, try another path and then produce an output. So it's a really fundamental part of the process. And yes, it definitely uses more computation.
Starting point is 00:35:17 100 times more? But it generally produces better answers. Do you think 100 times more is right? I mean, we're hearing that from Jensen. I mean, so I'm curious if that's your experience as somebody who's running these models. It definitely uses a lot more computation. And, you know, but I think the interesting thing is that you're not going to need to use those models all the time. You obviously need a hard problem.
Starting point is 00:35:43 ask it a tough question that requires this kind of chain of thought thinking. And, you know, many answers don't require that. And actually, you often prefer something that is fast, efficient, succinct, and instantaneous. Okay. And now we had a debate here on the show, and I'm hoping you can weigh in on this one too. I'm just throwing you all of our debates and getting answers, which is awesome. We love doing this. In terms of like how companies are thinking through the amount of money they're spending on serving these products and whether that can continue indefinitely. Let's just use this open AI image example, the image generator that they just released in chat chip PT. People are melting down their servers and they're creating anime images. But if you think about like the economic
Starting point is 00:36:32 activity generated by these images, it's quite low and it's quite expensive to serve. Or think about, for instance, me booking a ticket on, let's say, Kai. or ticket master through co-pilot instead of just going to, you know, ticket master or kayak on my own. It's a slightly better experience, but it's a very expensive experience to serve. And so those who say that this is coming to an end, this AI moment is coming to an end, basically say that this is all going to just be too expensive and not value at enough. We're going to be, you know, having chatbots book tickets, well, we can do it on the websites. We're going to be having image generators make us anime.
Starting point is 00:37:13 which does nothing but give us maybe 10 seconds of giggling, really good giggling, but 10 seconds of giggling, and then we move on. What, I mean, what do you think about that? Like, it's clearly, like, the servers are being used, but are they being used in a valuable way enough to make companies like yours keep going and building? It is a fair question. At the same time, as we've seen over and over again
Starting point is 00:37:36 in the history of technology, when something is useful, it gets cheaper and easier to use. and it spreads far and wide. And that increased adoption, because it's cheaper, has a sort of recursive effect on price because the more people use it, the more demand there is. And that then drives the cost of production down even more because of competition. And so I expect that to happen in this situation.
Starting point is 00:38:01 I think it's actually really good news for our data centers as well. You know, Microsoft is long committed to being carbon net negative by 2030, to be clean water positive by 2030 and to be a zero waste company, they're massive amazing commitments. And I think that's actually really exciting because we end up driving demand for the production of high quality renewable energy for our data centers. And that then obviously reduces the price. I think we've seen that with solar over the last 15 years,
Starting point is 00:38:34 which is like an unbelievable trajectory. So I think there's a lot of good news there, even if, as you say, you know, some of those use cases are just generating funny anime giggle picks, many of them will be doing very, very useful things in your life too. So, you know, there's always a bit of balance there. Yeah, I guess like Chris Dixon says, it can start as the next big thing will start as a game. And a lot of people laughed at these images and the way that they, you know, make you look like an anime character if you prompt them to do that. But I also saw Ethan Mollock from Wharton, prompting it to make infographics, and it handles it perfectly.
Starting point is 00:39:11 Right. Yeah, dude, I mean, like the intertubes would not be the intertubes without serious amount of cat memes, right? They make the world go around. Exactly. All right. I want to take a quick break and then talk a little bit about your relationship with OpenAI, and then maybe get your prediction on when we're going to see artificial general intelligence. We'll do that right after this. Hey, everyone, let me tell you about The Hustle Daily Show, a podcast filled with business, tech news, and original stories to keep you in the loop on what's trending. More than 2 million professionals read The Hustle's daily email for its irreverent and informative takes on business and tech news. Now, they have a daily podcast called The Hustle Daily Show,
Starting point is 00:39:52 where their team of writers break down the biggest business headlines in 15 minutes or less, and explain why you should care about them. So, search for The Hustle Daily Show and your favorite podcast app, like the one you're using right now. And we're back here on big technology podcast with Mustafa Soleiman. He is the CEO of Microsoft AI and one of Microsoft AI's big partners is Open AI. And I just can't help but think about where this partnership is going because we talked a little bit in the beginning about the assistant that you want to build. Something that knows your context, has memory of you, something that can help you get tasks done in the real world. Well,
Starting point is 00:40:33 Open AI wants to build that exact same thing. And so I'm curious, I mean, you guys have a deal, right, where they use your technology and they're supposed to feed some of their breakthroughs back to Microsoft. But at a certain point, why does it make sense for them to keep doing that if you're trying to build the same thing? Look, I mean, first of all, it's worth saying that this partnership started way back in 2019 when Microsoft had the foresight to put a billion dollars into a not-for-profit research lab. I think that's going to turn out to be one of the most impactful, most successful investments and partnerships of all time in technology.
Starting point is 00:41:11 And despite all the ups and downs, we actually have an amazing relationship with them. If you think about the fact that they are a rocket ship that has grown, you know, faster than any other technology company in living memory, delivered a product that people absolutely love, consistently delivered amazing research technology. You know, the first thing you have to do is take your hat off to them and give them maximum respect for that. At the same time, they're also still a startup, and, you know, they're busy sort of trying to figure out their, you know, product portfolio and their priorities. And, you know, whilst we have an incredibly deep partnership with them, which is going to last way through 2030 and beyond, they also have their priorities. We have our priorities. And that's just the nature of those partnerships.
Starting point is 00:41:55 They change over time, right? And as they're growing bigger and bigger, they have different priorities. And likewise, we're doing exactly the same. So I'm pretty confident that this is going to continue to be brilliant for both sides as it has been over the last five years. Okay, you said the partnership's going to last till 2030, but not if they declare that they've reached AGI. So what happens when they do that? You know, AGI is a very uncertain definition, right? Is it your definition or their definition that releases them from the contract, though?
Starting point is 00:42:26 You know, it's an interesting way to look at the world. You know, you think about it like this. if we really are on the cusp of producing something that is more valuable than all economically productive work that any human can produce, that, you know, I think one of the last things we're going to be worried about is our partnership with Open AI is going to profoundly change humanity. I think national governments will be very concerned and interested in how that plays out. And, you know, it's just going to change what it means to be human. And so I personally think that we're still a little way off from that.
Starting point is 00:43:01 I find it hard to judge. It doesn't instinctively feel to me like we're two to three years away. I know some people think that it is and I respect them deeply. Like a lot of smart people can disagree on stuff like that. I feel like we're still a good decade or so away. And when a scientist or a technologist, an entrepreneur like me says we're a decade away, that's just a hand wavy way of saying we're not really sure and it feels pretty far off. So, you know, but that's the best answer I can give.
Starting point is 00:43:27 It doesn't feel like it's imminent. And, you know, in the meantime, we're doing everything that we possibly can to build great products day-to-day. Okay. One more thing about OpenAI. Microsoft today, we're talking on Friday. So earlier this week is part of this $40 billion fundraising into Open AI. Open AI set the record for the largest VC round ever last year, $6.6 billion. This is $40 billion.
Starting point is 00:43:56 SoftBink's going to put $30 billion in. Microsoft is part of the 10 billion remaining. What do you get for the money? I think it's awesome. I mean, look, the more Open AI are successful, the more we are successful. Like, we will end up being one of the largest shareholders in the company. We have an amazing technology license from them. They, you know, use our infrastructure and our technology in terms of our Azure compute infrastructure and so on. So it's a great partnership. And, you know, in a partnership, we want to see them do the best that they can. That's why we participate in the round. Okay. And all right, so let's talk a little bit about the future of this
Starting point is 00:44:35 technology. I guess you already said you think, I was going to ask you when you think AGI is coming. You think decades away. That would actually make you a big, less optimistic than most of your counterparts, right? Demis is saying three to five years, I mean, people everywhere. I don't know. You might not think it's coming. We tend to think here. And we're probably less informed than you are that open AI might say it next year. And so we'll have to play this back if that happens. No, I didn't say decades plural. I said a decade, a decade. You know, but look, I think the truth is it's hard to judge. Like, could I imagine it happening within five years? Yeah, absolutely. It's, it is possible. The rate of progress over the last three or four
Starting point is 00:45:16 years has been electric. It's kind of unlike any other, you know, explosion of technology we've ever seen. The rate of progress is insane. Open source is on fire. They're doing incredible things. And every lab is, you know, every big company lab is investing everything that they've got in trying to make this possible. So yeah, I could certainly see a scenario where it's closer to five years. I'm just saying, you know, instinctively to me, it feels like there's still a lot of basics that we've got to get right. You know, we still have to nail hallucinations. We still have to nail those citations I mentioned. It's still not great at instruction following. It still doesn't quite do memory. It still doesn't personalize to every individual. But, you know, we're
Starting point is 00:45:57 seeing the glimmers of it doing all of those things. So I think that we're taking steady steps on the way there. Now, you were at Google for a while. You mentioned you worked on Lambda. I'm curious what you think happens. We don't even need to reach AGI for this question to come into play. What happens to search as we start to speak more with products like yours? You've mentioned in the past that you think search is horribly broken or I'm channeling your words, but something along that line. So what happens? I honestly think it's kind of amazing that we still all use search. It does feel like, you know, using a yellow pages or an A to Z back in the day, right? It, you know, I think it's going to fundamentally change. I think instead of browsing 10 blue
Starting point is 00:46:39 links, you're just going to ask your AI. It's going to give you a super succinct answer, show you images, maps, videos, all in the feed. You're going to give feedback and be like, oh, that's a bit strange. I prefer it a bit more like that. Or what does that look like? What about this? And it's just going to dynamically regenerate for you on the spot. So how does that change the business model? Well, I still think ads are going to play an enormous part of it. Hopefully, those ads are higher quality, more personalized, more useful.
Starting point is 00:47:09 There's nothing wrong with ads. We want them to be helpful to us. Like, I'm happy when I buy something that I found from an advert because it's what I really, really want. But I'm not happy when I feel like I'm getting spammed with low quality ads. And so that's the balancing act. that we've got to strike is to try and find ways to introduce ads into the, you know, the co-pilot experience in a way that's actually subtle and is really helpful to you.
Starting point is 00:47:34 Yeah, and that's really hard because let's say this is your best bud, and it's your inner circle of the five people you call when you're running out of ideas at Home Depot for it to then say, you know, I really appreciate you and I'm going to help you out here. But by the way, do you know there's a different side of glue that you might be interested in? And the finessing on that must be quite difficult. So we are running out of time. I just want to ask you one question about jobs because you're also pretty strident about the possibility that we might have some serious change here come to our work.
Starting point is 00:48:07 And you had said that AI is going to create a serious number of losers in white collar work. Maybe it already is. I've sort of changed my tune and thinking that, you know, we're all fine in the white collar work world. and now thinking, well, it's anyone's guess. So what's coming, Mustafa? I do think that that is the big story that we should be talking about. That's the transition that's going to happen over the next 15 years, is that it is going to be a cheap and basically abundant resource
Starting point is 00:48:38 to have these reasoning models that can take action in your workplace, that can orchestrate your apps and get things done for you on your desktop. Like, that really is quite a profound shift in how we work today. And I do think that, like, your day-to-day workflow just isn't going to look like this in 10 or 15 years' time. It's going to be much more about you managing your AI agent, you asking it to go do things, checking in on its quality, getting feedback, and getting into this, like, symbiotic relationship where you iterate with it and you create with it and solve with it. that's going to be massively more efficient, and I do think it's going to make everybody a lot more creative and productive. I mean, after all, it is intelligence that has produced everything that is of value in our human civilization. Like, everything around us is a product of smart human beings, getting together, organizing, creating, inventing, and producing everything that you see in your line of sight at this very moment.
Starting point is 00:49:38 And we're now about to make that very same technique, those set of capabilities, really change. cheap, if not like zero marginal cost. And so, you know, I think everyone gets a little bit caught up on the week-to-week day-to-day or definitions of these abstract ideas. Just focus on the capabilities, you know, it should really be thinking about these things as artificial capable intelligence. What can it do in practice and what is the value of that doing? I prefer that as a framing versus AGI because it's sort of more measurable and we can actually
Starting point is 00:50:12 look at it very, very explicitly in terms of. of its economic impact and its impact on work. I mean, you could argue that that's already here. And so just to sort of ask you one, follow up on that one, what would you tell young people to do today? Because, all right, I'm thinking customer service, probably not. Software engineering. I don't know.
Starting point is 00:50:32 I just wrote a story saying, you know, they can start to do the work of journalists. I mean, you've just released podcast five minutes ago. So what should young people do when they're thinking about a career? like saying what should young people do when they get access to the internet for the first time? Like part of it is sort of obvious where it's like use it, experiment, try stuff out, do crazy things, make mistakes, get it wrong. And, you know, part of it is like, well, I actually don't really know until people get a chance to really play with it. As we've seen over and over
Starting point is 00:51:05 in history of technology, you know, the things that people choose to do with their phones, with internet, with their laptops, you know, with the tools that they have are always like mind-blowing. They're always way more inventive and surprising than anything you could possibly think of ahead of time. And so then as you start to see people use it in a certain way, then, you know, as designers and creators of technology, we adapt what we put out there and try to make it more useful to those people. So I think the same applies to a 15-year-old who's, you know, a high school thinking about what they do next in college or whatever or whether or not they go. to college. And I think the answer is play with these things, try them out, keep an open mind,
Starting point is 00:51:46 try everything that you possibly can with these models. And then you'll start to see their weaknesses as well, by the way. And you'll start to chip away at the hype that I give because I'm super excited about it. I'm obviously a super optimistic, you know, techno person. But you'll see where it doesn't work. And you'll see its edges and where it makes mistakes and stuff like that. And I think that will give people a lot more concrete reassurance as to what trajectory of improvement we're on. All right. I just want to ask one last question just to wrap up everything we've talked about today. It's kind of an offbeat question, but I am curious, now that you're talking about how these bots are going to differentiate themselves based off the personality, we are
Starting point is 00:52:22 going to have advertising in them, but they might intermediate your interactions with other companies. What happens to brand in this new era? I think brand is actually more important than never in a way because there's sort of two modes of trust. There's trust based on utility where it's functionally correct. It's, you know, factually accurate. It does the thing that you've intended it to do and therefore you trust it to do the same thing again. But then there's also a kind of emotional trust where you trust it because it is polite and respectful, because it's funny, because it's familiar, you know, and that's really where brands come in, you know, trusted brands that are able to repeatedly deliver a reassuring message. I think people are
Starting point is 00:53:05 going to appreciate that more than more than ever before. Good stuff. Mustafa, this is the first interview we've done with a Microsoft AI executive. I hope not the last, anyone who's listening on the Microsoft team. Let's do this again. And Mustafa, I'm just so grateful to have your time today. Thank you so much for coming on the show. Thanks a lot. It's been really fun, really, really cool questions. Thank you. Awesome stuff. Well, thank you everybody for listening and we'll see you next time on Big Technology Podcast.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.