Big Technology Podcast - Microsoft's Head of Cloud & AI on the AI Buildout's Risks and ROI — With Scott Guthrie

Episode Date: October 1, 2025

Scott Guthrie is the executive vice president of Cloud and AI at Microsoft. Guthrie joins Big Technology Podcast to discuss the tech industry's massive AI infrastructure buildout and whether it is ove...rdoing it with the hundreds of billions of investment. Guthrie discusses the way Microsoft thinks about its OpenAI investment, whether it's worth investing in scaling pre-training, and Silicon Valley's growing debt problem Tune in for the second half where we discuss the longevity of the GPU, custom silicon, and competing with NVIDIA. --- Enjoying Big Technology Podcast? Please rate us five stars ⭐⭐⭐⭐⭐ in your podcast app of choice. Want a discount for Big Technology on Substack + Discord? Here’s 25% off for the first year: https://www.bigtechnology.com/subscribe?coupon=0843016b Questions? Feedback? Write to: bigtechnologypodcast@gmail.com

Transcript
Discussion (0)
Starting point is 00:00:00 Microsoft's head of cloud and AI joins us as we ask. Is this AI buildout going too far? That's coming up right after this. Welcome to Big Technology Podcast, a show for cool-headed and nuanced conversation of the tech world and beyond. We're joined today by Scott Guthrie. He's the head of cloud and AI at Microsoft, and he is the perfect guest to give us some context on the massive and some would say insane buildout of AI data centers taking place today. What does it mean? Is it going too far?
Starting point is 00:00:29 what will lead to. Scott, I'm so thrilled to have you on the show. Welcome. It's good to be here, Alex. Thanks for having me. All right, let me take you through the headlines over the past couple weeks. It's crazy that this has just been over the past few weeks, but here we go. NVIDIA agreed or announced that it would invest up to $100 billion in Open AI, starting out with $10 billion. Oracle announced it would invest $30 billion in Open AI or a $30 billion build out with the company, Anthropic raised $13 billion. You know, we're just talking about a cool 143 billion, no big deal.
Starting point is 00:01:03 Is this crazy? Is this overinvestment? Well, I think there's a great question. I'm sure that's top of mind for everyone. I think, you know, stepping back for a moment, I would say, if you look at AI and the impact, I think it's going to have in the economy, it's going to be, I think, the most profound technology shift in our lifetimes. And so I think if you look at the long.
Starting point is 00:01:29 term trend, I don't see, I don't worry about overinvesting. I think there will be a question on the horizon of different companies or making different strategies in terms of their investment and how they get the return in the one, two, three year horizon. So, you know, am I going to say that every company is perfectly timed? I don't, I'm not going to make that assertion. But at the same time, I do think the long term secular trend of AI is going to be that we're going to need more infrastructure. There's going to be more ROI. from it, and it's going to be more widely used. And so, you know, I think directionally from an industry perspective, the investments do
Starting point is 00:02:07 make sense and will ultimately yield pretty profound results. So you think that this level of buildout is healthy? I think we definitely are not nearly at the point at which there is too much AI infrastructure given, I think, the number of AI workloads that are coming for the world. And I think we're seeing over the last couple years, as people use AI, they get value, they use it more, the models get better, and people then use it even more for new use cases. And I think at this point, across the industry with AI, we're still more supply constrained than we are demand constrained. And I think, you know, I expect that to continue over the next couple of years as the technology continues to evolve and as people start to integrate AI into more and more workflows. Okay. So let me put it bluntly then. Microsoft has a partnership with OpenAI, has invested $13 billion thereabouts, has the capacity to build big data centers.
Starting point is 00:03:15 why did Microsoft make the decision not to do the $100 billion level build out with OpenAI or even the $30 billion that the company is doing with Oracle and leave it to other partners to do that? Well, we have a great partnership with Open AI and it's gone back many, many years and continues going forward. And we are building out and doing a lot of projects with Open AI. Across the Microsoft Cloud, we're building out AI data centers all over the world. And at the same time, you know, we are balancing, you know, our investment to make sure that it maximizes the AI infrastructure for both our first party Microsoft offerings that we're investing in our customers, AI offerings that we're investing in. and obviously opening eyes offerings that we're deeply enabling. So, you know, we are very invested.
Starting point is 00:04:19 I don't think it's a binary. Are we building out for opening or not? We definitely are building out for OpenEI. And at the same time, the way our partnership works is we're supportive of others participating in that as well. Okay, but I just want to put a fine point on it because, again, like if you believe that this technology is going to be massive. transformative, which you stated, and that we're not at the, at the sort of optimal point of the AI buildout yet, that there's room to continue to do more. Again, it's the partnership with what is the consensus leader in the space.
Starting point is 00:04:58 They needed more infrastructure. There must have been some calculation within your group or your company to say, is it worth it for us to be the one that goes out and builds this massive, massive footprint, you know, in partnership with them or somebody else. So I definitely understand there's multiple stakeholders, but what made Microsoft pause on that front? Well, we have a balanced view. And so it's, and we take a long-term view in terms of making sure that we're building out in all the locations that we want to build out that were being thoughtful in terms of kind of the investment spend and the infrastructure that we're building.
Starting point is 00:05:41 and also recognize that we don't have to do at all. And so I think we're always trying to kind of take a continually balanced view of that. And as you've seen from our GAPX and as you've seen from our earnings calls, we are investing a lot in infrastructure and building out like crazy. But again, at the same time, you know, we're always constantly reevaluating and watching closely, you know, which data centers in which markets to what specifications and making sure that we keep, you know, good discipline, as we're doing it, that optimizes for both the long-term, near-term, and mid-term horizons.
Starting point is 00:06:20 Okay. I'll just ask one more follow-up, and then we can move on. Good discipline. What about this would have been undisciplined to have gone to this level? Well, I don't think it's so much the volume level. I think it's one of the things that we do when we add new data center capacity or AI, infrastructure is, you know, making sure that, that we can use this infrastructure for a variety of different AI use cases. I think one of the things that's really going to differentiate
Starting point is 00:06:52 AI infrastructure companies in the future is that ability to kind of maximize yield on the infrastructure. Like, how are you driving down the cost of, you know, tokens per watt per dollar? And, you know, part of what makes the Microsoft portfolio so unique is the fact that, you know, that we have a lot of our own AI products, Microsoft 365 copilot, GitHub copilot, the work that we're doing with Nuance and Dragon and healthcare. We've got the world's largest consumer application with ChatGPT that runs on top of Azure,
Starting point is 00:07:30 and we have thousands, hundreds of thousands, and millions of businesses that are also building their own AI applications on top of us. And so as we think about, like, what market are we going to build a new data center? Is it for training? Is it for inferencing? And, you know, how do we make sure that that infrastructure is going to be
Starting point is 00:07:52 maximally used? You know, we feed in kind of each of these different customer scenarios into our calculus. And there are certain tranches of capacity that we're happy to build out because we can see very clear line of sight in terms of how we're going to maximize the usage and the revenue from it. And there's others that were maybe less. likely to see the immediate or the ROI that we'd like. And so we try to be disciplined about it.
Starting point is 00:08:19 As we've kind of shared in our blog post, you know, we do kind of look at every request first and we do have an opportunity on that. And as you've seen from our CAPEX, we are swinging in a lot of opportunities. But that doesn't, you know, we're not going to be undisciplined and say, blanket, we're going to do everything. You know, we know that, you know, some opportunities will have. more certain returns than others. And we're trying to make sure that we maximize our focus around those.
Starting point is 00:08:47 You know, as we're talking, I'm kind of laughing at myself because, you know, my question is basically boiling down to you've, like, talked about your CAPEX. Isn't Microsoft expected to spend like $80 billion on infrastructure this year in the neighborhood? I think that's what we shared in our last earn. And I'm like, well, why aren't you doing another $100 billion? in but but the fact that you're not as actually very interesting and and it goes to a point that you that you just made and I'm trying to read behind the line between the lines and you tell me if I'm if I'm getting this right you think about you're talking about where you invest and where
Starting point is 00:09:23 you're pretty sure you're going to get an ROI and to me I'm sitting in your shoes the question I would be asking is is it worth spending you know all that money on training where there's been a lot of noise about diminishing returns of training larger models with these unbelievably massive data centers. Now you have the startups like OpenAI and Anthropic, their belief in the scaling law seems unabated, and so the numbers get bigger and bigger, and they seem to believe that they'll continue to get an exponential return from training these bigger models. but is your decision in terms of being disciplined based on a belief that you're not,
Starting point is 00:10:08 you're not sure if scaling up will continue to work and therefore it's too big of a risk to make such a large bet on training an even bigger model in a bigger data center? Well, I think there's a couple different elements of that. I think one is recognizing that you want to have the best models. So, you know, training is super important because if you don't have the best models, then your actual ability to monetize AI goes down. You know, at, you know, part of what makes our partnership with OpenEI unique is the fact that we do have access to the best models,
Starting point is 00:10:46 frankly, whether they're trained on our infrastructure or anywhere else. You know, that's part of our partnership that's really important. And I also think when you think about training, training is evolving for maybe where simplistically we think of training a couple years ago of you do training in one place and then you do inferencing where you are executing the models and building applications.
Starting point is 00:11:13 You know, there's now multiple types of training. There's pre-training, there's post-training, there's reinforcement learning, there's fine-tuning. There's a lot of new techniques that both sometimes require lots of contiguous infrastructure and sometimes requires lots of infrastructure, but sometimes it's smaller sizes that can be used for very specific tasks. And so when we think about the investments of our infrastructure,
Starting point is 00:11:38 we're trying to think about all of this and compose it all end-to-end. You know, for us, that means, for example, we want to make sure that we have lots of inferencing capacity because that ultimately is how customers pay us and how ultimately you make money from any AI product that you build. And I think increasingly on the inferencing side, you know, one important element is the geopolitics of the world have gotten complicated over the last many years. And, you know, customers in Europe want to make sure that their AI is in Europe. And the customers in Asia are going to care about their AI in Asia.
Starting point is 00:12:17 Obviously, the customers in North America and the United States are going to care about their AI being delivered in North America. And so even as we build out our infrastructure, we want to think about it not just narrowly as we want to have one giant pool all in the U.S., we need to kind of be distributed around the world to kind of meet those geopolitical needs and to make sure that our AI is as close to the customers that are going to be using the AI as possible and can meet all of the data residency and data sovereignty needs. And so even if you look at our infrastructure builds around the world, we have regions in more countries, in more locations than any other.
Starting point is 00:12:56 structure provider. And again, as we balance out the investments we're making on AI info, you know, we're trying to keep that in mind versus narrowly put it all in one location. I totally understand that, but I have to go back to the diminishing returns of training question. Where do you stand on that? Well, I think if you look at training broadly, I think you're going to continue to see more value from the models by doing more training.
Starting point is 00:13:25 By kind of going back to my answer earlier, I don't know if that's always going to be pre-training. I think increasingly lots of post-training activities are going to significantly change the value of the model. And so by post-training, I mean take the base model and how do you add financial data or health care data or something that's very specific to an application or a use case. What's nice about post-training is that you don't have to do it in one large data center in one location. And so part of the technique that we've been focused on is how do we take this inferencing capacity around the world? And a lot of it is idle at night as people go to sleep. You know, how are we doing increasingly post-training in a distributed fashion across many, many different sites?
Starting point is 00:14:12 And then when employees come to work in the morning, we serve the applications. And so having that kind of flexibility and being able to dynamically schedule your AI infrastructure so that you're maximized. revenue generation and training, ideally in a very swappable dynamic way, I think is one of the things we're investing in heavily, and I think is one of the differentiators for Microsoft. Okay, but you'll forgive me for going back to this scaling pre-training question. I'm just trying to see what you believe here. And you haven't said it outright, but from your answers, it does seem to me like you believe that spending wildly on scaling pre-training is a bad bed?
Starting point is 00:14:56 I wouldn't necessarily say that. I think we've definitely seen as the scale infrastructure for pre-training has gotten bigger. We are seeing the models continually improve. And we're investing in those types of pre-training sites and infrastructure. We recently, for example, announced our Fairwater data, you know, center regions around the U.S. We have multiple Fairwaters. And, you know, we did a blog post recently of one of our new sites in Wisconsin. And these are, you know, hundreds of megawatts, hundreds of thousands of the latest G.B. 200s and G.B. 300 GPUs. And our, you know, we think
Starting point is 00:15:39 the largest contiguous block of GPUs anywhere in the world in one giant training infrastructure that can be used for pre-training. And so we're investing heavily in that, as you can see kind of from the photos from the sky in terms of massive infrastructure. And, you know, we do continue to see the scaling laws improve. Now, will the scaling laws improve linearly? Will they improve at the rate that they have? I think that is a question that everyone right now in the AI space is still trying to calculate.
Starting point is 00:16:15 But do I think they'll improve? Yes. And the question is really around what's the rate of improvement on pre-training. And I do think with post-training, we're going to continue to see dramatic improvements. And that's, again, why we're trying to make sure we have a balanced investment, both on pre-training and post-training infrastructure. And, yeah, and just to parse your words here, you can see improvement by making, by doubling the data center. But that's why I used the word bet, because are you going to get the same return if it doesn't improve exponentially and just improve? on the margins and that I think is the big question right now right it's a big question and and
Starting point is 00:16:52 you know the thing that also makes it the big question is it it's not like a law of nature that's immovable and so there could be one breakthrough that actually changes the scaling laws for better and and there could be a lack of breakthroughs that means again things will still improve but do they improve at the same rate that they historically did from a raw size and scale perspective. And that is the trillion-dollar questions. Okay, great. I do want to get to the ROI of AI spend in a moment.
Starting point is 00:17:31 You know, it's always great to have a chance to speak with someone who's in a position like you are within Microsoft because we get a chance to like take some headlines and which might paint a portion of a story and then ask you what the truth is. there were some stories over the past year talking about Microsoft had like canceled options to build data centers in certain locations and people took those headlines and they read into it that there was no demand for AI or that it wasn't going as well as Microsoft's telling us but what is the what is the reason for why those those data centers that there were the options and they were canceled what happened there well we're
Starting point is 00:18:13 constantly, I think in general, the headlines were focused on things that we canceled as opposed to all the things we signed. And so if you look at a given- It's amazing how news works that way, right? If it bleat it leads, so. If you look at kind of the overall investments, and certainly if you look at the overall cap-ex spend, it has been going up and up and up. And so as has, again, the revenue that comes from it. And so I think I would kind of focus on the overall picture. as opposed to individual tranches or individual projects that we potentially made decisions on. Now, you know, the thing that we did do and we continually do is look hard at every single investment
Starting point is 00:18:55 decision we make. We don't take this level of investment and this level of project and infrastructure lightly. It's critical that we invest wisely. It's critical if we invest that we make it successful and that we bring it to market on time with the right quality and the right security. And it's critical that we have the right go-to-market to monetize it. And so, you know, part of our calculus that we do as a leadership team is constantly looking at the variables for all of those.
Starting point is 00:19:24 And there are places and times when we slow down or pause projects, and there are times when we accelerate projects somewhere else. And kind of going back to my comment around the world, you know, also the, you know, regulation, geopolitics, of how AI is going to be used going forward has changed quite a bit. And what Europe thinks about where GPUs can be based has evolved quite a bit, I'd say, in the last 12 to 18 months. And I think it's going to continue to evolve around the world.
Starting point is 00:19:57 And so even as we think about the investments we're making, we're also being very, very thoughtful in terms of where geography-based are we investing so that we can, again, maximize the AI tokens we can serve in real production applications, and then ultimately use that maximization to ensure that we're delivering a good return on investment for every capital dollar we spend. Okay, and I have some technology questions for you, but just to keep on speaking about the financing of this stuff because it's so important. So there has been some interesting reporting about how the,
Starting point is 00:20:38 AI infrastructure build out has begun to be funded by debt, not just profits. Great story in the Wall Street Journal this week. It says debt is fueling the next wave of the AI boom. I'll read the beginning. In the initial years of the AI boom companies, comparisons to the dot-com bubble didn't make sense. Three years in growing level of debt are making them ring truer. Early on, wealthy tech companies were opening their wallets to out joust each other.
Starting point is 00:21:05 For leadership in AI, they were spending cash generated, largely from advertising. and cloud computing businesses. There was no debt-fueled splurge on computing and networking infrastructure like the one that inflated the bubble two and a half decades ago. However, that is starting to happen. Now, Open AI's deal with Oracle
Starting point is 00:21:24 has been pushed Oracle to start taking on debt. They say this is according to the story. Analysts at KeyBank Capital Markets estimated in a recent note that Oracle would have to borrow $25 billion a year over the next four years. Obviously, you guys are not Oracle, but you know, you're watching this happen as it plays out and see the parallels to the dot-com boom. I'm sure it's not fun. You've been at Microsoft for I think 27 years, 28 years, 28.
Starting point is 00:21:58 Sorry, I don't want to miss that last year there. So you've seen it, Scott. This seems to be an issue, at least from the outside. what do you think about it being on the inside? Well, I think obviously there's a tremendous amount of spend from lots of different companies. And I would say, yeah, the thing I can speak most to is what we're doing. And kind of, you know, per my comments earlier, I think we're trying to make sure that we have a smart investment play. and a long-term strategic play that allows us to ride the AI revolution
Starting point is 00:22:42 that we think is going to transform the world and do it in a way that leverages some of the strengths that we have at Microsoft, which is we have very good cash flow, we have a very diverse portfolio of businesses, in particular in the commercial enterprise space, whether it's cloud infrastructure, productivity applications, business applications, security, et cetera. All of them are going to be transformed by AI.
Starting point is 00:23:08 And, you know, if you look at, say, to your comment earlier on the Wall Street Journal post, I think if you even read further in the post, you know, it does show the ratios for different companies. And there are some companies that are 400 percent debt-to-equity ratios. And there are other companies that are much smaller, and that would be Microsoft. And I think, you know, we want to make sure that we're not, and I, I think, again, based on our CAPEX spend and the rate at which our CAPEX spend is going up, you know, we're not going to sit on the sidelines and not be bold as we invest. And at the same time, you know, I think the thing that our investors expect, and ultimately
Starting point is 00:23:47 I think every investor of every company will expect, is to see that revenue growing in terms of AI services and products that are being delivered in terms of net revenue recognized in a quarter and making sure that the proportionality of that to the spend and in particular to the obligations that maybe are being undertaken with debt are balanced. And that's the thing that we've been focused on. I think, you know, if you look at our last quarterly earnings, I think people were pretty pleased with getting the balance right there. And, you know, every quarter going forward, people are obviously going to be looking at
Starting point is 00:24:27 making sure that that balance is right so that they see us investing for the long term and going to win. And at the same time, doing it in a way that is sustainable and allows us to kind of ride through, you know, the ups and downs that inevitably will happen over the next many years as this technology, you know, transforms the world. What are the consequences if this goes wrong with the debt? You're obviously not taking on the same amount of debt. So there's a rationale behind it. What happens if it breaks? Well, I mean, we have the ability. We're not constrained.
Starting point is 00:25:08 I mean, our borrowing costs ironically right now are low. Yeah, but industry-wide, big picture industry-wide, not Microsoft specifically. Well, I think the thing that we, as an industry, I think, you know, again, you need to have that thesis of how you're going to use the infrastructure. And is it, do you have, I would focus less on the megawatts that sometimes get reported in the press. And more at where are those megawatts? And what are you gonna do with those megawatts? Is it gonna be ultimately capacity
Starting point is 00:25:41 that you can use to serve customers? Is it to build better models that help you serve customers? And you know, what is the line of sight in terms of the product services and revenue that comes from it? And I think that's a place where, again, Between Chatchibouti, which is the number one AI app in the world, between Microsoft 365, which is the number one enterprise AI app in the world,
Starting point is 00:26:04 and between GitHub, which is the number one developer AI app in the world, I feel good that we have applications using our infrastructure and maximizing it, and I feel good about the investments we're making in terms of capital spend and build out in the right locations to kind of continue to do that. And I think not every company probably has that level of, game plan and I don't think that maybe not every company is probably doing the same level of thoughtfulness of that and you know at some point you know different companies will probably be hit by it but you know we're very focused on what we do and how do we make sure that we stay
Starting point is 00:26:43 aggressive yet disciplined and make sure that we get that balance right all right i want to take a quick break and then i'm going to ask you a couple uh technology questions about the state of the buildout GPUs, custom silicon, and then maybe we can get a little bit into this ROI question. In fact, we will. We have to talk about the ROI of AI. We'll do that right after this. And we're back here on Big Technology podcast with Scott Guthrie, the head of cloud and AI at Microsoft. Scott, we have a Discord here at Big Technology, and I asked some of our members, you know,
Starting point is 00:27:16 what they would ask you. And we got a flood of excellent questions. And I think they were great because they focused on the technology. Some questions that I don't think you hear too often in the common conversation about this technology. So you're the perfect person to ask. I'm going to ask them to you. One of our members asked, what is the working life of a GPU and how long until they burn out? Are there use cases for GPUs once they are no longer top of the market?
Starting point is 00:27:45 We hear often about, well, unlike the laying of the fiber, the GPU depreciates after a couple of years. So I think this is a pretty important question. Can you handle that for us? Yeah, I think kind of going back to the comments that we had earlier on balance, I think as you think about your GPU build-out, one of the things that we think about
Starting point is 00:28:07 is the lifetime of the GPU and how we use it. I think, you know, what you use it for in your one or two might be very different than how you use it in your three, four, and five or six. And so, you know, I think that is something where, you know, so far we've always been able to use our GPUs, even ones that we deployed multiple years ago for different use cases and get positive ROI from it. And that's why our depreciation cycle for GPUs is what it is. But I do think that's, you know, as we build our infrastructure, we are definitely consciously thinking about that because you don't want to have your entire fleet in two years. suddenly have to be replaced because, you know, that that would be expensive. And so, you know,
Starting point is 00:28:55 we are very thoughtful on that. And again, I think I talked earlier about different training. I also think even as you think about training, we often, in the past, used to monolithically call training training training, there's lots of lots of different training use cases now. There's pre-training, there's synthetic data generation that goes into training. There's post-training with RL and fine-tuning. and other different techniques. And, you know, having infrastructure that's very fungible and that you can use for a variety of different training scenarios and at the same time be used for inferencing
Starting point is 00:29:33 where you ultimately enable an application to perform a query or perform an AI invocation is key. And I think that goes beyond just the GPS, Even though people often narrowly focus on that. It also is around the data center Architecture. It's around the storage and the compute that's near the GPUs and it also really comes to play with a network Because if you are for example building one large data center that only does training and it's not connected to a wide area network around the world that's close to the users It's hard to use that same infrastructure for inferencing because you can't go faster than the speed of light.
Starting point is 00:30:20 And so someone elsewhere around the world that wants to call that GPU, if you don't have the network to support it, you can't use it for those inferencing needs. And so again, going back to some of my comments earlier about how we're trying to be very thoughtful about where we place infrastructure and how we maximize the utilization,
Starting point is 00:30:40 we're definitely thinking of that, not just for this year or this quarter, but thinking about it on that four or five, or six-year horizon for how we want to basically leverage and use it. Okay. Here's another question. Are there any cool technological breakthroughs that would change the economies of data centers as we know them now?
Starting point is 00:31:00 GPU started as graphics processing units for video games. Are the resources you found that might do as well, but with fewer constraints? Well, I think one of the biggest changes that's happening right now from a data center perspective, and you're seeing this with the latest Nvidia GPUs, and I think you're going to see this in a more profound way over the next two years, is the shift from air-cooled data centers where you use, you know, effectively giant air-conditioning units or chillers, to a
Starting point is 00:31:36 liquid-liquid-cooled facility water, where you're actually pumping in water in order to cool the equipment in a closed loop circulating system. So in other words, you feed in cold water, you run it over the GPUs effectively, extract the water, cool it down again and then do it again throughout the building. That's a massive technology change. And it does mean that older data centers that are air cooled, you know, they can't just drop in liquid cooling and be effective. And so that is something that I think everyone that's in the AI space is designing for. It needs to be thoughtful of, again, with their infrastructure projects to make sure that they're ready for that technology shift.
Starting point is 00:32:18 It also is going to have a big difference and big impact in terms of the staffing. When you have an air-cooled data center, you'd have very few employees often per server. When you all start to involve water and liquid, you know, it's not massively more, but at the same time, it does change staffing because there are more things that break when you have pipes that are actually continuously flowing liquid into. a data center. So there's a lot of technology shifts that are happening right now, behind the scenes, beyond the GPUs. And then obviously, GPUs are the things that dominate the press in terms of innovations, both in terms of the silicon, but also in terms of the network. Because
Starting point is 00:33:00 at the end of the day, if you have a chip that can process a lot more information, but you don't have the ability to get that information to the chip and extract it or have it communicate with other chips, you know, then you don't get the yield out of it. And so I think it's fascinating right now in technology, the pace at which so many things are evolving so fast, both with the GPUs to the question, but then also the data centers, even the power and cooling infrastructure for the data centers and the network. And, you know, as a technologist, it's exciting times. Right. You mentioned staffing, so I want to ask a follow up on that front. I think for those who don't live, those who are not in this, deep into it, there is a perception that
Starting point is 00:33:48 data centers, they're placed near communities in some cases. They use up a lot of water. They don't provide a lot of jobs. Is that a misconception? I think it's a misconception. You give us some numbers to sort of flesh out what they actually bring to a community that they appear next to? Yeah. I mean, we've, we've talked. talked about with our Wisconsin Fairwater site that we did some press on recently and talked about, including with the governor of Wisconsin and others that were attending. You know, it's thousands of jobs that we've created on the construction of the site. I think we've shared over 3,000 jobs. And these are very skilled jobs. These are, you're talking about electricians. You're
Starting point is 00:34:40 talking about plumbers, you're talking about welders, you're talking about skilled tradescraft and, you know, high quality jobs. And I think, you know, if you look, we have a phenomenal work site, phenomenal workers there, and a phenomenal safety culture, which has allowed us to attract some of the best workers to work on that project. I think what people are missing sometimes when they say, okay, but when the project's done, how many people are going to be in the data center. And there will be, you know, hundreds of people that will be in the data center. What people are missing is the fact that right next to that data center, we're building another data center. And so those thousands of people that have been working on the first Fairwater
Starting point is 00:35:23 Data Center, we just announced are now going to be starting work on the second one. And then after the second one, we'll do a third one. And if you look at the land and you look at the power, we've accumulated in that area, it's multi-gigawatts of land, or multi-gawatts of power. And it's It's an awful lot of land. And so you're going to see us continue to employ thousands of very skilled tradescraft workers in that community. And as each one of those data centers comes online, we're going to add net new employees that will actually operate it and manage it.
Starting point is 00:35:58 So, you know, that would be an example, I think, of a community. And we, you know, we have over 400 data centers around the world. So they're not all that size. obviously, but replicating that and I think as more infrastructure gets built out, you're going to continue to see not just jobs created, but well-paying jobs that really require real tradecraft. Do you feel what, do you feel the, I don't know the right word to put it, the pressure of competing with China? Because from my understanding, China has a much looser regulation. approval process and they're just stacking, you know, data centers.
Starting point is 00:36:42 They have abundant electricity in the United States, in particular. I imagine Europe is the same way. That is not the case. So what's it like? Certainly, I think the world has a very different regulatory approval process. I mean, I think one thing that when I talk to people and they say, how can you build data centers faster, you know, there's obviously things that we can do from a technology and are doing from a technology and from a manufacturing perspective.
Starting point is 00:37:11 But, you know, candidly, you're in the U.S., the longest part of building a data center is getting permitting. It's not actually the construction. It's making sure that you, you know, get permitting approval for all the steps that you want to take. And, you know, different states and different parts of the country have different regulatory environments. And I think even if you look at a sort of a heat map,
Starting point is 00:37:34 if you will, of where data centers are being built in the U.S., You definitely see pockets, and I would say some of that approximates to where there is land and where there's power. And some of it really, you know, closely correlates with where it is easier or faster to kind of complete the permitting process. You know, in Wisconsin, we had a, you know, a phenomenal partnership with the governor and the local county. We were able to purchase some land and power that a manufacturer was previously going to use, and they pulled out of a project. And so, you know, I think that the local communities recognized if they weren't able to work with us, you know, they were going to lose jobs and have, you know, impact on the community. And they leaned in with us and, you know, can't say enough positives in terms of the speed with which, you know, we went through all the process. We got all the approvals.
Starting point is 00:38:27 It was a very thorough process. But it was streamlined so that we could move fast and that we could actually help ensure the jobs weren't lost and that instead they were created in the community. And I think there's more opportunities for public-private partnership like that that we'd certainly welcome as part of it. Okay. Another Discord question. What timeframes are you looking at to get ROI on these investments? So when would you stop investing if prior investments weren't showing returns? And how do you know when it's time to stop building? Well, I think, you know, at the end of the day, I think, you know, every quarter we share our revenue growth and we share our cap-ex spend. And, you know, to some extent, I think the, you know, markets keep companies honest in terms of that balance. And, you know, sometimes markets can be slightly irrational at times, but in the limit, the markets keep you honest.
Starting point is 00:39:25 And, you know, that's a big part of why we focus so much on making sure we get that balance right. Make sure we're, again, investing for the long run. And I don't think anyone, if you look at our CAPEX spend and our commitments and our investments would say that we are not being bold. But at the same time, you know, we have a report card every quarter where we need to kind of demonstrate and prove not just with press releases, but, you know, here's how much revenue growth we had. You know, last quarter we grew Azure 39% year over year on a very large number. And a lot of that was driven by AI and then also driven by the other systems that come with AI because there are databases and there's compute and the storage sold with that AI. And, you know, I think investors were happy with both the spend and the aggressiveness that we
Starting point is 00:40:20 were building out, but also the return. And, you know, I think that's going to be true, you know, forever. And, you know, making sure you get that balance right. And again, as part of that balance, markets want to know you're investing to win the long run. And at the same time, they want to make sure you have some level of discipline. And I think our portfolio, the balance that we have both across the products we build, but then also the fact that we have the largest AI product in the world called JetGBT, running on top of our cloud, gives us a unique opportunity to get that balance and that growth and that investment right.
Starting point is 00:40:57 Is Chat Chepti, by the way, going to stay? on Azure, even though Open AI is making these partnerships with Nvidia and Oracle? Yes. Okay, all right, it's good to get something definitive on that. You mentioned your 39% Azure growth, and I'm looking at your quarterly numbers every quarter and often talking about them on CNBC
Starting point is 00:41:21 and the numbers are massive. And the other side of it though is, so that's spend coming from clients, right? And there have been multiple studies that have come out recently that I've talked about how enterprises aren't getting the ROI that they've anticipated on their AI projects yet. When you see those studies, do they ring true to you? How do you react to them? Well, I think when you say AI in general, it's a very broad statement. This is generally, this is obviously, I mean, not obviously. It's in large
Starting point is 00:41:56 part, this is generative AI where companies everywhere have. try to adopt LLMs and try to put some version of that into play in there. And it's not recommender engines, basically. Yeah, but I think what you need to do is double click even further from Gen. I to get up co-pilot or healthcare or Microsoft 365 copilot or security products built with Geni. I do think ultimately, you know, the closer you can kind of double click on is this really delivering ROI, then you have much more precise data.
Starting point is 00:42:28 Because I do think a lot of companies have dabbled or done internal kind of, I'll call it proof of concepts. And some of them have paid off and some of them haven't. But, you know, but I think ultimately a lot of the solutions that are paying off that we continually hear from our clients and our customers is, you know, a bunch of the applications, for example, that we've built. I think similarly, you know, a bunch of the applications that our partners have built on top of it. us. And, you know, ultimately the Azure business is, you know, we get paid based on consumption. It's a consumption-based business, meaning if people aren't actually running something, we don't get paid. It's not like they're pre-buying a ton of stuff. You know, we recognize our revenue based on when it's used. And so, you know, the good news is when you look at our revenue growth,
Starting point is 00:43:19 it is, you know, it's not a bookings number. It's actually a consumption number. And you can tell that People are consuming more and, you know, the last two quarters are revenue growth accelerated on a big number. And that is a statement of the fact that I think people are getting a lot of ROI, at least with the projects that they're running on top of our cloud. Yeah, I think that's an important point to bring home. It is consumption-based. So you talked a little bit about water cooling versus air cooling. I love the term for the air cooling. It's called chillers.
Starting point is 00:43:54 And that's what my friends in high school called ourselves, you know, back in the day. And I want to end on the GPU side of things or the silicon side of things. What do you think the potential is for custom silicon in the AI world? I mean, like we talked about previously, GPUs were designed for gaming. It happened to do parallel processing. Actually ended up being really good for, you know, large language models, the training and the inference. What's your perspective on whether this industry is going to continue to run, on that type of chip and what the potential is for custom silicon.
Starting point is 00:44:33 I think a couple things. I think one is I think increasing the number of tokens you can get per watt per dollar is going to be the game over the next couple years and maximizing the ability of our cloud to deliver the best volume of tokens for every watt of power, for every dollar that's spent, where the dollar is spent on energy, it's spent on the GPUs, it's spent on the data center infrastructure, it's spent on the network, and it's spent on everything else, is the thing that we're laser focused on. And it is, you know, there's a bunch of steps as part of that, GPUs being a critical
Starting point is 00:45:17 component of it. And, you know, one of the things that our scale gives us the ability to do is, to invest for kind of non-linear improvements in that type of productivity and that type of yield. If you've got a million dollars of revenue on a couple hundred GPUs, you're not going to be investing in custom silicon. When you're at our scale, you will be. And you're not just investing in custom skillikin for GPUs for pre-training or for inferencing. You're looking at what can we be doing for synthetic data generation with silicon? What can we be doing? What can we be doing from a network compression perspective with Custom Silicon? What can we be doing from a security perspective? And we have bets across all of those, many of which are now in production and are
Starting point is 00:46:05 actually powering a lot of these AI experiences. In fact, I think every GPU server that we're running in the fleet right now is using Custom Silicon at the networking compression storage layer that we've built. Now, the GPUs themselves are also going to be a prize that people are going to try to optimize, like the actual instructions for doing the GPUs. Invidio is a fantastic partner of ours. We're probably one of, if not the biggest customer in the world of theirs. And we partner super deeply with Jensen and his team. You know, at the same time, and partly why they're so successful is they're executing incredibly well. You know, at the same time, if you look at the history of silicon, not every single company, or it's rare that have a silicon company that every single
Starting point is 00:46:55 year is doing the absolute perfect work that's differentiated. And kudos to Jensen for what he's done, and I know he's going to keep trying to do it going forward. But, you know, there will be other opportunities from other companies where people are going to look for a niche that's going to be big enough in this AI space to be truly differentiated versus what NVIDIA is delivering. and then we're doing our own silicon investment in-house, because we're going to be going after those same opportunities.
Starting point is 00:47:23 And ultimately, the way we've tried to build our infrastructure, none of our customers know when they're using Microsoft 365 or GitHub or any company on what silicon they're running on. And we're going to be constantly tuning the use cases based on the applications. And if we find ways that are breakthroughs, we're absolutely going to be taking advantage of them for those use cases. And again, at our balance of scale and our balance of use cases, I'm very confident that we're going to find use cases where custom silicon will make a difference.
Starting point is 00:47:55 And I'm also very confident we're going to continue to be a great partner to Nvidia and others in the world that are going to be selling us great solutions. All right, Scott, I want to end on this because I've always been curious about the human aspect of this. Like, you're going out and working on designing your own chips that are trying to be better than GPUs for certain parts of this. application layer and training. And then you said one of Nvidia's biggest customers, if not its biggest customer. So is this like a situation where like you go to Jensen and you're like, we're going to just both give a shot at building this stuff and made the best chip win
Starting point is 00:48:35 and it's friendly, like friendly competition or is there any awkwardness in there? Because you're like kind of building the thing that is making them the most valuable company in the world. Well, I think probably different companies handle that differently. I think the nice thing about Microsoft is A, we've been around a while, and I think also we're, you know, we compete almost in every market in some way, shape, or form. So, like, there's none of my partners that I'm not also a competitor with, I think, is probably a true thing. It's crazy.
Starting point is 00:49:03 And the important thing is, I think you have that enterprise maturity to be able to recognize, you know, I want Jensen to do the best possible work because it's going to benefit me. And we've leaned in. We were the very first cloud to deliver live GB200s, which is a massive architectural shift for NVIDIA. That's the first of their liquid. The great spackwell. Yeah.
Starting point is 00:49:27 And we were the first one running, first rack running, the first cluster running, the first data center running of any cloud or neocloud provider in the world. And so, you know, that's an example where we really leaned in and moved at the speed of light together. And we're going to continue doing those types of projects. And at the same time,
Starting point is 00:49:43 know, he recognizes and understands we're going to be doing lots of things. And I also recognize he's going to work with other providers as well. So I think the ability to kind of keep a complete thought and recognize it's not zero sum on every single decision. And that at the end of the day, you know, it's a market. We're all going to compete. And we're also going to partner. And, you know, I think we have the maturity at Microsoft to do that, again, the balance. I think I've said balance multiple times. I do think balance in life, but I'll especially, in business and especially in technology, that is the devil's in the detail. But if you can get that right and do it consistently, those are the companies that win. And those are the companies
Starting point is 00:50:25 that really have the ability to set the agenda. And that's what we're focused on. Well, Scott, I just want to say thank you for taking the time. I know you don't do this often. So I appreciate why did you say, okay, hey, I want to come out and speak about this today? well a bunch of people internally said hey you're going to talk to Alex and so that's always a good advice to follow okay so it's it's fun to get a chance to do and uh really look i really enjoyed the conversation as did I yeah thank you again for taking the time again and I know it's rare for you to come out and speak about these things uh you're running a massive massive and fast growing business and so it was great to be able to speak with you and get into peak I get a
Starting point is 00:51:09 peek into it today and look as to what the rest of the industry is doing and your perspective on that. So thanks for coming on the show, Scott. Appreciate it. Thanks for having me, Alex. All right, everybody, thank you so much for listening and watching. We'll be back on Friday to break down the week's news with Max Zhef of TechCrunch. It's going to be a great episode, and we hope to see you there. Thanks again, and we'll see you next time on Big Technology Podcast.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.