a16z Podcast - Jensen Huang and Arthur Mensch on Winning the Global AI Race

Episode Date: March 21, 2025

The global race for AI leadership is no longer just about companies—it’s about nations. AI isn’t just computing infrastructure; it’s cultural infrastructure, economic strategy, and national se...curity all rolled into one.In this episode, Jensen Huang, founder and CEO of NVIDIA, and Arthur Mensch, cofounder and CEO of Mistral, sit down to discuss sovereign AI, national AI strategies, and why every country must take ownership of its digital intelligence.How AI will reshape global economies and GDPThe full AI stack—from chips to models to AI factoriesWhy AI is both a general purpose technology and deeply specializedThe open-source vs. closed AI debate and its impact on sovereigntyWhy no one will build AI for you—you have to do it yourselfIs this the most consequential technology shift of all time? If so, the stakes have never been higher.Resources: Find Arthur on X: https://x.com/arthurmenschFind Anjney on X: https://www.linkedin.com/in/anjney/Find NVIDIA on X: https://x.com/nvidiaFind Mistral: https://x.com/MistralAI Stay Updated: Let us know what you think: https://ratethispodcast.com/a16zFind a16z on Twitter: https://twitter.com/a16zFind a16z on LinkedIn: https://www.linkedin.com/company/a16zSubscribe on your favorite podcast app: https://a16z.simplecast.com/Follow our host: https://twitter.com/stephsmithioPlease note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details please see a16z.com/disclosures.

Transcript
Discussion (0)
Starting point is 00:00:00 This is the greatest force of reducing the technology divide the world's ever known. It will have an impact on GDP of every country into double digits in the coming years. Nobody's going to do this for you. You've got to do it yourself. It's up to organizations, to enterprises, to countries, to build what they need. The stakes at play are basically the equivalent of modern digital colonialization. AI isn't just computing infrastructure. cultural infrastructure. The race for AI dominance is not only constrained to companies, but is increasingly
Starting point is 00:00:37 capturing the attention of countries. And that includes the infrastructure spanning every layer of the stack. The chips, the models, the applications, plus the energy required to run these, quote, AI factories, the talent needed to produce them, and well-designed policy that helps not hinders this entire ecosystem. And all of this together is, turning critical. Set up is always hard.
Starting point is 00:01:03 This is no different. The only question is do you need to do it? If you want to be part of the future, and this is the most consequential technology of all time, not just our time, of all time, digital intelligence, how much more valuable, how much more important can it be? In today's episode, we explore sovereign AI
Starting point is 00:01:24 and this regional race for AI infrastructure across countries big and small. And there is truly no one better to discuss this than our guests, Jensen Huang and Arthur Mench. Jensen, of course, is the inimitable co-founder and longtime CEO of NVIDIA, a company known for its constant reinvention and ability to place critical bets like the GPU or graphics processing unit that has propelled it to be one of the largest companies at over $3 trillion in market cap as of this recording. Of course, the products that NVIDIA makes, like the GPU, are also the backbone
Starting point is 00:01:58 to so much of our digital world today. Arthur, on the other hand, is the co-founder and CEO of Mistral, a leading AI lab that focuses on customizable, open-source frontier models, but also a growing number of tools to help companies and even countries engage with AI. Today, Arthur and Jensen sit down with A16Z general partner, Anjani Amira, as they explore the role of digital intelligence at the nation level and how countries should think about ownership, codifying their culture, and the role that open source should play.
Starting point is 00:02:30 All right, let's get started. As a reminder, the content here is for informational purposes only, should not be taken as legal, business, tax, or investment advice, or be used to evaluate any investment or security and is not directed at any investors or potential investors in any A16C fund. Please note that A16D and its affiliates may also maintain investments in the companies discussed in this podcast. For more details, including a link to our investments,
Starting point is 00:02:55 please see A16C.com slash Disclosures. Today we're talking about sovereign AI, all things, national infrastructure and open source. So let's just start with the first question I usually get from nation-state leaders, which is, is AI actually a general purpose technology? In the history of humanity, we've had maybe a handful of these, 22, 24 economists call these specific technologies that accelerate economic progress broadly across society. Electricity, the printing press. And the question everybody's asking right now, is that the right way to think about AI?
Starting point is 00:03:36 Or why isn't AI just another important, but ultimately narrow technology? I think it's a general purpose technology because it basically revisits entirely the way we are building software and the way we are using machines. And so in the same way that Internet was a general purpose technology, AI is a general purpose technology here. It allows to build agents that are doing things on your behalf. And in that respect, it can be used in any vertical of the industry. It can be used for services, for public services. It can be used to change the life of citizens. It can be used for agriculture. It can obviously be used for defense purposes. So it covers everything that a state needs to worry about.
Starting point is 00:04:16 And so in that respect, it's very natural that any state makes it a priority and makes it a dedicated nationwide strategy. By the way, everything Arthur said is 100% correct. It is also exactly the reason why everybody's given up
Starting point is 00:04:30 and it's precisely wrong. And the reason for that is this. If it's a general purpose technology and one company can build the ultimate general purpose technology then why shouldn't
Starting point is 00:04:40 anybody else do it? And that is the flaw. Right. But that's also the mind trick to convince everyone that intelligence is only something that a few people ought to go build.
Starting point is 00:04:54 Everybody ought to sit back and wait for it. I would advise that everybody engage AI. And it is not just a few companies in the world who should build it. Everybody should build it. Nobody's going to care more about the Swedish culture and the Swedish language and the Swedish people and the Swedish ecosystem more than Sweden. Nobody's going to care about the ecosystem of Saudi Arabia more than Saudi Arabia. and nobody's going to care about Israel more than Israel,
Starting point is 00:05:22 despite the fact that the technology is general purpose and absolutely true, how could intelligence not be general purpose? It is also hyper-specialized. And the reason for that is because, let's face it, I don't think I'm waiting around for a general-purpose chatbot to be an expert in a particular area of disease. I still think that I would prefer to have somebody who is hyper-specialized in that field to fine-tune, to train, and post-trained, if you will,
Starting point is 00:05:52 an AI model that's going to be specialized in that. It's the general purpose technology the same way a programming language is the general purpose technology. And in addition to that, it's also a culture-carrying technology. So I think what that means is that there's an infrastructure.
Starting point is 00:06:10 There are chips that obviously not every country are going to build. There are general-purpose models like base models, compression of the web, that are eventually going to be open source and that can serve as the right basis for constructing specialized systems. But beyond that, I think it's up to organizations,
Starting point is 00:06:28 to enterprises, to countries, to build what they need. So the way to make it work is to take a general purpose model, like an open source model, for instance, and to get the knowledge you have specifically or ask your citizens or ask your employees to distill their knowledge into the systems into the agents that are going to be working on your behalf so that progressively those agents become more accurate
Starting point is 00:06:53 and following the instructions and the specifications that the country or an enterprise may have. So you need vertical experts or you need cultural experts or you need people with a certain national agenda to partner with technological companies that can expose the open source infrastructure in a way that is easy to use.
Starting point is 00:07:16 use, and in a way that is easy to specialize. So I think that's where the frontier lies. It's a very horizontal technology. To make anything useful out of it, you need the partnership between the horizontal providers and the vertical experts. But unlike previous general purpose technology waves in history, like electricity or the printing press, how is this one different? If I'm a nation state leader and I'm trying to understand what the right framework is for
Starting point is 00:07:39 me to think about AI in my country, should I think about it like digital labor or should I think about it as akin to bridges? I think it's similar to electricity in the sense that it will have an impact on GDP of every country in the double digits in the coming years. So that means that from an economical point of view, every nation needs to worry about it because if they don't manage to set up infrastructure, to set up their own sovereign capacities at the right place,
Starting point is 00:08:05 that means that this is money that might flow back to other countries. So that's changing the economic equilibrium across the world. In the sense, that's not very different from electricity. 100 years ago, if you weren't building electricity factories, you were preparing yourself to buy it from your neighbors, which at the end of the day isn't great because it creates some dependencies. I think in that sense, it is similar. What is fairly different, I think there's two things.
Starting point is 00:08:29 First of all, it's kind of an amorphic technology. If you want to create digital labor with it, you need to shape it. You need to have infrastructure, talent, and software. And the talent needs to be created locally. I think this is quite important. And the reason for that is that, in contrast with electricity,
Starting point is 00:08:46 this is a content-producing technology. So you have agents that are producing content, that are producing text, producing images, producing voice, interacting with people. And when you're producing content and interacting with society, you become a social construct.
Starting point is 00:09:00 And in that respect, social constructiary cultures and values of either an enterprise or a country. And so if you want those values not to disappear and not to depend on a central provider, you need to engage with it more profoundly that you would be to engage with electricity, for instance.
Starting point is 00:09:17 Would you agree with that, Jensen? A couple of ways to think about it. Your country's digital intelligence is not likely something you would want to outsource to a third party without some consideration. Your digital intelligence is just now a new infrastructure for you. Your telecommunications, your healthcare, your education, your highways, your electricity.
Starting point is 00:09:45 This new layer is your digital intelligence. It's your responsibility to decide how you want this digital intelligence to evolve and whether you want to outsource it so that you could never have to worry about intelligence again, or this is something that you feel you want to engage, maybe even control and shape into a national infrastructure. Of course, it has all the things that author said, AI factories, infrastructure, etc. there's another way you could think about it is your digital workforce. Now this is a new layer and you've got to decide whether the digital workforce of your country or your company is something that you decide to outsource, hope it evolves the way that you would like it to, or is it
Starting point is 00:10:29 something that you want to engage, maybe even decide to control and nurture and make better. we hire general purpose employees all the time. We hire them out of school. Some of them are more general purpose than others. Some of them are more intelligent than others. But once they become our employees, we decide to onboard them, train them, guardrail them, evaluate them, continuously improve them. We make the investment necessary to make general purpose intelligence into super, intelligence that we could benefit from.
Starting point is 00:11:10 And so I think that that second layer, thinking about it as digital workforce, in both cases, it contributes to the national economy. In both cases, it contributes to social advance. In both cases, that it contributes to the culture. And I think that in both cases, a country needs to play a very active role in it. And so I think it's back to your original question about sovereign AI. how to think about it. Yes, it is definitely a general purpose technology, but you have to decide how to shape it. Your country's digital data belongs to you. Your national library,
Starting point is 00:11:51 your history, for so long as you want to digitize it, you could make it available to everybody in the world. You could also make it available to companies or researchers and institutions in your own country. It belongs to you. Of course, These are all vaporous things. They're very soft ideas. But it does belong to you. And you could decide it belongs to you in the sense that this is where you came from. You could decide how to put it to use for the benefit of your people.
Starting point is 00:12:21 And it belongs to you in the sense that it's your responsibility to shape its future. Sovereign AI. It's your responsibility. There are several other types of assets that nation states fund and protect, the military, your electricity grid. Let's say I have understood now the criticality of AI infrastructure and sovereign AI. Do I have to now take control of every part of the stack? So Jensen mentions, I guess, digital workforce. And I think it's a very good analogy that you need an onboarding platform for your AI workforce,
Starting point is 00:12:54 which means you need to be able to customize the models and pour the knowledge that are sitting in your national libraries into the model so that suddenly speaks better your language. You need to get your systems to know about your lows. So that suddenly the guard wells that are set when you're deploying an AI software are compliant. And so that onboarding platform that requires to customize, to guardrail, to evaluate. And then when noticing that certain things needs to be improved,
Starting point is 00:13:24 to fix things, to debug things, that's the platform that we are building. So being able to deploy systems that are easy to tune and working with these platform providers to do the custom systems. And once the custom systems are made, it's important to be able to maintain them yourself. So that means being able to deploy them on your own infrastructure,
Starting point is 00:13:47 being able to ask your technological partners to potentially disappear from the loop. Your IT department is going to become the HR department of your digital workforce. And they're going to use these tools that Arthur describes to onboard AI's fine-tune AI, guard rail them, evaluate them, continuously improve them. Right.
Starting point is 00:14:07 And that flywheel will be managed by the modern version of the IT department. Right. And we'll have biological workforce and we'll have a digital workforce. It's fantastic. And so nobody's going to do this for you. You've got to do it yourself. Right. That's why even though we have so many technology companies in the world, every company
Starting point is 00:14:27 still has their own IT department. I've got my own IT department. I'm not going to outsource it to somebody else. In the future, there'll be even more important. to me because they'll be helping us manage these digital workforces. You're going to do this in every country. You're going to do this in every company within those countries. And so the space for what Arthur is describing to take this general purpose technology, but to really fine-tune it into domain experts. They're national experts or their industrial experts or their corporate
Starting point is 00:14:59 experts or functional experts. This is the future, the giant future space of AI. So you both said something that I just want to make sure I'm understanding correctly. You called it a soft concept like your culture, and you said there are a bunch of norms that the training data has that you customize the models on. You said norms that exactly means it's soft versus rules, which are more hard, or algorithms and laws, which are very specific. There's different things that you want to incorporate into your AI systems.
Starting point is 00:15:29 There are some elements of style and of knowledge that you're not going to enforce through strict guardrails that you can enforce through continuous training of models, for instance. You take preferences and you distill it into the models themselves. And then you have a set of laws, you have a set of policies if you're in a company, and those are strict. And so usually the way you build it is that you connect the models
Starting point is 00:15:52 to the strict rules and you make sure that every time it answers, you verify that the rules are respected. on one side you're pouring and compressing knowledge in a soft way into the models and on the other side you're making sure that you have a certain number of policies and rules that are strictly enforced and that have a hundred percent accuracy so on one side this is soft this is preference this is culture preference somebody's preference is multidimensional you know what you prefer it depends it's implicit many times in communication there's so many features that defines my preference
Starting point is 00:16:23 it takes AI to be able to precisely comply with the description that Arthur was describing just now. Could you imagine if a human had to write this in Python, describe every one of these, capture every one of these things in C++. Based on this, I prefer that, but if you did that, I prefer that other thing. And, I mean, the number of rules would be insane, which is the reason why AI has the ability to codify all of this. It's a new programming model that can deal with the ambiguity of life. Well, it sounds like you're saying AI isn't just
Starting point is 00:16:54 computing infrastructure, it's also cultural infrastructure. Yes, it is. Is that right? And it's about making sure that your cultural infrastructure and the human expertise that are in your company
Starting point is 00:17:04 or in your country makes it to the AI systems. Right. Culture reflects your values. We were just talking about how each one of these AI models, AI services,
Starting point is 00:17:14 respond differently to the type of questions you're asked. Right. Because they codify the values of their service or the values of their company into each one of their services. Could you imagine this now amplified at an international scale?
Starting point is 00:17:30 This is an inherent limitation of centralized AI models, where you're thinking that you can encode some universal values and some universal expertise into a general purpose model. At some point, you need to take the general purpose model and ask a specific population of employees or of citizens, what are their preferences and what are their expectations? And you need to make sure that your, specializing the model in a soft way and in a hard way for rules and through culture and preferences.
Starting point is 00:17:58 And so that part is not something that you can outsource as a country. It's not something that you can outsource as an enterprise. You need to own it. Well, then is it an exaggeration to say if it is cultural infrastructure and I don't own sovereignty of it, the stakes at play are basically the equivalent of modern digital colonialization. If you're saying, Ange, you've got to think about AI is almost like your digital workforce and another country or somebody who's not my sovereign nation can decide what my workforce can and can't do. That's a problem. Some of it is universal. For example, it is possible for certain companies to serve nations and society and companies around the world
Starting point is 00:18:40 because it's basically universal. But it cannot be the only digital intelligence layer. It has to be augmented by something regional. You know, I think McDonald's is pretty good everywhere. All right. Kentucky fried chicken is pretty good everywhere. But you still want the local style, local taste that augments on the last mile. That's right. The local cafes, the mom and pop restaurants, because it defines the culture. It defines society. It defines us. I think it's terrific that you have Walmart everywhere that you can count on everywhere. You know, I think it's fine. But you need to have local taste, local style, local preference, local excellence, local services. Let me swing in another way. It is very likely that in the context of our digital workforce in the future, we will have some digital
Starting point is 00:19:27 workers, which are generic. Right. They're just really good at doing maybe basic research or something basic. Good college level graduate. Or they're useful for every company. It's unnecessary for me to create something new. Right. I think Excel is pretty good.
Starting point is 00:19:44 Microsoft Office is universally excellent. Right. I'm perfectly fine with it. Good reference architecture base. That's right. Then there's industry-specific tools, industry-specific expertise that is really important. For example, we use synopsis and cadence.
Starting point is 00:19:59 Arthur doesn't have to, because it's specific to our industry, not his. We probably both use Excel, probably both use PDFs. We both use browsers. And so there's some universal things that we can all take advantage of, and there'll be universal digital workers that we can take advantage of. And then there'll be industry-specific, and then there'll be company-specific. right inside our company we have some special skills that are very important to us that defines us it's highly biased if you will very guardrail to doing very specific work highly biased to
Starting point is 00:20:32 the needs and the specialties of our company right and so we become superhuman in those areas well your digital workforce is kind of be the same and AI is going to be the same there'll be some that you just take off the shelf the new search will likely be some AI right the new research will probably be some AI but then there'll be industrial versions of AIs that we'll maybe get from Cadence and others, and then we'll have to groom our own using Arthur's tools. Right. And we'll have to fine tune them, we'll onboard them, make them incredible.
Starting point is 00:21:02 I very much agree with this vision of having a general purpose model and then some layer of specialization for industries and then an extra layer of specialization for companies. You will have a tree of AI systems that are more and more specialized. And maybe to give a concrete example with what we recently did, so we released in January a model called Mistral Smol, and it's a general purpose model. So it speaks all of the languages, it knows mostly about most things. But then what we did is that we took it and we started a new family of specialized models that were specialized in languages.
Starting point is 00:21:34 So we took more languages in Arabic, more languages in Indian languages, and we retrained the model. And so we distil this extra knowledge that the initial model hadn't seen. And so in doing that, we actually made it much, much better in being idiomatic when it speaks Arabic and when it speaks languages from the Indian peninsula. And so language, it's probably like the first thing you can do when you're specializing a model. The good thing is that for a given size of model, you can get a model that is much better if you choose to specialize it in a language. So today, our model, it's the 24B, it's called Mistral Saba, it's a model
Starting point is 00:22:09 tune in Arabic, is outperforming every other language model that are like five times larger. And the reason for that is that we did this specialization. And so that's the first layer. And then If you think of a second layer, you can think of verticals. So if you want to build a model which is not only good at Arabic, but also good at handling legal cases in Saudi Arabia, for instance, well, you need to specialize it again. So there's some extra work that needs to be done in partnership with companies to make sure that not only your system is good at speaking a certain language, but it's good at speaking a certain language and understanding the legal work that is done in this language.
Starting point is 00:22:46 And so it's true for any combination that you can think of, of vertical and language. I see. You want to have a medical diagnosis assistant in French. Well, you need to be good at French, but you also need to understand how to be good at speaking the French language of physicians. And so those two things, it's very hard to do as a general purpose model provider. If this is true and what you're describing is real, that I need the capabilities to customize this AI layer on my local norms, my local data. which is fairly sophisticated from a technical capability perspective. How would you advise a big nation to think about the stack we're talking about,
Starting point is 00:23:24 the chips, the compute, the data center, the models that sit on top the applications, and then ultimately what you were describing as the AI nurse or the AI doctor? And how would you advise someone that's a smaller nation differently? I would say you need to buy and to set up the horizontal part of the stack. So you need the infrastructure. You need the inference primitives, you need the customization primitives, you need the observability, you need the ability to connect visuals to models, to connect models to sources of information, of real-time information.
Starting point is 00:23:58 Those are primitives that are fairly well factorized across the different countries, across the different enterprises. And once you have that, these are things that can be bought. Then you can start working. Then you can start building. You build from these primitives according to your values, according to your expertise, and thanks to your local talent. The question is, where is the frontier
Starting point is 00:24:17 and between what is horizontal and horizontal, if you're a small enterprise or a small country, you should probably buy. And what is vertical and specific to you, and that's definitely something that you need to build.
Starting point is 00:24:29 You have to get it in your head that it's not as hard as you think it is. First of all, because the technology is getting better, it's easier. Could you imagine doing this five years ago? It's impossible.
Starting point is 00:24:41 Could you imagine doing this five years from now? It'll be trivial. And so we're somewhere in that middle. The only question is, do you have to do it? Right. The truth of the matter is, I hate onboarding employees. And the reason for that is because it takes a lot of work.
Starting point is 00:24:55 But once you set up an HR organization and leadership mentoring organization and processes, then your ability to onboard employees is easier and is systematically more enjoyable for everybody involved. But in the very beginning is hard. Set up is always hard. Set up is always hard. This is no different. The only question is, do you need to do it? If you want to be part of the future, and this is the most consequential technology of all time, not just our time, of all time. Digital intelligence, how much more valuable, how much more important can it be? And so if you
Starting point is 00:25:34 come to the conclusion this is important to you, then you have to engage it as soon as you can, learn along the way, and just know that it's getting easier and easier all the time. The fact that the matter is if we try to do agentic systems even three years ago, it was incredibly hard. Right. But agentic systems are a lot easier today. And all of the tools necessary for curating data sets, for onboarding the digital employees to evaluating the employees, the guard railing, digital employees, all of those are getting better all the time.
Starting point is 00:26:02 The other thing about technology is when it becomes faster, it's easier. Right. Could you imagine back in the old days? Of course, I had the benefit of seeing computers from its earliest days. and the performance of the computers were so frustratingly slow. Everything you did was hard. But these days,
Starting point is 00:26:18 the type of things we do is just magical because it's also fast. And so whether it's motivated by your institutional need to engage the most consequential technology of all time or the fact that it's getting better all the time, so it's not that hard,
Starting point is 00:26:37 I think the number of excuses is running out. So let's talk about that for a second because change is hard. I've got an endless list. If I'm a nation state leader, I'm facing increasing amounts of geopolitical risk. I don't know who my allies are. Elections are coming. There's any number of things I've got to deal with. But now let's say I understand that this is important. You guys spend so much time talking nation state leaders who are thinking about what are the risks of adopting AI too fast. And you're right, the zeitgeist has shifted based on the Paris Action Summit. It seemed like there's a tone of optimism more than there was a tone of pessimism a year ago.
Starting point is 00:27:09 but what are the most common questions you get from nation state leaders when they're asking about risks and how to think about them? So I've heard several questions, but one of the risk is to see your population start getting afraid of the technology for fear of it replacing them. And that is something that can actually be prevented. If we collectively make sure that everybody get access to the technology and is trained in using it.
Starting point is 00:27:34 So the skilling of the various citizens of the populations is extremely important. And stating AI as an opportunity for them to actually work better and showing the purpose of it through applications, through things that they can actually install on their smartphone, through public services. We're working, for instance, with the French unemployment system to actually connect opportunities of jobs to unemployed people through AI agents
Starting point is 00:28:01 that are being actionated by, obviously, human operators within the agency. And that is one opportunity. That's a very palatable opportunity for people to find a job better. And so that's part of the thing that can make sure that population understand the opportunity and the fact that AI is really just a new change for them to adopt, just the same way they had to adopt personal computers in the 90s and Internet in the 2000s. The common aspect with these changes is that you need people to embrace the technology. And I think the biggest problem that nation states may have is to
Starting point is 00:28:37 see AI increase the digital divide that is already relatively big but if we work together and if done in the right way we can make sure that AI is actually reducing the digital divide. AI is a new way to program a computer. It is because by typing in some words you can make the computer do something. Just like we did in the past.
Starting point is 00:28:58 Right. I know you talk to it. You can interact with it in a whole lot of ways. You can make the computer do things for you a lot easier today than it was before. Right. The number of people
Starting point is 00:29:10 who could prompt chat GPT and do productive things just from a human potential perspective is vastly greater than the number of people who can program C++ ever. And therefore, we have closed the technology divide.
Starting point is 00:29:28 Probably the greatest equalizer we've seen. It is by definition. Right. The greatest equalizer of technologies of all time. But you still need, to have citizens to know about it. I think that's the thing.
Starting point is 00:29:40 I'm just describing the fact. Yes. The fact is there are more people who program computers using chat GPT today than there are people who program computers using C++. Right. That's a fact. And so the fact is this is the greatest force
Starting point is 00:29:57 of reducing the technology divide the world's ever known. Right. It's just perceived, and what Arthur's saying, the perception through, I don't know who and I'm talking about it and I don't know how talking about it
Starting point is 00:30:11 but the fact that the matter is it is not stopping right it's not stopping anything the number of people who are actively using chat GPT today is off the charts
Starting point is 00:30:20 I think it's terrific it's completely terrific anybody who's talking about anything else apparently isn't working and so I think people realize the incredible capabilities
Starting point is 00:30:32 of AI and how it's helping them with their work I use it every single day. I used it this morning. And so every single day I use it. And I think that deep research is incredible. Right. My goodness, the work that Arthur and all of the computer scientists around the world are doing is incredible.
Starting point is 00:30:49 And people know it. People are picking it up, obviously, right? Just the number of active users. Let's talk about open source for a bit. Because both of you have talked quite publicly about the importance of open models in the context of sovereignty. At DeepMind, you're part of the Chinchilla skating laws, which were openly published, your co-founder, Guillaume, created Lama. And then last year, NVIDIA and Mistral worked on a jointly trained model called Mistral Nemo. Why are open models
Starting point is 00:31:14 such a big part of your focus? Because it's an horizontal technology and enterprises and states are going to be eventually willing to deploy it on their own infrastructure. Having this openness is important from a sovereignty perspective. That's the first point. And then the second point of importance is that releasing open source models is a way. to accelerate progress and we created Mistral on the basis that what we've seen during our early career when we were doing AI in between 2010 and 2020 was an acceleration of progress because every lab was building on top of each other. And that's something that kind of disappeared with the first large language models from open AI in particular. And so spinning back that open
Starting point is 00:32:00 fly wheel of I contribute something and then another lab is contributing something else. And then And we iterate from that is the reason why we created Mistral. And I think we did a good job at it because we started to release models and then META started to release models as well. And then we had Chinese company like Deep Seek release stronger models. And everybody benefited from it. So coming back to Mistral Nemo, one difficulty of creating AI models in an open way is that this is more a cathedral than a bazaar setting when it comes to open source. Because you have large spend to do to build a model. And so what we did with Nvidia team is really.
Starting point is 00:32:34 to mix the two teams together, have them work on the same infrastructure, the same code, have the same problems, and combine their expertise to build the same model. And that has been very successful because Nvidia brought a lot of things we didn't know. I think we brought things that Nvidia didn't know. And at the end of the day,
Starting point is 00:32:51 we produced something that was at the time the best model for its size. And so we really believe in such collaborations and we think that we should do them more and at a higher scale. And not only with only two companies, but probably with three or four. And that's the way open source is going to prevail.
Starting point is 00:33:06 I completely agree. The benefit of open source, in addition to accelerating and elevating the basic science, the basic endeavor of all of the general models and the general capabilities, is the open source versions also activate a ton of niche markets and niche innovation. All of a sudden, in health care, life science, sciences, physical sciences, robotics, transportation, the number of industries that were activated as a result of open source capabilities are sufficiently good is incredible. Don't ignore the incredible capabilities of open source, particularly in the fringe, the niche. But mission critical where data might be sensitive. Yeah, it could be, for example, in mining energy. Right. Who's going to go create an AI company to go mine energy? Energy is really important,
Starting point is 00:34:01 but the mining of energy is not that big of a market. And so open source activates every single one of them. Financial services, it turns out, activates them. Healthcare, defense. You pick your favorites. Anything that is mission critical and that requires to do one's own deployment, that potentially requires to do on the edge deployment as well. Right.
Starting point is 00:34:20 And anything that requires some strong auditing and the ability to do a thorough evaluation of it. You can evaluate a model much better if you have access to the weights and if you only have access to APIs. And so if you want to build certainty around the fact that your system is going to be 100% accurate, I don't think you should be using a close source model. And you have to connect it into your flywheel. How are you going to connect your local data?
Starting point is 00:34:44 Yeah, you have to connect it into your own, your local data, your own local experience. The more you use it, the better it becomes that flywheel. You can't do it without open source. But let's say I'm a nation state leader. I've been considering open source. I'm starting to hear things like, hey, open source is a threat to national security. we should not be exporting our models because these open models actually give away
Starting point is 00:35:06 a ton of nation-state secrets or more importantly, the bad guys can use these open models too. And so this is a threat to security. Instead, what we should be doing is locking down maybe development amongst two or three labs that have the infrastructure to get licenses from the government to do training, to do the right safety and certification. I've certainly been hearing that a lot.
Starting point is 00:35:25 How should I think about that versus what you're telling me, which is actually no open is better for mission-critical industries? collaboration in between labs is going to be critical for humanity success and if one state decides to lock things down the only thing that is going to happen is that another state will take the leadership because cutting yourself from the open fly wheel has just too high of a cost for you to maintain competitivity if you do that
Starting point is 00:35:48 this is a debate that has occurred in the United States and effectively if there's some export control over weight this is not going to stop any country in Europe any country in Asia to continue its progress. And they will collaborate to actually accelerate that progress. So I think we just need to embrace the fact that this is an horizontal technology
Starting point is 00:36:09 very similar to programming languages. Programming languages, they are all open source, right? So I think AI just needs to be open source in that respect. We're glad to see that this realization that we could accelerate together by being more open about the way we build the technology. And so it's great to see that open source has a lot of good days
Starting point is 00:36:29 before it. It is impossible to control. Software is impossible to control. If you want to control it then somebody else's will emerge and become the standard just as Arthur mentioned. And the question is, is open source safer? Open source enables more transparency, more researchers, more people to scrutinize the work.
Starting point is 00:36:54 The reason why every single company in the world is built, every cloud service provider is built on open source is because it is the safest technology of all. Give me an example of a public cloud today that's built on an infrastructure stack that isn't open source. You start from open source, you could customize it. But the benefit of open source is the contribution of so many people and the scrutiny, very importantly, You can't just put any random stuff into open source. You'll get laughed off the internet. You've got to put good stuff on the open source because the scrutiny is intense.
Starting point is 00:37:35 And so I think open source provides all of that. Great collaboration to accelerate innovation, escalate excellence, ensure transparency, attract scrutiny, all of that improves safety. In a sense, you're saying it's partly more secure because as we've seen with open source databases, storage, networking, compute, you get mass red-teaming. The whole world can help you red-team your technology versus just a small group of researchers
Starting point is 00:38:03 inside your company. Is that roughly, right? A right way to think about it? Exactly. By pulling a lot of organizations together to come up with a technology that they can all use and specialize on their own domains,
Starting point is 00:38:14 you're forcing the technology to be good for every one of them. And so that means you're removing biases. You're really making sure that the general purpose models that you're building are as good as possible and don't have failures. And I think open source in that respect
Starting point is 00:38:29 is also a way to reduce the number of failure points. If as a company today, I decide to rely fully on a single organization and on its safety principles, on its red-teaming organization as well, I'm trusting it a little too much. Whereas if I'm building my technology on open-source models, I'm trusting the world to make sure that the basis on which I'm building is secure.
Starting point is 00:38:52 So that's a reduction of failure points. And it's obviously something that you need to do as an enterprise or as a country. We're going to transition a little bit now into company building, which is something a lot of people are excited to hear from both of you about. So let's start with you, Jensen. You've remarked that NVIDIA is the smallest big company in the world. What enabled you to operate that way? Our architecture was designed for several things.
Starting point is 00:39:14 It was designed to adapt well in a world of change, either caused by us or affecting us. And the reason for that is because technology changes fast. And if you over-correct on controllability, then you are underserving a system's ability to become agile and to adapt. And so our company uses words like aligned instead of use words like control. I don't know that one time I've used the word control in talking about the way that the company works. We care about minimum bureaucracy, and we want to make our processes as lightweight as possible.
Starting point is 00:40:02 Now, all of that is so that we can enhance efficiency, enhance agility, and so on and so forth. We avoid words like division. When Nvidia was first started, it was modern to talk about divisions. Right. And I hated the word divide. Why would you create an organization that's fundamentally divided? I hated the word business units. The reason for that is because why should anybody exist as one?
Starting point is 00:40:27 Why don't you leverage as much of the company's resources as possible? I wanted a system that was organized much more like a computing unit, like a computer, to deliver on an output as efficiently as possible. And so the company's organization looks a little bit like a computing stack. And what is this mechanism that we're trying to create? And in what environment are we trying to survive in? Is this much more like a peaceful countryside or is this like much more like a concrete jungle? What kind of environment are you in because the type of system you want to create should be consistent with that?
Starting point is 00:41:07 And the thing that always strikes me odd is that every company's org chart looks very similar, but they're all different things. One's a snake, the other one's a elephant, the other one's a cheetah. and everybody is supposed to be somewhat different in that forest, but somehow they all get along. Same exact structure, same exact organization, doesn't seem to make sense to me. I agree that it feels like companies have personalities, and despite the fact that they're organized sometimes similarly,
Starting point is 00:41:37 I should say that obviously we have a lot of things to learn, and I mean, the company is not even two years old. I guess one challenge we have with Mistrial, and I think our competitors have the same challenge, is that this is one of the first time that the software company is actually a deep tech company that is driven by science.
Starting point is 00:41:54 Science doesn't have the same timescales as software. You need to operate on a monthly basis. Sometimes you don't know exactly when the thing will be ready. But on the other hand, you have customers asking, when is the next model coming up? When is this capability going to be available, et cetera? And so you need to manage expectation. And I think for us, the biggest challenge,
Starting point is 00:42:15 and I think we're starting to do a good job at it, is to manage the hinge in between the product requirements and what the science is able to do. Research and product. Research and product. And you don't want the research team to be fully dedicated to making the product work. So you need to work, and I think we've started to do it,
Starting point is 00:42:36 on making sure that you have several frequencies in your company. You have fast frequencies on the product side, iterating every week. And you have slow frequencies on the science side that are looking at why profoundly the product is failing on certain domains and how they could fix it through research, through new data, through new architecture, through new paradigm. And I think that's fairly new.
Starting point is 00:42:58 This is not something that you would find in a typical SaaS company because this is inherently a science problem. I mean, Nvidia is one of the most successful companies that over a 30-year timeline has figured out a way to keep science and research ahead of the rest of the world, whether it was CUDA back in 2012, where those fundamental systems research, or Cosmos today, which is now saying, you know, it's definitely state of the art on how simulation should work out.
Starting point is 00:43:25 We've harmonized exactly what Arthur just said. Is that heuristic right for you? Yeah, we harmonized that inside our company. We have basic research, applied research, and then we have architecture, and then we have product development, and we have multiple layers of it. And these layers are all essential. Right. And they all have their own time clock in the case of basic research.
Starting point is 00:43:44 the frequency could be quite low. On the other hand, all the way to the product side, we have a whole industry of customers who are counting on us. And so we have to be very precise. And somewhere between basic research and discovering hopefully surprises that nobody expects.
Starting point is 00:44:02 On the one hand, on the other hand, to be able to deliver on what everyone expects. Predictably. These two extremes, we manage harmoniously inside our company. There's so many fascinating things about this market, but there's one in particular that I want to call out. Both of you have customers that are also your competitors. And those competitors are huge and highly capitalized tech giants. Nvidia sells GPUs to AWS, which is building its own chips called Traneum.
Starting point is 00:44:27 And Arthur, your training models that you sell through AWS and Azure who have funded labs like Anthropic and Open AI. So how do you win an environment like this and how do you manage those relationships? Because we talked about company building internally, but now I'm curious externally. How do you survive? Jensen said it well. You give up control but you work on alignment and despite the fact that sometimes you have certain companies can be competitors
Starting point is 00:44:51 you may have a line interests and you can work on specific agendas that are shared. You have to have your own place. Obviously these cloud service providers aren't working with Arthur because they already have the same thing. They just want two of the same things. It's because Arthur and
Starting point is 00:45:08 Mistral has a position in the world that is unique to Mistral, and they add value in a particular place that is unique. A lot of the conversation we've had today are areas that Mistral and the work and their position in the world makes them uniquely good at. And we are different. We're not just another ASIC. We can do things for the CSPs and do things with the CSPs that are not possible for them to do themselves. For example, NVIDE's architecture is in every cloud. And in a lot of ways, we have their first onboarding for amazing future startups. And the reason for that is because by onboarding to Nvidia, they don't have to make a
Starting point is 00:45:52 strategic or business or otherwise commitment to a major cloud. They could go into every cloud. And they could even decide to build their own system they like because the economics turns out to be better for them at some point or they would like access to capabilities that we have that are somewhat protective within the clouds. And so whatever the reasons are, in order to be a good partner to somebody, you still have to have a unique position. You need to have a unique offering. And I think Ms. Straw has a very unique offering.
Starting point is 00:46:20 We have a very unique offering. And our position in the world is important to even the people we compete against. And so I think when we are comfortable within that and comfortable with our own skin, then we can be excellent partners to all of the CSPs. And we want to see them succeed. I know that it's a weird thing to say when you see them as a competitor, which is the reason we don't see them as a competitor. We see them as a collaborator who happens to compete with us as well. And probably the single most important thing that we do for all the CSPs is bring them business. And that's what a great computing platform does.
Starting point is 00:46:53 We bring people business. I remember when Arthur and I first met, we sat down in London at a late night restaurant and sketched out the plan for his Series A. And we were figuring out why he needed so much capital for the Series A, which in hindsight was remarkably efficient. I think the Mistral Series A we put together was half a billion relative to other folks who had to spend multiple billions to get to the same place
Starting point is 00:47:13 but I asked him what chips would you like to run on? And you looked at me so absurdly as if I had asked you a question that how could it be an answer other than Nvidia other than H-100s and I think that ecosystem
Starting point is 00:47:24 has been the startup ecosystem that Nvidia has invested in creates so much business for the clouds what is the philosophy that led you to invest so deeply in startups
Starting point is 00:47:35 and founders so early on, even before anybody knew about them. Two reasons, I would say. One, the first reason is I rarely call us as a GPU company. What we make is a GPU, but I think of Nvidia as a computing company. If you're a computing company, the most important thing you think about is developers. Right. If you're a chip company, the most important thing you think about is a chip. And all of our strategies, all of our actions, all of our priorities, all of our focus, all of our investments, 100% of it is aligned with the
Starting point is 00:48:05 attitude that is developer first. It's about the computing platform first, another way of saying ecosystem. Right. And so everything starts there. Everything ends there. GTC is a developers conference. All of our initiatives inside the company is developer first. So that's number one.
Starting point is 00:48:22 The second thing is we were pioneering a new computing approach that was very alien to the world of general purpose computing. And so this accelerated computing approach was rather alien and counterintuitive and rather awkward for a very long time. And so we're constantly seeking out looking for the next incredible breakthrough, the next impossible thing to do without accelerated computing. And so it's very natural that I would find and would seek out researchers and great thinkers like Arthur because, you know, I'm looking for the next killer app. And so that's kind of a natural intuition, natural instinct of somebody who is creating something new. And so
Starting point is 00:49:04 If there's an amazing computer science thinker that we haven't engaged with, that's my bad. We've got to get on it. That's a perfect segue from a computing perspective. What are the most significant trends you see on the horizon? And in particular, for an audience who might be prime ministers, a president, so ministers of IT in some of the world's fastest growing markets, trying to understand where computing is going. How would you guide them? We're moving toward workloads that are more and more asynchronous. So workloads where you give a task to an AI system and then you, you, you, you.
Starting point is 00:49:34 wait for it to do 20 minutes of research before returning. So that's definitely changing a bit the way you should be looking at infrastructure because that creates more load. So I guess it's a bull case for data centers and for Nvidia. As I've said, I guess, in the beginning of this episode, all of this is not going
Starting point is 00:49:50 to happen well if you don't have the right onboarding infrastructure for the agents. If you don't have a proper way for your AI systems to learn about the people they interact with and to learn from the people they interact with. So that aspect of Learning from human interaction is going to be extremely important in the coming years.
Starting point is 00:50:09 And there's another aspect which is around personalization of having, I guess, models and systems consolidate the representation of their users to be as useful as possible. I think we are in the early stage of that. But that's going to change, again, pretty profoundly the interaction we have with machines that will know more about us and know more about our taste and how to be as useful as possible toward us. As a leader of a country, I want to think about education, about making sure that I have a local talent pool that understands AI enough to create specialized AI systems. And I want to think about infrastructure, both on the physical side, but also on the software side. So what are the right primities, what is the right partner to work with that is going to provide you with the platform of onboarding. And so those two things are important. If you have this and you have the talent, and if you do deep partnerships, the economy of your state is going to be profoundly changed. The last 10 years, we've seen extraordinary change in computing.
Starting point is 00:51:09 From hand coding to machine learning, from CPUs to GPUs, from software to AI, across the entire stack, the entire industry has been completely transformed. And we're going through that still. The next 10 years is going to be incredible. Of course, the industry has been wrapped up in talking about scaling laws. And pre-training is important, of course, and continues to do. be. Now we have post-training. And post-training is thought experiments and practice and tutoring and coaching and all of the skills that we use as humans to learn the idea that thinking and agentic
Starting point is 00:51:47 and robotic systems are now just around the corner. It's really quite exciting. And so what it means to computing is very profound. People are surprised that Blackwell is such a great leap over Hopper. And the reason for that is because we built Blackwell for inference. And just in time, because all of a sudden, thinking is such a big computing load. And so that's one layer, is there's a computing layer. The next layer is the type of AIs that we're going to see. There's the agentic AI, the informational, digital worker AIs. But we now have physics AI that's making great progress. And then there's physical AI that's making great progress. And physics AI is, of course, things that obeyed the physical laws and the atomic laws and the chemical laws and all of the various
Starting point is 00:52:34 physical sciences that are going to see some great breakthroughs, and I'm very excited about that. That affects industry, that affects science, affects higher education and research, and then physical AI, AI that understand the nature of the physical world, from friction to inertia, the cause and effect, object permanence, those kind of basic things that humans have common sense, but most AIs don't. And so I think that that's going to enable a whole bunch of robotic systems that are going to have great implications in manufacturing and others. The U.S. economy is very heavily weighted on knowledge workers. And yet many of the other countries are very heavily weighted on manufacturing. And so I think for many of the prime ministers and the leaders of countries to realize that the AIs that they need to transform and to revolutionize their industries that are so vital to them,
Starting point is 00:53:26 whether it's energy focused or manufacturing focused, it's just around the corner, and they ought to stay very alert to this. I would encourage people not to over-respect the technology. And sometimes when you over-admire a technology, over-respect the technology, you don't end up engaging it. You're afraid of it somehow. Some of the things that we said today about AI closing the technology divide
Starting point is 00:53:50 is really something that ought to be recognized. this is of such incredible national interest that you have the responsibility to engage it and you know exciting times ahead that was incredible thank you both so much for making time if they want to go learn more they want to figure out how to partner with the do companies call us you're kidding me you can call us yes you're two of us you're kidding we'll start with listing to this podcast
Starting point is 00:54:13 and then giving them a speed dial we'll put their numbers in the show notes jensen and vidia.com job done you heard it here we're very responsive I can attest to that. All right. Thank you so much, guys. All right, thank you, Ann. Thank you.
Starting point is 00:54:27 All right. Thank you. All right, that is all for today. If you did make it this far, first of all, thank you. We put a lot of thought into each of these episodes, whether it's guests, the calendar, Tetras, the cycles with our amazing editor, Tommy, until the music is just right. So if you like what we put together, consider dropping us a line at rate thispodcast.com slash A16C. And let us know what your favorite episode is. It'll make my day.
Starting point is 00:54:52 and I'm sure Tommy's too. We'll catch you on the flip side.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.