a16z Podcast - Jensen Huang and Arthur Mensch on Winning the Global AI Race
Episode Date: March 21, 2025The global race for AI leadership is no longer just about companies—it’s about nations. AI isn’t just computing infrastructure; it’s cultural infrastructure, economic strategy, and national se...curity all rolled into one.In this episode, Jensen Huang, founder and CEO of NVIDIA, and Arthur Mensch, cofounder and CEO of Mistral, sit down to discuss sovereign AI, national AI strategies, and why every country must take ownership of its digital intelligence.How AI will reshape global economies and GDPThe full AI stack—from chips to models to AI factoriesWhy AI is both a general purpose technology and deeply specializedThe open-source vs. closed AI debate and its impact on sovereigntyWhy no one will build AI for you—you have to do it yourselfIs this the most consequential technology shift of all time? If so, the stakes have never been higher.Resources: Find Arthur on X: https://x.com/arthurmenschFind Anjney on X: https://www.linkedin.com/in/anjney/Find NVIDIA on X: https://x.com/nvidiaFind Mistral: https://x.com/MistralAI Stay Updated: Let us know what you think: https://ratethispodcast.com/a16zFind a16z on Twitter: https://twitter.com/a16zFind a16z on LinkedIn: https://www.linkedin.com/company/a16zSubscribe on your favorite podcast app: https://a16z.simplecast.com/Follow our host: https://twitter.com/stephsmithioPlease note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details please see a16z.com/disclosures.
Transcript
Discussion (0)
This is the greatest force of reducing the technology divide the world's ever known.
It will have an impact on GDP of every country into double digits in the coming years.
Nobody's going to do this for you. You've got to do it yourself.
It's up to organizations, to enterprises, to countries, to build what they need.
The stakes at play are basically the equivalent of modern digital colonialization.
AI isn't just computing infrastructure.
cultural infrastructure.
The race for AI dominance is not only constrained to companies, but is increasingly
capturing the attention of countries.
And that includes the infrastructure spanning every layer of the stack.
The chips, the models, the applications, plus the energy required to run these, quote,
AI factories, the talent needed to produce them, and well-designed policy that helps not
hinders this entire ecosystem.
And all of this together is,
turning critical.
Set up is always hard.
This is no different.
The only question is do you need to do it?
If you want to be part of the future,
and this is the most consequential technology of all time,
not just our time, of all time,
digital intelligence, how much more valuable,
how much more important can it be?
In today's episode, we explore sovereign AI
and this regional race for AI infrastructure
across countries big and small.
And there is truly no one better to discuss this than our guests, Jensen Huang and Arthur Mench.
Jensen, of course, is the inimitable co-founder and longtime CEO of NVIDIA,
a company known for its constant reinvention and ability to place critical bets like the GPU
or graphics processing unit that has propelled it to be one of the largest companies
at over $3 trillion in market cap as of this recording.
Of course, the products that NVIDIA makes, like the GPU, are also the backbone
to so much of our digital world today.
Arthur, on the other hand, is the co-founder and CEO of Mistral,
a leading AI lab that focuses on customizable, open-source frontier models,
but also a growing number of tools to help companies and even countries engage with AI.
Today, Arthur and Jensen sit down with A16Z general partner,
Anjani Amira, as they explore the role of digital intelligence at the nation level
and how countries should think about ownership, codifying their culture,
and the role that open source should play.
All right, let's get started.
As a reminder, the content here is for informational purposes only,
should not be taken as legal, business, tax, or investment advice,
or be used to evaluate any investment or security
and is not directed at any investors or potential investors in any A16C fund.
Please note that A16D and its affiliates
may also maintain investments in the companies discussed in this podcast.
For more details, including a link to our investments,
please see A16C.com slash Disclosures.
Today we're talking about sovereign AI, all things, national infrastructure and open source.
So let's just start with the first question I usually get from nation-state leaders,
which is, is AI actually a general purpose technology?
In the history of humanity, we've had maybe a handful of these, 22, 24 economists call
these specific technologies that accelerate economic progress broadly across society.
Electricity, the printing press.
And the question everybody's asking right now, is that the right way to think about AI?
Or why isn't AI just another important, but ultimately narrow technology?
I think it's a general purpose technology because it basically revisits entirely the way
we are building software and the way we are using machines.
And so in the same way that Internet was a general purpose technology, AI is a general
purpose technology here. It allows to build agents that are doing things on your behalf. And in that
respect, it can be used in any vertical of the industry. It can be used for services, for public
services. It can be used to change the life of citizens. It can be used for agriculture. It can
obviously be used for defense purposes. So it covers everything that a state needs to worry about.
And so in that respect, it's very natural that any state makes it a priority and makes it a dedicated
nationwide strategy.
By the way,
everything Arthur said
is 100% correct.
It is also
exactly the reason
why everybody's given up
and it's precisely wrong.
And the reason for that is this.
If it's a general purpose
technology
and one company
can build the ultimate
general purpose technology
then why shouldn't
anybody else do it?
And that is the flaw.
Right.
But that's also
the mind trick
to convince everyone
that intelligence
is only something that a few people ought to go build.
Everybody ought to sit back and wait for it.
I would advise that everybody engage AI.
And it is not just a few companies in the world who should build it.
Everybody should build it.
Nobody's going to care more about the Swedish culture and the Swedish language and the
Swedish people and the Swedish ecosystem more than Sweden.
Nobody's going to care about the ecosystem of Saudi Arabia more than Saudi Arabia.
and nobody's going to care about Israel more than Israel,
despite the fact that the technology is general purpose
and absolutely true, how could intelligence not be general purpose?
It is also hyper-specialized.
And the reason for that is because, let's face it,
I don't think I'm waiting around for a general-purpose chatbot
to be an expert in a particular area of disease.
I still think that I would prefer to have somebody who is hyper-specialized in that field
to fine-tune, to train, and post-trained, if you will,
an AI model that's going to be specialized in that.
It's the general purpose technology
the same way a programming language
is the general purpose technology.
And in addition to that,
it's also a culture-carrying technology.
So I think what that means is that
there's an infrastructure.
There are chips that obviously not every country
are going to build.
There are general-purpose models
like base models, compression of the web,
that are eventually going to be open source
and that can serve as the right basis
for constructing specialized systems.
But beyond that, I think it's up to organizations,
to enterprises, to countries, to build what they need.
So the way to make it work is to take a general purpose model,
like an open source model, for instance,
and to get the knowledge you have specifically
or ask your citizens or ask your employees
to distill their knowledge into the systems
into the agents that are going to be working on your behalf
so that progressively those agents become more accurate
and following the instructions
and the specifications that the country or an enterprise may have.
So you need vertical experts
or you need cultural experts
or you need people with a certain national agenda
to partner with technological companies
that can expose the open source infrastructure
in a way that is easy to use.
use, and in a way that is easy to specialize.
So I think that's where the frontier lies.
It's a very horizontal technology.
To make anything useful out of it, you need the partnership between the horizontal
providers and the vertical experts.
But unlike previous general purpose technology waves in history, like electricity or the printing
press, how is this one different?
If I'm a nation state leader and I'm trying to understand what the right framework is for
me to think about AI in my country, should I think about it like digital labor or should
I think about it as akin to bridges?
I think it's similar to electricity in the sense that it will have an impact on GDP
of every country in the double digits in the coming years.
So that means that from an economical point of view,
every nation needs to worry about it
because if they don't manage to set up infrastructure,
to set up their own sovereign capacities at the right place,
that means that this is money that might flow back to other countries.
So that's changing the economic equilibrium across the world.
In the sense, that's not very different from electricity.
100 years ago, if you weren't building electricity factories,
you were preparing yourself to buy it from your neighbors,
which at the end of the day isn't great because it creates some dependencies.
I think in that sense, it is similar.
What is fairly different, I think there's two things.
First of all, it's kind of an amorphic technology.
If you want to create digital labor with it,
you need to shape it.
You need to have infrastructure, talent, and software.
And the talent needs to be created locally.
I think this is quite important.
And the reason for that is that,
in contrast with electricity,
this is a content-producing technology.
So you have agents that are producing content,
that are producing text,
producing images, producing voice,
interacting with people.
And when you're producing content
and interacting with society,
you become a social construct.
And in that respect,
social constructiary cultures and values
of either an enterprise or a country.
And so if you want those values
not to disappear and not to depend
on a central provider,
you need to engage with it more profoundly
that you would be to engage with electricity, for instance.
Would you agree with that, Jensen?
A couple of ways to think about it.
Your country's digital intelligence
is not likely something you would want to outsource
to a third party without some consideration.
Your digital intelligence is just now a new infrastructure for you.
Your telecommunications, your healthcare, your education,
your highways, your electricity.
This new layer is your digital intelligence.
It's your responsibility to decide how you want this digital intelligence to evolve
and whether you want to outsource it so that you could never have to worry about intelligence again,
or this is something that you feel you want to engage, maybe even control and shape into a national infrastructure.
Of course, it has all the things that author said, AI factories, infrastructure, etc.
there's another way you could think about it is your digital workforce. Now this is a new layer
and you've got to decide whether the digital workforce of your country or your company is something
that you decide to outsource, hope it evolves the way that you would like it to, or is it
something that you want to engage, maybe even decide to control and nurture and make better.
we hire general purpose employees all the time.
We hire them out of school.
Some of them are more general purpose than others.
Some of them are more intelligent than others.
But once they become our employees, we decide to onboard them, train them, guardrail them, evaluate them, continuously improve them.
We make the investment necessary to make general purpose intelligence into super,
intelligence that we could benefit from.
And so I think that that second layer, thinking about it as digital workforce, in both
cases, it contributes to the national economy.
In both cases, it contributes to social advance.
In both cases, that it contributes to the culture.
And I think that in both cases, a country needs to play a very active role in it.
And so I think it's back to your original question about sovereign AI.
how to think about it. Yes, it is definitely a general purpose technology, but you have to
decide how to shape it. Your country's digital data belongs to you. Your national library,
your history, for so long as you want to digitize it, you could make it available to everybody
in the world. You could also make it available to companies or researchers and institutions
in your own country. It belongs to you. Of course,
These are all vaporous things.
They're very soft ideas.
But it does belong to you.
And you could decide it belongs to you in the sense that this is where you came from.
You could decide how to put it to use for the benefit of your people.
And it belongs to you in the sense that it's your responsibility to shape its future.
Sovereign AI.
It's your responsibility.
There are several other types of assets that nation states fund and protect, the military, your electricity grid.
Let's say I have understood now the criticality of AI infrastructure and sovereign AI.
Do I have to now take control of every part of the stack?
So Jensen mentions, I guess, digital workforce.
And I think it's a very good analogy that you need an onboarding platform for your AI workforce,
which means you need to be able to customize the models and pour the knowledge that are sitting in your national libraries into the model
so that suddenly speaks better your language.
You need to get your systems to know about your lows.
So that suddenly the guard wells that are set
when you're deploying an AI software are compliant.
And so that onboarding platform that requires to customize,
to guardrail, to evaluate.
And then when noticing that certain things needs to be improved,
to fix things, to debug things,
that's the platform that we are building.
So being able to deploy systems that are easy to tune
and working with these platform providers
to do the custom systems.
And once the custom systems are made,
it's important to be able to maintain them yourself.
So that means being able to deploy them on your own infrastructure,
being able to ask your technological partners
to potentially disappear from the loop.
Your IT department is going to become the HR department
of your digital workforce.
And they're going to use these tools that Arthur describes
to onboard AI's fine-tune AI,
guard rail them, evaluate them, continuously improve them.
Right.
And that flywheel will be managed by the modern version of the IT department.
Right.
And we'll have biological workforce and we'll have a digital workforce.
It's fantastic.
And so nobody's going to do this for you.
You've got to do it yourself.
Right.
That's why even though we have so many technology companies in the world, every company
still has their own IT department.
I've got my own IT department.
I'm not going to outsource it to somebody else.
In the future, there'll be even more important.
to me because they'll be helping us manage these digital workforces. You're going to do this in
every country. You're going to do this in every company within those countries. And so the space
for what Arthur is describing to take this general purpose technology, but to really fine-tune it
into domain experts. They're national experts or their industrial experts or their corporate
experts or functional experts.
This is the future, the giant future space of AI.
So you both said something that I just want to make sure I'm understanding correctly.
You called it a soft concept like your culture, and you said there are a bunch of norms
that the training data has that you customize the models on.
You said norms that exactly means it's soft versus rules, which are more hard, or algorithms
and laws, which are very specific.
There's different things that you want to incorporate into your AI systems.
There are some elements of style and of knowledge
that you're not going to enforce through strict guardrails
that you can enforce through continuous training of models, for instance.
You take preferences and you distill it into the models themselves.
And then you have a set of laws,
you have a set of policies if you're in a company,
and those are strict.
And so usually the way you build it is that you connect the models
to the strict rules and you make sure that every time it answers,
you verify that the rules are respected.
on one side you're pouring and compressing knowledge in a soft way into the models
and on the other side you're making sure that you have a certain number of policies
and rules that are strictly enforced and that have a hundred percent accuracy
so on one side this is soft this is preference this is culture preference
somebody's preference is multidimensional you know what you prefer it depends it's implicit
many times in communication there's so many features that defines my preference
it takes AI to be able to precisely comply with the description that Arthur was describing just now.
Could you imagine if a human had to write this in Python, describe every one of these,
capture every one of these things in C++.
Based on this, I prefer that, but if you did that, I prefer that other thing.
And, I mean, the number of rules would be insane, which is the reason why AI has the ability to codify all of this.
It's a new programming model that can deal with the ambiguity of life.
Well, it sounds like you're saying
AI isn't just
computing infrastructure,
it's also cultural infrastructure.
Yes, it is.
Is that right?
And it's about making sure
that your cultural infrastructure
and the human expertise
that are in your company
or in your country
makes it to the AI systems.
Right.
Culture reflects your values.
We were just talking about
how each one of these
AI models,
AI services,
respond differently
to the type of questions
you're asked.
Right.
Because they codify
the values of their service
or the values of their company into each one of their services.
Could you imagine this now amplified at an international scale?
This is an inherent limitation of centralized AI models,
where you're thinking that you can encode some universal values
and some universal expertise into a general purpose model.
At some point, you need to take the general purpose model
and ask a specific population of employees or of citizens,
what are their preferences and what are their expectations?
And you need to make sure that your,
specializing the model in a soft way and in a hard way for rules and through culture and preferences.
And so that part is not something that you can outsource as a country. It's not something
that you can outsource as an enterprise. You need to own it.
Well, then is it an exaggeration to say if it is cultural infrastructure and I don't own
sovereignty of it, the stakes at play are basically the equivalent of modern digital
colonialization. If you're saying, Ange, you've got to think about AI is almost like
your digital workforce and another country or somebody who's not my sovereign nation can decide
what my workforce can and can't do. That's a problem. Some of it is universal. For example,
it is possible for certain companies to serve nations and society and companies around the world
because it's basically universal. But it cannot be the only digital intelligence layer. It has to be
augmented by something regional. You know, I think McDonald's is pretty good everywhere. All right. Kentucky
fried chicken is pretty good everywhere. But you still want the local style, local taste that augments
on the last mile. That's right. The local cafes, the mom and pop restaurants, because it defines
the culture. It defines society. It defines us. I think it's terrific that you have Walmart
everywhere that you can count on everywhere. You know, I think it's fine. But you need to have local
taste, local style, local preference, local excellence, local services. Let me swing in another way. It is very
likely that in the context of our digital workforce in the future, we will have some digital
workers, which are generic.
Right.
They're just really good at doing maybe basic research or something basic.
Good college level graduate.
Or they're useful for every company.
It's unnecessary for me to create something new.
Right.
I think Excel is pretty good.
Microsoft Office is universally excellent.
Right.
I'm perfectly fine with it.
Good reference architecture base.
That's right.
Then there's industry-specific tools,
industry-specific expertise that is really important.
For example, we use synopsis and cadence.
Arthur doesn't have to, because it's specific to our industry, not his.
We probably both use Excel, probably both use PDFs.
We both use browsers.
And so there's some universal things that we can all take advantage of,
and there'll be universal digital workers that we can take advantage of.
And then there'll be industry-specific, and then there'll be company-specific.
right inside our company we have some special skills that are very important to us that defines us
it's highly biased if you will very guardrail to doing very specific work highly biased to
the needs and the specialties of our company right and so we become superhuman in those areas
well your digital workforce is kind of be the same and AI is going to be the same there'll be
some that you just take off the shelf the new search will likely be some AI right the new research will
probably be some AI but
then there'll be industrial versions of AIs that we'll maybe get from Cadence and others,
and then we'll have to groom our own using Arthur's tools.
Right.
And we'll have to fine tune them, we'll onboard them, make them incredible.
I very much agree with this vision of having a general purpose model and then some layer of
specialization for industries and then an extra layer of specialization for companies.
You will have a tree of AI systems that are more and more specialized.
And maybe to give a concrete example with what we recently did, so we released in January
a model called Mistral Smol, and it's a general purpose model.
So it speaks all of the languages, it knows mostly about most things.
But then what we did is that we took it and we started a new family of specialized models
that were specialized in languages.
So we took more languages in Arabic, more languages in Indian languages, and we retrained
the model.
And so we distil this extra knowledge that the initial model hadn't seen.
And so in doing that, we actually made it much, much better in being idiomatic when it speaks
Arabic and when it speaks languages from the Indian peninsula. And so language, it's probably
like the first thing you can do when you're specializing a model. The good thing is that
for a given size of model, you can get a model that is much better if you choose to specialize
it in a language. So today, our model, it's the 24B, it's called Mistral Saba, it's a model
tune in Arabic, is outperforming every other language model that are like five times larger. And
the reason for that is that we did this specialization. And so that's the first layer. And then
If you think of a second layer, you can think of verticals.
So if you want to build a model which is not only good at Arabic, but also good at handling
legal cases in Saudi Arabia, for instance, well, you need to specialize it again.
So there's some extra work that needs to be done in partnership with companies to make sure
that not only your system is good at speaking a certain language, but it's good at speaking
a certain language and understanding the legal work that is done in this language.
And so it's true for any combination that you can think of, of vertical and language.
I see.
You want to have a medical diagnosis assistant in French.
Well, you need to be good at French, but you also need to understand how to be good at speaking the French language of physicians.
And so those two things, it's very hard to do as a general purpose model provider.
If this is true and what you're describing is real, that I need the capabilities to customize this AI layer on my local norms, my local data.
which is fairly sophisticated from a technical capability perspective.
How would you advise a big nation to think about the stack we're talking about,
the chips, the compute, the data center, the models that sit on top the applications,
and then ultimately what you were describing as the AI nurse or the AI doctor?
And how would you advise someone that's a smaller nation differently?
I would say you need to buy and to set up the horizontal part of the stack.
So you need the infrastructure.
You need the inference primitives, you need the customization primitives, you need the observability,
you need the ability to connect visuals to models, to connect models to sources of information,
of real-time information.
Those are primitives that are fairly well factorized across the different countries, across the
different enterprises.
And once you have that, these are things that can be bought.
Then you can start working.
Then you can start building.
You build from these primitives according to your values, according to your expertise, and
thanks to your local talent.
The question is, where is the frontier
and between what is horizontal
and horizontal,
if you're a small enterprise
or a small country,
you should probably buy.
And what is vertical and specific
to you, and that's definitely
something that you need to build.
You have to get it in your head
that it's not as hard as you think it is.
First of all,
because the technology is getting better,
it's easier.
Could you imagine doing this
five years ago?
It's impossible.
Could you imagine doing this
five years from now?
It'll be trivial.
And so we're somewhere in that middle.
The only question is, do you have to do it?
Right.
The truth of the matter is, I hate onboarding employees.
And the reason for that is because it takes a lot of work.
But once you set up an HR organization and leadership mentoring organization and processes,
then your ability to onboard employees is easier and is systematically more enjoyable for everybody involved.
But in the very beginning is hard.
Set up is always hard.
Set up is always hard.
This is no different. The only question is, do you need to do it? If you want to be part of the future,
and this is the most consequential technology of all time, not just our time, of all time.
Digital intelligence, how much more valuable, how much more important can it be? And so if you
come to the conclusion this is important to you, then you have to engage it as soon as you can,
learn along the way, and just know that it's getting easier and easier all the time. The fact that
the matter is if we try to do agentic systems even three years ago, it was incredibly hard.
Right.
But agentic systems are a lot easier today.
And all of the tools necessary for curating data sets, for onboarding the digital employees to
evaluating the employees, the guard railing, digital employees, all of those are getting
better all the time.
The other thing about technology is when it becomes faster, it's easier.
Right.
Could you imagine back in the old days?
Of course, I had the benefit of seeing computers from its earliest days.
and the performance of the computers
were so frustratingly slow.
Everything you did was hard.
But these days,
the type of things we do is just magical
because it's also fast.
And so whether it's motivated
by your institutional need
to engage the most consequential technology
of all time
or the fact that it's getting better
all the time, so it's not that hard,
I think the number of excuses is running out.
So let's talk about that for a second
because change is hard. I've got an endless list. If I'm a nation state leader, I'm facing
increasing amounts of geopolitical risk. I don't know who my allies are. Elections are coming.
There's any number of things I've got to deal with. But now let's say I understand that this is
important. You guys spend so much time talking nation state leaders who are thinking about what are
the risks of adopting AI too fast. And you're right, the zeitgeist has shifted based on the Paris
Action Summit. It seemed like there's a tone of optimism more than there was a tone of pessimism a year ago.
but what are the most common questions you get from nation state leaders
when they're asking about risks and how to think about them?
So I've heard several questions,
but one of the risk is to see your population start getting afraid of the technology
for fear of it replacing them.
And that is something that can actually be prevented.
If we collectively make sure that everybody get access to the technology
and is trained in using it.
So the skilling of the various citizens of the populations
is extremely important.
And stating AI as an opportunity for them to actually work better
and showing the purpose of it through applications,
through things that they can actually install on their smartphone,
through public services.
We're working, for instance, with the French unemployment system
to actually connect opportunities of jobs to unemployed people through AI agents
that are being actionated by, obviously, human operators within the agency.
And that is one opportunity.
That's a very palatable opportunity for people to find a job better.
And so that's part of the thing that can make sure that population understand the opportunity
and the fact that AI is really just a new change for them to adopt,
just the same way they had to adopt personal computers in the 90s and Internet in the 2000s.
The common aspect with these changes is that you need people to embrace the technology.
And I think the biggest problem that nation states may have is to
see AI increase the digital divide
that is already relatively big
but if we work together and if done in the right way
we can make sure that AI is actually reducing the digital divide.
AI is a new way to program a computer.
It is because by typing in some words
you can make the computer do something.
Just like we did in the past.
Right.
I know you talk to it.
You can interact with it in a whole lot of ways.
You can make the computer do things for you
a lot easier today
than it was before.
Right.
The number of people
who could prompt chat GPT
and do productive things
just from a human potential perspective
is vastly greater
than the number of people
who can program C++ ever.
And therefore,
we have closed the technology divide.
Probably the greatest equalizer we've seen.
It is by definition.
Right.
The greatest equalizer
of technologies of all time.
But you still need,
to have citizens to know about it.
I think that's the thing.
I'm just describing the fact.
Yes.
The fact is there are more people
who program computers using chat GPT today
than there are people who program computers using C++.
Right.
That's a fact.
And so the fact is this is the greatest force
of reducing the technology divide the world's ever known.
Right.
It's just perceived, and what Arthur's saying,
the perception through,
I don't know who
and I'm talking about it
and I don't know how
talking about it
but the fact that the matter is
it is not stopping
right
it's not stopping anything
the number of people
who are actively using
chat GPT today
is off the charts
I think it's terrific
it's completely terrific
anybody who's talking about
anything else
apparently isn't working
and so I think
people realize
the incredible capabilities
of AI and how it's
helping them with their work
I use it every single day.
I used it this morning.
And so every single day I use it.
And I think that deep research is incredible.
Right.
My goodness, the work that Arthur and all of the computer scientists around the world are doing is incredible.
And people know it.
People are picking it up, obviously, right?
Just the number of active users.
Let's talk about open source for a bit.
Because both of you have talked quite publicly about the importance of open models in the context of sovereignty.
At DeepMind, you're part of the Chinchilla skating laws, which were openly
published, your co-founder, Guillaume, created Lama. And then last year,
NVIDIA and Mistral worked on a jointly trained model called Mistral Nemo. Why are open models
such a big part of your focus? Because it's an horizontal technology and enterprises and states
are going to be eventually willing to deploy it on their own infrastructure. Having this
openness is important from a sovereignty perspective. That's the first point. And then the
second point of importance is that releasing open source models is a way.
to accelerate progress and we created Mistral on the basis that what we've seen during our early
career when we were doing AI in between 2010 and 2020 was an acceleration of progress because
every lab was building on top of each other. And that's something that kind of disappeared with
the first large language models from open AI in particular. And so spinning back that open
fly wheel of I contribute something and then another lab is contributing something else. And then
And we iterate from that is the reason why we created Mistral.
And I think we did a good job at it because we started to release models and then META started to release models as well.
And then we had Chinese company like Deep Seek release stronger models.
And everybody benefited from it.
So coming back to Mistral Nemo, one difficulty of creating AI models in an open way is that this is more a cathedral than a bazaar setting when it comes to open source.
Because you have large spend to do to build a model.
And so what we did with Nvidia team is really.
to mix the two teams together,
have them work on the same infrastructure,
the same code, have the same problems,
and combine their expertise to build the same model.
And that has been very successful
because Nvidia brought a lot of things we didn't know.
I think we brought things that Nvidia didn't know.
And at the end of the day,
we produced something that was at the time
the best model for its size.
And so we really believe in such collaborations
and we think that we should do them more
and at a higher scale.
And not only with only two companies,
but probably with three or four.
And that's the way open source is going to prevail.
I completely agree.
The benefit of open source, in addition to accelerating and elevating the basic science, the basic endeavor of all of the general models and the general capabilities, is the open source versions also activate a ton of niche markets and niche innovation.
All of a sudden, in health care, life science,
sciences, physical sciences, robotics, transportation, the number of industries that were activated
as a result of open source capabilities are sufficiently good is incredible. Don't ignore
the incredible capabilities of open source, particularly in the fringe, the niche.
But mission critical where data might be sensitive. Yeah, it could be, for example, in mining
energy. Right. Who's going to go create an AI company to go mine energy? Energy is really important,
but the mining of energy is not that big of a market.
And so open source activates every single one of them.
Financial services, it turns out, activates them.
Healthcare, defense.
You pick your favorites.
Anything that is mission critical and that requires to do one's own deployment,
that potentially requires to do on the edge deployment as well.
Right.
And anything that requires some strong auditing
and the ability to do a thorough evaluation of it.
You can evaluate a model much better if you have access to the weights
and if you only have access to APIs.
And so if you want to build certainty around the fact that your system is going to be 100% accurate,
I don't think you should be using a close source model.
And you have to connect it into your flywheel.
How are you going to connect your local data?
Yeah, you have to connect it into your own, your local data, your own local experience.
The more you use it, the better it becomes that flywheel.
You can't do it without open source.
But let's say I'm a nation state leader.
I've been considering open source.
I'm starting to hear things like, hey, open source is a threat to national security.
we should not be exporting our models
because these open models actually give away
a ton of nation-state secrets
or more importantly, the bad guys can use these open models too.
And so this is a threat to security.
Instead, what we should be doing is locking down
maybe development amongst two or three labs
that have the infrastructure to get licenses from the government
to do training, to do the right safety and certification.
I've certainly been hearing that a lot.
How should I think about that versus what you're telling me,
which is actually no open is better
for mission-critical industries?
collaboration in between labs is going to be critical for humanity success
and if one state decides to lock things down
the only thing that is going to happen is that another state will take the leadership
because cutting yourself from the open fly wheel
has just too high of a cost for you to maintain competitivity if you do that
this is a debate that has occurred in the United States
and effectively if there's some export control over weight
this is not going to stop any country in Europe
any country in Asia to continue its progress.
And they will collaborate
to actually accelerate that progress.
So I think we just need to embrace the fact
that this is an horizontal technology
very similar to programming languages.
Programming languages, they are all open source, right?
So I think AI just needs to be open source in that respect.
We're glad to see that this realization
that we could accelerate together
by being more open about the way we build the technology.
And so it's great to see
that open source has a lot of good days
before it. It is
impossible to control. Software is
impossible to control. If you want to control it
then somebody else's will emerge and become the standard
just as Arthur mentioned.
And the question is, is open source safer?
Open source enables more transparency, more researchers,
more people to scrutinize the work.
The reason why every single company in the world
is built, every cloud service provider is built on open source is because it is the safest technology
of all. Give me an example of a public cloud today that's built on an infrastructure stack
that isn't open source. You start from open source, you could customize it. But the benefit of
open source is the contribution of so many people and the scrutiny, very importantly,
You can't just put any random stuff into open source.
You'll get laughed off the internet.
You've got to put good stuff on the open source because the scrutiny is intense.
And so I think open source provides all of that.
Great collaboration to accelerate innovation, escalate excellence, ensure transparency,
attract scrutiny, all of that improves safety.
In a sense, you're saying it's partly more secure because as we've seen with open source databases,
storage, networking, compute,
you get mass red-teaming.
The whole world can help you red-team your technology
versus just a small group of researchers
inside your company.
Is that roughly, right?
A right way to think about it?
Exactly.
By pulling a lot of organizations together
to come up with a technology
that they can all use
and specialize on their own domains,
you're forcing the technology
to be good for every one of them.
And so that means you're removing biases.
You're really making sure
that the general purpose models
that you're building
are as good as possible and don't have failures.
And I think open source in that respect
is also a way to reduce the number of failure points.
If as a company today, I decide to rely fully on a single organization
and on its safety principles,
on its red-teaming organization as well,
I'm trusting it a little too much.
Whereas if I'm building my technology on open-source models,
I'm trusting the world to make sure that the basis on which I'm building
is secure.
So that's a reduction of failure points.
And it's obviously something that you need to do as an enterprise or as a country.
We're going to transition a little bit now into company building,
which is something a lot of people are excited to hear from both of you about.
So let's start with you, Jensen.
You've remarked that NVIDIA is the smallest big company in the world.
What enabled you to operate that way?
Our architecture was designed for several things.
It was designed to adapt well in a world of change,
either caused by us or affecting us.
And the reason for that is because technology changes fast.
And if you over-correct on controllability,
then you are underserving a system's ability to become agile and to adapt.
And so our company uses words like aligned instead of use words like control.
I don't know that one time I've used the word control in talking about the way that the company works.
We care about minimum bureaucracy, and we want to make our processes as lightweight as possible.
Now, all of that is so that we can enhance efficiency, enhance agility, and so on and so forth.
We avoid words like division.
When Nvidia was first started, it was modern to talk about divisions.
Right.
And I hated the word divide.
Why would you create an organization that's fundamentally divided?
I hated the word business units.
The reason for that is because why should anybody exist as one?
Why don't you leverage as much of the company's resources as possible?
I wanted a system that was organized much more like a computing unit, like a computer,
to deliver on an output as efficiently as possible.
And so the company's organization looks a little bit like a computing stack.
And what is this mechanism that we're trying to create?
And in what environment are we trying to survive in?
Is this much more like a peaceful countryside or is this like much more like a concrete jungle?
What kind of environment are you in because the type of system you want to create should be consistent with that?
And the thing that always strikes me odd is that every company's org chart looks very similar, but they're all different things.
One's a snake, the other one's a elephant, the other one's a cheetah.
and everybody is supposed to be somewhat different in that forest,
but somehow they all get along.
Same exact structure, same exact organization,
doesn't seem to make sense to me.
I agree that it feels like companies have personalities,
and despite the fact that they're organized sometimes similarly,
I should say that obviously we have a lot of things to learn,
and I mean, the company is not even two years old.
I guess one challenge we have with Mistrial,
and I think our competitors have the same challenge,
is that this is one of the first time
that the software company
is actually a deep tech company
that is driven by science.
Science doesn't have the same timescales as software.
You need to operate on a monthly basis.
Sometimes you don't know exactly when the thing will be ready.
But on the other hand, you have customers asking,
when is the next model coming up?
When is this capability going to be available, et cetera?
And so you need to manage expectation.
And I think for us, the biggest challenge,
and I think we're starting to do a good job at it,
is to manage the hinge in between the product requirements
and what the science is able to do.
Research and product.
Research and product.
And you don't want the research team
to be fully dedicated to making the product work.
So you need to work, and I think we've started to do it,
on making sure that you have several frequencies in your company.
You have fast frequencies on the product side,
iterating every week.
And you have slow frequencies on the science side
that are looking at why profoundly the product is failing on certain domains
and how they could fix it through research,
through new data, through new architecture, through new paradigm.
And I think that's fairly new.
This is not something that you would find in a typical SaaS company
because this is inherently a science problem.
I mean, Nvidia is one of the most successful companies
that over a 30-year timeline has figured out a way
to keep science and research ahead of the rest of the world,
whether it was CUDA back in
2012, where those fundamental systems research, or Cosmos today, which is now saying, you know,
it's definitely state of the art on how simulation should work out.
We've harmonized exactly what Arthur just said.
Is that heuristic right for you?
Yeah, we harmonized that inside our company.
We have basic research, applied research, and then we have architecture, and then we have product
development, and we have multiple layers of it.
And these layers are all essential.
Right.
And they all have their own time clock in the case of basic research.
the frequency could be quite low.
On the other hand, all the way to the product side,
we have a whole industry of customers
who are counting on us.
And so we have to be very precise.
And somewhere between basic research
and discovering hopefully surprises
that nobody expects.
On the one hand, on the other hand,
to be able to deliver on what everyone expects.
Predictably.
These two extremes, we manage harmoniously inside our company.
There's so many fascinating things
about this market, but there's one in particular that I want to call out. Both of you have
customers that are also your competitors. And those competitors are huge and highly capitalized
tech giants. Nvidia sells GPUs to AWS, which is building its own chips called Traneum.
And Arthur, your training models that you sell through AWS and Azure who have funded
labs like Anthropic and Open AI. So how do you win an environment like this and how do you
manage those relationships? Because we talked about company building internally, but now I'm curious
externally. How do you survive?
Jensen said it well. You give up
control but you work on alignment
and despite the fact that sometimes
you have certain companies can be competitors
you may have a line interests
and you can work
on specific agendas that are shared.
You have to have your own place.
Obviously these cloud service providers
aren't working with Arthur because
they already have the same thing. They just want two of the same
things. It's because Arthur and
Mistral has a position in the world
that is unique to Mistral, and they add value in a particular place that is unique.
A lot of the conversation we've had today are areas that Mistral and the work and their position
in the world makes them uniquely good at. And we are different. We're not just another ASIC.
We can do things for the CSPs and do things with the CSPs that are not possible for them
to do themselves. For example, NVIDE's architecture is in every cloud.
And in a lot of ways, we have their first onboarding for amazing future startups.
And the reason for that is because by onboarding to Nvidia, they don't have to make a
strategic or business or otherwise commitment to a major cloud.
They could go into every cloud.
And they could even decide to build their own system they like because the economics turns
out to be better for them at some point or they would like access to capabilities that we
have that are somewhat protective within the clouds.
And so whatever the reasons are, in order to be a good partner to somebody, you still have to have a unique position.
You need to have a unique offering.
And I think Ms. Straw has a very unique offering.
We have a very unique offering.
And our position in the world is important to even the people we compete against.
And so I think when we are comfortable within that and comfortable with our own skin, then we can be excellent partners to all of the CSPs.
And we want to see them succeed.
I know that it's a weird thing to say when you see them as a competitor, which is the reason we don't see them as a competitor.
We see them as a collaborator who happens to compete with us as well.
And probably the single most important thing that we do for all the CSPs is bring them business.
And that's what a great computing platform does.
We bring people business.
I remember when Arthur and I first met, we sat down in London at a late night restaurant and sketched out the plan for his Series A.
And we were figuring out why he needed so much capital for the Series A, which in hindsight was remarkably efficient.
I think the Mistral Series A we put together
was half a billion
relative to other folks
who had to spend
multiple billions to get to the same place
but I asked him
what chips would you like to run on?
And you looked at me so absurdly
as if I had asked you a question
that how could it be an answer
other than Nvidia
other than H-100s
and I think that ecosystem
has been
the startup ecosystem
that Nvidia has invested in
creates so much business
for the clouds
what is the philosophy
that led you to invest
so deeply in startups
and founders so early on, even before anybody knew about them.
Two reasons, I would say.
One, the first reason is I rarely call us as a GPU company.
What we make is a GPU, but I think of Nvidia as a computing company.
If you're a computing company, the most important thing you think about is developers.
Right.
If you're a chip company, the most important thing you think about is a chip.
And all of our strategies, all of our actions, all of our priorities, all of our focus, all of our investments, 100% of it is aligned with the
attitude that is developer first.
It's about the computing platform first, another way of saying ecosystem.
Right.
And so everything starts there.
Everything ends there.
GTC is a developers conference.
All of our initiatives inside the company is developer first.
So that's number one.
The second thing is we were pioneering a new computing approach that was very alien to the world
of general purpose computing.
And so this accelerated computing approach was rather alien and
counterintuitive and rather awkward for a very long time. And so we're constantly seeking out
looking for the next incredible breakthrough, the next impossible thing to do without accelerated
computing. And so it's very natural that I would find and would seek out researchers and
great thinkers like Arthur because, you know, I'm looking for the next killer app. And so that's
kind of a natural intuition, natural instinct of somebody who is creating something new. And so
If there's an amazing computer science thinker that we haven't engaged with, that's my bad.
We've got to get on it.
That's a perfect segue from a computing perspective.
What are the most significant trends you see on the horizon?
And in particular, for an audience who might be prime ministers, a president, so ministers of IT in some of the world's fastest growing markets, trying to understand where computing is going.
How would you guide them?
We're moving toward workloads that are more and more asynchronous.
So workloads where you give a task to an AI system and then you, you, you, you.
wait for it to do 20 minutes of research
before returning. So that's definitely
changing a bit the way you should
be looking at infrastructure because that creates
more load. So I guess it's a
bull case for data centers and for Nvidia.
As I've said, I guess, in the beginning of this episode,
all of this is not going
to happen well if you don't have the right
onboarding infrastructure for the agents.
If you don't have a proper way
for your AI systems to learn about
the people they interact with
and to learn from the people they interact with.
So that aspect of
Learning from human interaction is going to be extremely important in the coming years.
And there's another aspect which is around personalization of having, I guess, models and systems
consolidate the representation of their users to be as useful as possible.
I think we are in the early stage of that.
But that's going to change, again, pretty profoundly the interaction we have with machines
that will know more about us and know more about our taste and how to be as useful as possible
toward us. As a leader of a country, I want to think about education, about making sure that I have a local talent pool that understands AI enough to create specialized AI systems. And I want to think about infrastructure, both on the physical side, but also on the software side. So what are the right primities, what is the right partner to work with that is going to provide you with the platform of onboarding. And so those two things are important. If you have this and you have the talent, and if you do deep partnerships,
the economy of your state is going to be profoundly changed.
The last 10 years, we've seen extraordinary change in computing.
From hand coding to machine learning, from CPUs to GPUs, from software to AI,
across the entire stack, the entire industry has been completely transformed.
And we're going through that still.
The next 10 years is going to be incredible.
Of course, the industry has been wrapped up in talking about scaling laws.
And pre-training is important, of course, and continues to do.
be. Now we have post-training. And post-training is thought experiments and practice and tutoring and
coaching and all of the skills that we use as humans to learn the idea that thinking and agentic
and robotic systems are now just around the corner. It's really quite exciting. And so what it
means to computing is very profound. People are surprised that Blackwell is such a great leap over
Hopper. And the reason for that is because we built Blackwell for inference. And just in time, because
all of a sudden, thinking is such a big computing load. And so that's one layer, is there's
a computing layer. The next layer is the type of AIs that we're going to see. There's the
agentic AI, the informational, digital worker AIs. But we now have physics AI that's making
great progress. And then there's physical AI that's making great progress. And physics AI is, of course,
things that obeyed the physical laws and the atomic laws and the chemical laws and all of the various
physical sciences that are going to see some great breakthroughs, and I'm very excited about that.
That affects industry, that affects science, affects higher education and research, and then physical AI,
AI that understand the nature of the physical world, from friction to inertia, the cause and effect,
object permanence, those kind of basic things that humans have common sense, but most AIs don't.
And so I think that that's going to enable a whole bunch of robotic systems that are going to have great implications in manufacturing and others.
The U.S. economy is very heavily weighted on knowledge workers.
And yet many of the other countries are very heavily weighted on manufacturing.
And so I think for many of the prime ministers and the leaders of countries to realize that the AIs that they need to transform and to revolutionize their industries that are so vital to them,
whether it's energy focused or manufacturing focused,
it's just around the corner,
and they ought to stay very alert to this.
I would encourage people not to over-respect the technology.
And sometimes when you over-admire a technology,
over-respect the technology, you don't end up engaging it.
You're afraid of it somehow.
Some of the things that we said today about AI closing the technology divide
is really something that ought to be recognized.
this is of such incredible national interest
that you have the responsibility to engage it
and you know exciting times ahead
that was incredible thank you both so much for making time
if they want to go learn more they want to figure out how to partner with the do companies
call us you're kidding me you can call us yes
you're two of us you're kidding we'll start with listing to this podcast
and then giving them a speed dial we'll put their numbers in the show notes
jensen and vidia.com job done
you heard it here we're very responsive
I can attest to that.
All right.
Thank you so much, guys.
All right, thank you, Ann.
Thank you.
All right.
Thank you.
All right, that is all for today.
If you did make it this far, first of all, thank you.
We put a lot of thought into each of these episodes, whether it's guests, the calendar, Tetras, the cycles with our amazing editor, Tommy, until the music is just right.
So if you like what we put together, consider dropping us a line at rate thispodcast.com slash A16C.
And let us know what your favorite episode is.
It'll make my day.
and I'm sure Tommy's too.
We'll catch you on the flip side.