Lex Fridman Podcast - #494 – Jensen Huang: NVIDIA – The $4 Trillion Company & the AI Revolution
Episode Date: March 23, 2026Jensen Huang is the co-founder and CEO of NVIDIA, the world’s most valuable company and the engine powering the AI computing revolution. Thank you for listening ❤ Check out our sponsors: https...://lexfridman.com/sponsors/ep494-sc See below for timestamps, transcript, and to give feedback, submit questions, contact Lex, etc. Transcript: https://lexfridman.com/jensen-huang-transcript CONTACT LEX: Feedback – give feedback to Lex: https://lexfridman.com/survey AMA – submit questions, videos or call-in: https://lexfridman.com/ama Hiring – join our team: https://lexfridman.com/hiring Other – other ways to get in touch: https://lexfridman.com/contact EPISODE LINKS: NVIDIA: https://nvidia.com NVIDIA on X: https://x.com/nvidia NVIDIA AI on X: https://x.com/NVIDIAAI NVIDIA on YouTube: https://youtube.com/@nvidia NVIDIA on Instagram: https://www.instagram.com/nvidia/ NVIDIA on LinkedIn: https://www.linkedin.com/company/nvidia/ NVIDIA on Facebook: https://www.facebook.com/NVIDIA/ NVIDIA on GitHub: https://github.com/NVIDIA Nemotron: https://developer.nvidia.com/nemotron SPONSORS: To support this podcast, check out our sponsors & get discounts: Perplexity: AI-powered answer engine. Go to https://perplexity.ai/ Shopify: Sell stuff online. Go to https://shopify.com/lex LMNT: Zero-sugar electrolyte drink mix. Go to https://drinkLMNT.com/lex Fin: AI agent for customer service. Go to https://fin.ai/lex Quo: Phone system (calls, texts, contacts) for businesses. Go to https://quo.com/lex OUTLINE: (00:00) – Introduction (00:26) – Sponsors, Comments, and Reflections (06:34) – Extreme co-design and rack-scale engineering (09:20) – How Jensen runs NVIDIA (28:41) – AI scaling laws (43:41) – Biggest blockers to AI scaling laws (45:25) – Supply chain (47:20) – Memory (53:25) – Power (58:45) – Elon and Colossus (1:02:13) – Jensen’s approach to engineering and leadership (1:07:38) – China (1:15:51) – TSMC and Taiwan (1:21:06) – NVIDIA’s moat (1:26:43) – AI data centers in space (1:30:31) – Will NVIDIA be worth $10 trillion? (1:40:40) – Leadership under pressure (1:54:26) – Video games (2:01:18) – AGI timeline (2:03:31) – Future of programming (2:17:02) – Consciousness (2:23:23) – Mortality PODCAST LINKS: – Podcast Website: https://lexfridman.com/podcast – Apple Podcasts: https://apple.co/2lwqZIr – Spotify: https://spoti.fi/2nEwCF8 – RSS: https://lexfridman.com/feed/podcast/ – Podcast Playlist: https://www.youtube.com/playlist?list=PLrAXtmErZgOdP_8GztsuKi9nrraNbKKp4 – Clips Channel: https://www.youtube.com/lexclips
Transcript
Discussion (0)
The following is a conversation with Jensen Huang, CEO of Invida,
one of the most important and influential companies in the history of human civilization.
Invita is the engine powering the AI revolution,
and a lot of its success can be directly attributed to Jensen's share force of will
and his many brilliant bets and decisions as a leader, engineer, and innovator.
And now a quick few second mention of his sponsor.
Check them out in the description or at Lex Friedman.com slash sponsors.
It is, in fact, the best way to support this podcast.
We got Shopify for selling stuff online,
Element for Electrolites, Finn for customer service AI agents,
quo for a phone system like calls, texts, contacts for your business,
and perplexity for curiosity-driven knowledge exploration.
Choose wise to my friends.
And now onto the full ad reads.
I try to make them interesting, but if you skip,
Please still check out our sponsors.
I enjoy their stuff.
Maybe you will too.
To get in touch with me, for whatever reason,
go to Lexfremant.com slash contact.
All right.
Let's go.
This episode is brought to you by Shopify,
a platform designed for anyone to sell anywhere
with a great-looking online store.
Now, I know it's an incredible platform for selling stuff.
It's a mechanism by which you can buy stuff on the internet.
But the thing I like to celebrate is engineering.
They just recently tweeted about,
SimGim, which runs simulated shopping sessions by the hundreds of thousands daily.
I personally love the idea that things at scale, especially now with the LLM models, can be simulated.
You basically want to be simulating human behavior, human decision making, human choice.
In this particular context, of course, is shopping.
It's really fascinating.
And they describe in their blog posts how they're leveraging Nvidia stacks.
to accomplish this task.
But you should know, in general,
that you can sign up for a $1 per month trial period
at Shopify.com slash Lex.
That's all lowercase.
Go to Shopify.com slash Lex to take your business
to the next level today.
This episode is also brought to you by Element,
my daily zero-sugar, delicious electrolyte mix.
That, as far as I know,
has very little to do with artificial intelligence
and GPUs and CPCUs and CPCUs.
be used in the revolution that we're experiencing in the tech sector.
And I think that's beautiful because I got a chance to train a bunch of world-class
fighters, wrestlers, grapplers recently.
I'm going to be traveling to parts of the world that doesn't really have much.
And I think in those parts of the world is where the mind can reconnect with the things
that are truly important that are truly timeless.
Anyway, in those parts of the world, I often get...
pretty out there in terms of physical strain and diet and dehydration and so on.
So elements, one of the crucial things in my bag.
Really, water and salt.
And really nice, delicious, well-balanced salt, meaning sodium, potassium, magnesium, electrolytes.
Element is my go-to.
Watermelon salt, my favorite flavor.
Get a free eight-count sample pack with any purchase.
try it to drinkelement.com slash Lex.
This episode is also brought to you by Finn,
a powerful AI system that focuses on customer service.
It's trusted by over 6,000 companies.
It has a 65 average resolution rate
and is built to handle complex,
multi-step queries like returns, exchanges, and disputes.
This is such a fascinating problem.
Because customer problem,
the bulk of them fall into a very specific set of categories,
but there's nuanced details within those categories
that make all the difference.
And it can be an incredibly frustrating thing for a human being,
like myself, I swear, I promise.
Definitely not a robot.
Wouldn't tell you if I was.
But it's frustrating for a human
to come to the customer service process
and to know that your problem kind of is like this problem,
there's all these details.
can provide about the system you're operating on the specifics of the puzzle you're trying to
solve. But there's details that you just know in your gut that this is important, especially if
you kind of thought through the problem. I've been to there is quite a bit. You want to have some
level of personalization that can get to the tricky aspect, the perspective on the problem that
really would lead you down the road to a solution. Anyway, love this problem. Really glad Finn is focusing
on it. Go to fin.com. A.I. slash Lex.
to learn more about transforming your customer service and scaling your support team.
That's fin.aI slash Lex.
This episode is also brought to you by Quo, spelled QUO,
also known as a company with just three letters,
will win you a game of Scrabble.
That is not a joke.
It feels like a joke I have made before,
but let's run with it.
It's a dad joke.
It's a bad dad joke.
The only,
The only thing worse than a dad joke is a bad dad joke.
But here we go.
Spell Q-U-O, a business phone platform for calling and messaging.
Basically, you have a bunch of people trying to help a larger group of people,
and you want to orchestrate how they communicate with each other.
And this is just the system that does it extremely well, period.
Quo integrates AI into the whole shabang, organizing everything, generating summaries,
highlighting the next steps, all that kind of stuff. It just does it well. The interface on top of the
AI is also really strong. So try Quo for free, plus get 20% off your first six months when you go to
quo.com slash Lex. That's QUO.com slash Lex. This is the Lex Friedman podcast, and now, dear friends,
here's Johnson Huang. You've propelled Nvidia into a
new era in AI, moving beyond its focus on chip scale design to now rack scale design.
And I think it's fair to say that winning for Nvidia for a long time used to be about
building the best GPU possible. And you still do. But now you've expanded that to extreme
co-design of GPU, CPU, memory, networking, storage, power, cooling, software, the rack itself,
the pod that you've announced, and even the data center. So let's talk about extreme code
design? What is the hardest part of co-designing a system with that many complex components and design
variables? Yeah, thanks for that question. So first of all, the reason why extreme code design is
necessary is because the problem no longer fits inside one computer to be accelerated by one GPU.
The problem that you're trying to solve is you would like to go faster than the number of
computers that you add. So you added, you know, 10,000 computers, but you would like it to go a
million times faster. Then all of a sudden, you have to take the algorithm, you have to break up the
algorithm, you have to refactor it, you have to shard the pipeline, you have to shard the data,
you have to shard the model. Now all of a sudden, when you distribute the problem this way,
not just scaling up the problem,
but you're distributing the problem,
then everything gets in the way.
This is the Amdahl's Law problem,
where the amount of speed up you have for something
depends on how much of the total workload it is.
And so if computation represents 50% of the problem,
and I sped up computation infinitely, like a million times,
you know, I only sped up the total workload
by a factor of two.
Now all of a sudden,
not only do you have to distribute
computation, you have to
shard the pipeline somehow,
you also have to solve the networking
problem because you've got
all of these computers are all connected
together. And so
distributed computing at the scale that we
do,
the CPU is a problem, the GPU
is a problem, the networking is a problem,
the switching is a problem,
and distributing the workload across
all these computers are a problem.
It's just a massively complex computer science problem.
And so we just got to bring every technology to bear.
Otherwise, we scale up linearly or we scale up based on the capabilities of Moore's law,
which has largely slowed because Dernard's scaling has slowed.
I'm sure there's tradeoffs there.
Plus, you have a complete disparate disciplines here.
I'm sure you have specialists in each one of these,
high bandwidth memory.
the network and the NV link, the NICs, the optics and the copper that you're doing,
the power delivery, the cooling, all of that.
I mean, there's like world experts in each of those.
How do you get them in a room together to figure out?
That's why my staff is so large.
What's the, can you take me through the process of the specialists and the journalists,
like, how do you put together the rack when you know the set of things you have to
shove into Iraq together?
Yeah.
Like, what does that process look like of designing it all together?
There's the first question, which is, what is Extreme Co-designed?
you're you were optimizing across the entire stack of software from architectures to chips to systems
to system software to the algorithms to the applications that's one layer the second thing that you
and i just talked about is goes beyond CPUs and GPUs and networking chips and scale up switches and
scale out switches and then of course you've got to include power and cooling and all of that
because you know all these computers are extremely extremely power power high
hungry. They do a lot of work and they're very energy efficient, but they in aggregate still
consume a lot of power. And so that's one, the first question is what is it? The second question is
why is it? And we just spoke about the reason, you know, you want to distribute the workload so
that you can exceed the benefit of just increasing the number of computers. And then the third question
is, how is it? How do you do it? And that's the, that's kind of the miracle.
of this company. You know, when you're designing a computer, you have to have operating system of
computers. When you're designing a company, you should first think about what is it that you want
the company to produce? You know, I see a lot of companies' organization charts, and they all look the same.
Hamburger organization charts, soft organization charts, and car company organization charts,
they all look the same. And it doesn't make any sense to me. You know, the goal of a company is to be the
machinery, the mechanism, the system that produces the output, and that output is the product
that we'd like to create. It is also designed the architecture of the company should reflect
the environment by which it exists. It almost directly says what you should do with the organization.
My direct staff is 60 people. You know, I don't have one-on-ones with them because it's impossible.
You can't have 60 people on your staff if you're, you know, going to get work done and
So you still have 60 reports.
You still have, of course.
More, yeah.
And most stars at least have a foot in engineering.
Almost all of them.
There's experts in memory.
There's experts in CPUs.
There's experts in optical.
That's incredible.
Yeah, GPUs and architecture, algorithms, design.
So you constantly have an eye in the entire stack.
And you're having intense discussions about the design of the entire stack.
And no conversation is ever one person.
That's why I don't do one-on-ones.
We present a problem and all of us attack it.
You know, because we're doing extreme code design.
And literally the company is doing a stream code design all the time.
So even if you're talking about a particular component, like cooling, networking,
everybody's listening in.
Yeah, exactly.
And they can contribute, well, this doesn't work for the power distribution.
This doesn't work for the memory.
This doesn't work for this.
Exactly.
And whoever wants to tune out, tune out.
You know what I'm saying?
And the reason for that is because the people who are on the staff,
they know when to pay attention.
They're supposed, you know,
something they could have contributed to,
they didn't contribute it to,
I'm going to call them out, you know,
and so, hey, come on, let's get in here.
So as you mentioned,
Nvidia's this company that's adapting to the environment.
So at which point can you say,
did the environment change,
and began adapting sort of secretly
in the early days?
days from GPU for gaming, maybe the early deep learning revolution to we're now going to start
thinking of it as an AI factory. What does NVIDIA do is produces AI, let's build a factory that makes AI.
I could reason through it just systematically. We started out as an accelerator company,
but the problem with accelerators is that the application domain is too narrow. It has the benefit
of being incredibly optimized for the job. You know, any specialist has that benefit.
The problem with intense specialization is that, of course, your market reach is narrower.
But that's even fine.
The problem is the market size also dictates your R&D capacity.
And your R&D capacity ultimately dictates the influence and impact that you can possibly have in computing.
And so when we first started out as an accelerator, a very specific accelerator,
we always knew that that was going to be our first step.
We had to find a way to become accelerated computing.
But the problem is when you become a computing company,
it's too general purpose, and it takes away from your specialization.
I connected two words that actually have fundamental tension.
The better computing company we become, the worse we became as a specialist.
The more of a specialist, the less capacity we have to do overall computing.
And so that, and I connected those two words together on purpose,
that the company has to find that really narrow path step by step by step
to expand our aperture of computing,
but not give up on the most important specialization that we had.
Okay, so the first step that we took beyond acceleration was we invented the programmable pixel shader.
So that was the first step towards programmability.
It was our first journey towards moving into the world of computing.
The second thing that we did was we created, we put FP32 into our shaders.
That FP32 step, I-Tripply compatible FP32, was a huge step in the direction of computing.
It was the reason why all of the people who were working on stream processors and other types of data flow processors
discovered us.
And they say, hey, all of a sudden,
you know, we might be able to use this GPUs
as incredibly computationally intensive.
And it's now, you know, compliant with IAAA.
I could take my software that I was writing,
you know, previously on CPUs.
And I can, you know, see about, you know,
using the GPU for them.
And which led us to create, put C on top of FP32
was it called, we call CG.
That CG path took us to eventually Kuda.
Kuda, step by step by step.
Well, putting Kuda on G-Force, that was a strategic decision that was very, very hard to do
because it cost the company enormous amounts of our profits, and we couldn't afford it at the time.
But we did it anyways because we wanted to be a computing company.
A computing company has a computing architecture.
A computing architecture has to be compatible across all of the chips that we build.
Can you take me to that decision?
So putting Kuna on G-Force could not afford to do.
Can you explain that decision?
Why boldly choose to do that anyway?
Can you explain that decision?
That was the first, I would say that that was the first,
the first strategic decision that is as close to an existential threat.
But people who don't know, it turned out to be, spoiler alert,
one of the most incredibly brilliant decisions.
ever made by a company.
So CUDA turned out to be
an incredible foundation for computation
in this AI infrastructure world.
So you're just setting the context.
It turned out to be a good decision.
Yeah, it turned out to a big good decision.
I think the...
So here's the way it went.
So we invented this thing called CUDA,
and it expanded the aperture
of applications that we can accelerate
with our accelerator.
The question is,
How do we attract developers to Kuta?
Because a computing platform is all about developers.
And developers don't come to a computing platform
just because, you know, it could perform something interesting.
They come to a computing platform because the install base is large.
Because a developer like anybody else wants to develop software that reaches a lot of people.
So the install base is, in fact, the single most important part.
part of an architecture. The architecture could attract enormous amounts of criticism. For example,
no architecture has ever attracted more criticism than the X-86, you know, as a less than elegant
architecture. But yet, it is the defining architecture of today. It gives you an example that, in fact,
So many risk architectures, which were beautifully architected, incredibly well designed by some of the brightest computer scientists in the world, largely failed.
And so I've given you two examples where one is, you know, one is elegant, the other one's barely aesthetic.
And so yet X86 survived.
Install base is everything.
Install base defines an architecture.
Not everything else is secondary.
Okay, and so there were other architectures at the time.
Kuda came out, OpenCL was here.
There were, you know, there's several other competing architectures.
But the thing that, the decision that we made that was good was we said, hey, look, ultimately, it's about installed base.
And what is the best way we could get a new computing architecture into the world?
By that time frame, G-Force had become successful.
We were already selling millions and millions of G-Force GPUs a year.
And we said, you know, we had to put Kuda on G-Force and put it into every single PC
whether customers use it or not.
And use it as a starting point of cultivating our installed base.
Meanwhile, we'll go and attract developers and went to universities and wrote books
and taught classes and put Kuda everywhere.
And eventually, people would discover, and at the time, the PC was the primary computing
vehicle. There was no cloud. And we could put a supercomputer in the hands of every researcher
in school, every scientist, you know, every engineering school, every student in school. And eventually,
something amazing will happen. Well, the problem was, coulda increased our cost of that GPU,
which is a consumer product, so tremendously, it completely consumed all of the company's gross
profit dollars. And so at the time, the company was probably, you know, worth, I don't know, at the time,
eight, wasn't it like $8 billion or something, $6, $7 billion or something like that?
After we launched Kuta, I recognized that it was going to add so much cost, but it was something
we believed in. You know, our market cap went down to like $1.5 billion. And so we were down there
for a while. And we clawed our way back slowly. But we carried Kuda on G-Force. I always say that
Nvidia is the house that G-Force built because it was G-Force that took Kuda out to everybody.
Researchers, scientists, they discovered Kuda on G-Force because they were all, you know, many of them
were gamers, many of them built their own PCs anyways. In a university lab, many of them built
clusters themselves, you know, using PC components.
And so that, you know, that's kind of how we got going.
And then that became the platform and the foundation for the deep learning revolution.
That was also another great, great observation, yeah.
That existential moment, do you remember, like, what were those meetings like?
What were those discussions like deciding as a company, risking everything?
Well, I had to make it clear to the board what we're trying to do.
and the management team knew our gross margins were to get crushed.
So you could imagine a world where G-Force would carry the burden of Kuda,
and none of the gamers would appreciate it,
and none of the gamers would pay for it.
You know, they only pay certain price,
and it doesn't matter what your cost is.
And so, you know, we increased our cost by 50%.
And that consumed, and we were a 35% gross margin company.
And so it was quite a difficult decision to make.
But you could imagine that someday this would go into workstations
and it would go into supercomputers and in those segments,
maybe we can capture more margin.
So you could reason your way into being able to afford this,
but it still took a decade.
But that's more like conversation with the board convincing them,
but you psychologically, because Nvidia has continued to make
bold bets
that predict the future
and in part,
especially now, define the future.
So I'm almost looking for
wisdom about how you're able to make those decisions
to make leaps like that as a company.
Well, first of all,
I'm informed by a lot of curiosity.
At some point,
there's a reasoning system
that that convinces me so clearly this outcome will happen, that this will happen.
And so I believe it in my mind. And when I believe it in my mind, you know, you know how it is.
You manifest a future. And that future is so convincing, there's no way it won't happen.
There's a lot of suffering in between, but you've got to believe what you believe.
So you envision the future.
Yeah.
And you essentially from a sort of engineering perspective manifested.
Yeah.
And you reason about how to get there.
You reason about why it must exist.
And, you know, I reason, we all reason it here.
The management team would reason about it, all the people that we spent a lot of time reasoning about it.
The thing that the next part of it is probably a skill thing, which is, you know, oftentimes in leadership,
the leadership stays quiet or they learn about something
and then they do some manifesto
and it's a brand new year
and somehow at the end of the year,
next year we're going to have a brand new plan
big huge layoff this way,
big huge organization changed this way,
new mission statement,
brand new logos,
you know, that kind of stuff.
We've just never,
I never do things that way.
When I learned about something
and it's starting to influence how I think,
I'll make it very clear to everybody near me that, you know, this is interesting.
This is going to make a difference.
This is going to impact that.
And I reason about things step by step by step.
Oftentimes, I've already made up my mind, but I'll take every possible opportunity,
external information, new insights, new discoveries, new engineering, you know,
revelations, new milestones develop.
I'll take those opportunities.
and I'll use it to shape everybody else's belief system.
And I'm doing that literally every single day.
I'm doing that with my board.
I'm doing that with my management team.
I'm doing that with my employees.
I'm trying to shape their belief system such that when I come the day I say,
hey, let's buy Melanox.
It's completely obvious to everybody that we absolutely should.
on the day that on the day that I that I said hey guys let's go all in on deep learning and let me tell you why
I've already been laying down the bricks to different organizations inside the company every organization
and every everybody many of the people might have heard everything most of the company heard here's of
course pieces of it and on the day that I announce it um everybody's kind of bought in to
many pieces of it. And in a lot of ways, I like to announce these things. And I imagine that the employees
are kind of saying, you know, Jensen, what took you so long? And in fact, I've been shaping their
belief system for some time. And therefore, leadership, sometimes it looks like you're leading
from behind. But you've been shaping there, you know, to the point where on the day that I
declared it, 100% buy-in. But that's what you want. You want to bring everybody a
long. You know, otherwise, we announced something about deep learning and everybody goes,
what are you talking about? You know, you announce something about let's go all in on this thing,
and your management team, your board, your employees, your customers, they're kind of like,
where's this coming from? You know, this is insane. And so, so GTC, in fact, if you go back in time,
you look at, look at the keynotes, I'm also shaping the belief system of my partners and the industry
and I'm using that to shape, you know, the belief system of my own employees.
And so by the time that I announced something, like, for example, we just announced GROC.
We've been late, I've been talking about the stepping stones for two and a half years.
You just go back and go, oh, my gosh, they've been talking about it for two and a half years.
And so I've been laying the foundation step by step by step.
So when the time comes, you announce it, everybody's, you know, what took you so long?
But it's not just inside the company, you're shaping the landscape, the broader global landscape of innovation.
Like putting those ideas out there, you really are manifesting reality.
We don't build computers.
We actually don't build clouds.
We don't, as it turns out, we're a computing platform company.
And so nobody can buy anything from us.
That's the weird thing.
You know, we vertically design, vertically integrate to design and optimize.
But then we open up the entire platform at every single layer.
to be integrated into other companies' products and services and clouds and supercomputers and
OEM computers.
And so the amazing thing is I can't do what I do without having convinced them first.
And so most at GTC is about manifesting a future that by the time that my product is ready,
they're going, what took you so long?
Yeah.
So one of the things you've been a believer for a long time is,
scaling laws broadly defined. So are you still a believer in the scaling laws? Yeah, we have more
scaling laws now. So I think you've outlined four of them with pre-training, post-training,
test time, and agentic scaling. What do you think, when you think about the future, deep future
and the near-term future, what are the blockers that you're most concerned about to keep you up
at night, that you have to overcome in order to keep scaling? Well, we can go back and reflect on
what people thought were blockers.
So in the beginning, we were the first,
the pre-training
law, you know, people thought,
well, rightfully so,
that the amount of data that we have,
high quality data that we have,
will limit the intelligence that we achieve.
And that scaling law was an important,
very important scale law.
The larger the model,
the correspondingly more data,
results in a better,
results in a smarter AI.
And so that was pre-training.
And Ilius,
Suscover, Ilya said, we're out of data or something like that, pre-training is over or something
like that. The industry panicked, you know, that this is the end of AI. And of course, of course
that's obviously not true. We're going to keep on scaling the amount of data that we have to
train with. A lot of that data is probably going to be synthetic. And that also confused people,
you know, and what people don't realize is they've kind of forgotten that most of the data that
that we are training
that we teach each other with
inform each other with.
This is synthetic.
You know,
it's synthetic because
it didn't come out of nature.
You created it.
I'm consuming it.
I modify it.
Augmented.
I regenerate it.
Somebody else consumes it.
And so we've now reached a level
where AI is able to
take ground truth,
augment it,
enhance it, synthetically generate an enormous amount of data,
and that part of post-training continues to scale.
And so the amount of data that we could use
that is human-generated will be smaller and smaller,
smaller, the amount of data that we use to train model
is going to continue to scale.
To the point where we're no longer limited,
training is no longer limited by data,
is now limited by compute.
And the reason for that is most of the data
synthetic. Then the next phase is a test time. And I still remember people, people telling me that
inference, oh yeah, that's easy. Pre-training, that's hard. These are giant systems that people
are talking about. Inference must be easy. And so inference chips are going to be little tiny chips,
and, you know, they're not like Nvidia's chips. Oh, those are going to be complicated and expensive.
And, you know, we could make, and this is, and in the future inference is going to be the biggest market.
It's going to be easy, and we're going to commoditize, and, you know, everybody can build their own chips.
And that was always illogical to me because inference is thinking.
And I think thinking is hard.
Thinking is way harder than reading.
You know, pre-training is just memorization and generalization, you know, and looking for patterns and relationships.
You're reading and reading versus thinking, reasoning, solving problems, taking,
unexplored experiences, new experiences, and breaking it down into decomposing it into, you know,
solvable pieces that we then go off either through first principle reasoning or, you know,
through previous examples, prior experiences, you know, or just exploration and search and, you know,
trying different things.
And that whole process of test time scaling,
inference is really about thinking.
And it's about reasoning.
It's about planning.
It's about search.
And so how could that possibly be compute light?
And we were absolutely right about that.
So test time scaling is intensely compute intensive.
Then the question is, okay, now we're at inference and we're at test time scaling.
What's beyond that?
Well, obviously, we have now created, you know, one agenic person.
And that one agentic person has a large language model that we've now, you know, developed.
But during test time, that agentic system goes off and does research and bangs on databases and it goes and, you know, uses tools.
And one of the most important things it does is spins off and spawns off a whole bunch of subagents,
which means we're now creating large teams.
It's so much easier to scale Nvidia
by hiring more employees
than it is to scale myself.
And so the next scaling law is the agentic scaling law.
It's kind of like multiplying AI.
Multiplying AI, we could spin off agents
as fast as you want to spend off agents.
And so, you know, you have four scaling laws.
And as we use the agentic systems,
they're going to create a lot more data,
they're going to create a lot of experiences.
Some of it we're going to say,
wow, this is really good.
We ought to memorize this.
That data set then comes all the way back to pre-training.
We memorize and generalize it.
We then refine it and fine-tune it back into post-training.
Then we enhance it even more with test time,
you know, in the agentic systems,
you know, put it out into the industry.
And so this loop, the cycle,
is going to go on and on and on.
It kind of gums down to basically
intelligence is going to scale
by one thing and this compute.
But there's a tricky thing there
that you have to anticipate and predict,
which is some of these components
requires different kind of hardware
to really do it optimally.
So you have to anticipate
where the AI innovation is going to lead.
For example, make sure of experts
with sparsity.
Perfect.
With hardware, you can't just pivot
on a week's notice, you have to anticipate
what that's going to look like.
That's so scary and difficult to do, right?
For example, these AI model architectures
are being invented about once every six months.
Yeah.
Right?
And system architectures and hardware architectures
kind of every three years.
And so you need to anticipate
what likely is going to happen
to three years from now.
And there's a couple of ways that you could do that.
First of all, we could do research internally ourselves.
And that's one of the reasons why we have basic research.
We have applied research.
We create our own models.
And so we have hands-on life experience right here.
This is part of the co-design that I'm talking about.
We're also the only AI company in the world that works with literally every AI company in the world.
And to the extent that we can, we try to get a sense of what are the challenges that people are experiencing.
So you're listening to the whispers across the industry, the Aalibs,
That's right. You've got to listen and learn from everybody and have a, have a,
and then the last part is to have an architecture that's flexible that can adapt and move with the wind.
And one of the benefits of Kuda is that it's, you know, on the one hand, an incredible accelerator.
On the other hand, it's really flexible.
And so that balance, incredible balance between specialization, otherwise we can't accelerate the CPU versus generalization.
so that we can adapt with changing algorithms,
that's really, really important.
That's the reason why Kuta has been so resilient on the one hand,
and yet we continue to enhance it.
We're at Kuta 13.2,
and so we're evolving the architecture so fast
that we can stay with the modern algorithms.
For example, when mixture of experts came out,
that's the reason why we had MVLink-72,
instead of MVLink 8.
We could now take an entire
$4 trillion, $10 trillion per
parameter model and put it in one
computing domain as if it's running
on one GPU.
People
probably didn't notice.
I said it, but
if you look at the architecture of
the Grace Blackwell
racks, it was completely
focused on doing one thing,
processing the LLM.
All of a sudden, one year
you're looking at a Vera Rubin rack.
It has storage accelerators.
It has this incredible new CPU called Vera.
It has Vera Rubin and NVLink-72 to run the LLMs.
It also has this new additional rack called GROC.
And so this entire rack system is completely different than the previous one.
And it's got all these new components in it.
And the reason for that is because the last one was designed to run
M-O-E large language models, inference,
and this one is to run agents,
and agents bang on tools.
Obviously, the design of the system
had to have been done
before Clawed Code Codex, OpenClaw,
so you were anticipating the future, essentially.
And that comes from what?
From the whispers, from the understanding of what all the state of the artist?
No, it's easier than that.
You just reason about it.
First of all, it's the reason.
No matter what happens at some point,
in order for that large language model to be a digital worker,
let's just use that metaphor.
Let's say that we want the LLM to be a digital worker.
What does it have to do?
It has to access ground truth.
That's our file system.
It has to be able to do research.
It doesn't know everything.
We don't have,
and I don't want to wait until this AI becomes,
you know, universally smart about everything past, present, and future before I make it useful.
And so, therefore, I might as well let it go do research.
It's obviously, if it wants to help me, it's got to use my tools.
You know, a lot of people would say, you know, AI is going to completely destroy software.
We don't need software anymore.
We don't need tools anymore.
That's ridiculous.
Let's use the, let's use a thought experiment.
And you could just sit there and you're a glass of whiskey.
and think about all these things
and it would become completely obvious.
Like if I were to create
the most amazing agent
that we can imagine in the next 10 years,
let's say be a human or robot.
If that human robot were to be created,
is it more likely that the human robot
comes into my house
and uses the tools that I have
to do the work that it needs to do?
Or does this hand turns into a 10-pound hammer?
in one instance turns into a scalpel in another instance. And in order to boil water, it beams
microwaves out of its fingers. Or is it more likely just to use the microwave? You know, and the first
time it goes up to the microwave, it probably doesn't know how to use it. But that's okay. It's
connected to the internet. It reads the manual of this microwave, reads it, instantly becomes an expert.
And so it uses it. And so I think the, I just,
described, in fact, almost all of the properties of OpenClaw, you know, that it's going to use
tools, that it's going to access files, it's going to be able to do research. It has I.O. subsystem.
And when you're done reasoning through it, reasoning about it through it in that way,
then you say, oh my gosh, the impact to the future computing is deeply profound.
And the reason for that is, I think we've just reinvented the computer.
and then now you say, okay, when did we reason about that?
When did we reason about open claw?
If you take the open claw schematic that I used at GTC, you will find it two years ago.
Literally two years ago at GTC, I was talking about agentic systems that exactly reflect open claw today.
And of course, the confluence of many things had to happen.
all, we need a clot and GPD and, you know, all of these models to reach a level of capabilities.
So their innovation and their breakthroughs and their continued advances was really important.
And then, of course, somebody had to create an open source, you know, project that was sufficiently
robust, you know, and sufficiently complete, and that we can all put to work.
And I think OpenClaw did for agenic systems,
what ChatGPT did for generative systems.
And I just think it's a very big deal.
Yeah, it's a really special moment.
I'm not exactly sure why it captured so much of the world's attention,
but it did more than Claude Code and Codex and so on.
Because consumers could reach it.
Sure, yeah.
But there's also so much of this is vibes.
And Peter, I had a podcast with him,
a wonderful human being.
So part of it is also the human being.
humans that represent the thing.
No doubt.
Part of it is memes and the,
because we're all trying to figure it out.
There's really serious and complicated security concerns about when you have such
powerful technology, how do you hand over your data so they can do useful stuff, but then
there are scary things associated with that.
And we as a civilization, as individual people, and as a civilization figuring out how to
find that right balance.
Yeah, we, we jumped on it right away, and we sent a bunch of security experts as way.
And we did this thing called OpenShel.
It's already been integrated.
into open claw.
And they put forward
Nemo Claw. Yeah, exactly.
They install is super easy.
It makes sure that it's secure.
We give you two out of three rights.
Agentic systems can access
sensitive information. It can execute
code and it can communicate externally.
We could keep things safe
if we give you two out of those three
capabilities at any time, but not
all three. And out of those
two out of three capabilities, we also give you
access control based on whatever rights that you're given by enterprise.
And then we connected to a policy engine that all these enterprises already have.
And so we're going to try to do our best to help open claw become a better claw.
So you eloquently explained how we have a long history of blockers that we thought were going to be
blockers and we overcame them.
But now looking into the future, what do you think might be the blockers?
Now that it's clear that agents will be everywhere.
So it's obviously we're going to need compute.
So what is going to be the blocker for that scaling?
Power is a concern, but it's not the only concern.
But that's the reason why we're pushing so hard on extreme code design
so that we can improve the tokens per second per watt orders of magnitude every single year.
And so in the last 10 years, Moore's Law would have progressed computing about 100 times
in the last 10 years.
we progressed and scaled up computing by a million times in the last 10 years.
And so we're going to keep on doing that through extreme code design.
So energy efficiency per per watt completely affects the revenues of a company.
It affects the revenues of a factory.
And we're just going to push that to the limit so that we could keep on driving token costs down as fast as we can.
You know, our computer price is going up, but our token generation effectiveness is going up so much faster that token cost is coming down.
It's just, it's coming down an order of magnitude every year.
So power, that's an interesting one.
So the way to try to get around the power blocker is to try to, with the tokens per second per watt, try to make it more and more efficient.
Of course, there's the question of how do we get more power?
We should also get more power.
That's a really complicated one.
you've talked about small module nuclear power plants.
There's all kinds of ideas for energy.
How much does it keep you up at night,
the bottlenecks in the supply chain of AI,
like ASML with EU V lithography machines,
DSMC with advanced packaging, like COOS,
and SKHinnix with high bandwidth memory?
All the time, and we're working on all the time.
No company in history has ever grown at a scale that we're growing
while accelerating that growth.
It's incredible.
Yeah.
And it's hard for people to even understand this.
In the overall world of AI computing, we're increasing share.
And so supply chain, upstream and downstream, are really important to us.
I spent a lot of time informing all the CEOs that I work with.
What are the dynamics that's going to cause the growth to continue or even accelerate?
It's part of the reasons why to the entire right-hand.
hand side of me were CEOs of practically the entire IT industry upstream and practically the
entire infrastructure industry downstream. And there were all, there were several hundred CEOs.
And I don't think there's ever been keynotes where several hundred CEOs show up. And part of it is,
I'm telling them about our business condition now. I'm telling them about the growth drivers
in the very near future and what's happening.
And I'm also describing where are we going to go next
so that they could use all of this information
and all of the dynamics that are here
to inform how they want to invest.
And so I informed them that way
like I inform my own employees.
And then of course, then I make trips out to them
and make sure that, hey, listen,
I want you to know this quarter, this coming year,
this next year, these things are going to happen.
and if you look at the CEOs of the DRAM industry,
the number one DR RAM in the world was DDR memory for CPUs and data centers.
About three years ago, I was able to convince several of the CEOs that even though at the time,
HBM memory was used quite scarcely, you know, and barely by supercomputers,
that this was going to be a mainstream memory for data.
centers in the future. And at first it sounded ridiculous, but several of the CEOs believed me
and decided to invest in building HBO memories. Another memory was rather odd to put into a
data center is the low power memories that we used for cell phones. And we wanted them to adapt them
for supercomputers in the data center. And they go, cell phone memory for supercomputers?
And I explained to them why. Well, look at these two memories, LPDDR5.
HBM4, the volumes are so incredible.
All three of them had record years in history,
and these are 45-year companies.
And so, you know, that's part of my job
is to inform and shape, inspire, you know.
So you're not just manifesting the future
and maybe inspiring Nvidia,
the different engineers of the company,
you're manifesting the supply chain
of the future. So you're having conversations with TSM, with ASML. Upstream, downstream, downstream.
So that's the thing. GEV, Caterpillar. Yeah, that's downstream from us. Yeah, yeah.
Yeah, the whole thing. I mean, but that's so, there's so much incredibly difficult engineering
that happens in the entire semiconductor industry. And it just feels scary how intricate the supply chain is,
many components there are, but it works somehow.
Exactly. The deep science, the deep engineering, the incredible manufacturing, and so much of
the manufacturing is already robotics, but we have a couple of hundred suppliers that contribute
the technology that goes into our 1.3 million component rack. Each rack is 1.3, 1.5 million
components. There are 200 suppliers across the Veraruban rack. So it's interesting that you
don't list that as the thing that keeps you up at night in the list of blockers.
But I'm doing all the things necessary to see.
I can go to sleep because I checked it off.
I said, okay, you know, I go, I can go to sleep, I go, well, let's see.
Let's reason about this.
What's important for us?
Okay, let's reason about this.
Because we change the system architecture from the original DGX-1 that you remembered to
MVLink 72 Rack scale computing.
What does that mean?
What does that mean to software?
What does that mean to engineering?
What does that mean to how we design and test?
And what does that mean to the supply chain?
Well, one of the things that meant was we moved supercomputer integration at the data center
into supercomputer manufacturing in the supply chain.
If you're doing that, you also have to recognize you're going to move.
and if you're in a total footprint of whatever data center you're going to build,
let's say you would like to have, you know, 50 gigawatts of supercomputers that are running
simultaneously. And it takes one week to manufacture that 50 gigawatts of supercomputers,
then each week in the supply chain, the supercomputers are going to need a gigawatt of power.
And so we're going to need the supply chain
to increase the amount of power it has
to build, test,
to build and test the supercomputers
in the supply chain before I ship it.
Well, MVLINK72 literally built supercomputers
in the supply chain
and ships them two, three tons at a time per rack.
It used to be, they used to come in parts
and we used to assemble them inside the data center.
But that's impossible now
because NVLing 72 is so dense.
And so that's an example.
And I would have to go
into, you know, I fly into the supply chain, go meet my partners saying, hey, I said, guess what?
So here's what we're going to do with, this is the way we used to build our DGXs.
We're going to build them this way.
This is going to be so much better because we're going to need them for inference.
The market for inference is, you know, coming.
The inflection point for inference is coming.
It's going to be a big market.
And so I first explained to them what's going on, why it's going to happen.
And then I ask them to make several billion.
dollars of capital investments each. And because they, you know, they trust me. And I, I'm very
respectful of them. And I give them every opportunity to question me. And I spend time to explain
things to people. And I reason about it. I draw them pictures. And I reason about it in first
principles. And by the time I'm done with them, there's no what to do. So a lot of us about
relationships and building a shared view of the future. Yeah. But do you worry about
certain bottlenecks. I mean, what are the biggest bottlenecks in the supply chain? Are you worried about
ASMLs, EV tooling? Are you worried about the packaging, co-os packaging of TSM about how fast it
could scale? Like you said, you're not only growing incredibly fast, you're accelerating a growth.
So it feels like everybody in the supply chain, and those are certainly bottlenecks would have to
scale up. Are you having conversations with them? Like, how can you scale this faster? Do you worry about
it? No. Okay. Because
I told them what I needed.
They understood what I need.
They told me what they're going to go do,
and I believe what they're going to do. Interesting.
Yeah. That's great to hear. So maybe
if we can just link on the power for a little bit,
what are your hopes for
how to solve the energy problem?
One of the areas that I would love
us to talk about and just get the message out.
You know, our
power grid,
is designed for the worst-case condition with some margin.
Well, 99% of the time
were nowhere near the worst-case condition
because the worst-case condition is a few days in the winter,
a few days in the summer, and extreme weather.
Most of the time, we're nowhere near the worst-case condition,
and we're probably running around, call it 60% of peak.
And so 99% of the time, our power grid,
has excess power
and they're just sitting idle
but they have to be there sitting idle
because just in case
when the time comes
hospitals have to be powered
and infrastructure has to be powered
and airports have to run
and so and so forth
and so the question that I have
is whether we could go
and help them understand
and create contractual agreements
and design computer architecture
systems data centers
such that
when they need
the maximum power for infrastructure in society
that the data centers would get less.
But that's in a very rare instance anyways.
And during that time, we either have a backup generator
for that little part of it,
or we just have our computers shift to workload somewhere else,
or we have the computers just run slower.
You know, we could degrade our performance,
reduce our power consumption,
and provide for a, you know,
a slightly longer latency response
you know, when somebody asked for, you know, ask for an answer.
And so I think that that way of using computers,
of building data centers,
instead of expecting 100% up time,
and these contracts that are really, really quite rigorous,
it's putting a lot of pressure on the grid to be able to,
now they're going to have to increase from their maximum.
I just want to use their excess.
It's just sitting there.
Yeah, that's not talked about enough.
So what's stopping there?
Is it regulation?
Is it bureaucracy?
I think it's a through-way problem.
It starts with the end customer.
The end customer puts requirements on the data centers
that they can never not be available.
Okay, so that the end customer expects perfection.
Now, in order to deliver that perfection,
you need a combination of backup generators
and your grid power supplier to deliver on perfection.
And so everybody's got to have six-nines.
Well, I think, first of all, right now,
we ought to have everybody understand that when the customer
asked for these things, you have somebody in your
data center operations team disconnected from the CEO.
I bet the CEO doesn't know this.
I'm going to talk to all the CEOs.
Their CEOs are probably not paying any attention
to the contracts that are being signed.
And so everybody wants to sign the best contract, of course.
and they go down to the cloud service providers
and the two contract negotiators
that I could just see them now
negotiating these multi-year contracts
both sides want the best contract
as a result
the CSPs then have to go down to the utilities
and they expect the nine to six-nines
and so I think I think the first thing
is just make sure that all of the customers
the CEOs of the customers realize what they're asking for.
Now, the second thing is we have to build data centers that gracefully degrade.
And so if the power, the utility, the grid tells us, listen, we're going to have to back you down to about 80%.
We're going to say that's no problem at all.
We're just going to move our workload around.
We're going to make sure that data is never lost, but we can reduce the computing rate and use less energy.
The quality of service degrades a little bit for the critical workloads.
I shift that somewhere else right away, so I don't have that problem. And so, you know,
whoever, whichever data center still has 100% uptime. And so how difficult of an engineering problem
is that the smart, dynamic allocation of power in the data center? As soon as you could specify,
you could engineer it. Beautifully put. So long as it obeys the laws of physics on first
principles, I think we're good. What was the third thing you were mentioning?
So the second thing is the data centers. And the third thing is, we need to you,
utilities to also recognize that this is an opportunity.
And instead of saying, look, it's going to take me five years to increase my grid capability.
If you have, if you're willing to take power of this level of guarantee, I can make them available for you next month and at this price.
And so if utilities also offered more segments of power delivery,
promises, then I think everybody will figure out what to do with it.
But there's just way too much waste in the grid right now.
We should go after it.
You've highly lauded Elon and XAI's accomplishment in Memphis
in building Colossus supercomputer, probably in record time in just four months.
It's now at 200,000 GPUs and growing very quickly.
Is there something that you can speak to the understand about his approach
that's instructive to the broadly
to all the data center creators
that enable that kind of accomplishment
his approach to engineering, his approach to the whole
management of construction, everything.
First of all, Elon is deep in so many different topics,
yet he's also a really good systems thinker.
And so he's able to think through multiple disciplines.
And he obviously
pushes things, questions everything,
where they're number one, is it necessary?
Number two, does it have to be done this way?
And in other words, you know, does it have to take this long?
And so he has the ability to question everything
to the point where everything is down to its minimal amount
that's necessary.
You can't take anything else out.
and yet
the
the
the necessary
capabilities
of the product
retains
you know
and so he's
he is as minimalist
as you could
possibly imagine
and he does it
at a system
scale
I also love
the fact that
he is
he is represented
he is
present
at the point of action
you know
he'll just go there
if there's a problem
he'll just go there
and you show me the problem.
You know, when you do all of this in combination,
you overcome a lot of previous,
this is the way we do it.
You know, I'm waiting for them.
You know, I mean, just everybody has a lot of excuses.
And so, and then the last thing is
when you act personally with so much urgency,
it causes everybody else to act with urgency,
you know?
And every supplier has a lot of customers going on.
And every supplier has a lot of projects going on.
And he makes it his business that he's the top priority of everybody else's, you know, projects.
And so he does that by demonstrating it.
Yeah, I've been in a bunch of those meetings.
It's fun to watch because really not enough people ask the question, like, okay, so can this be done a lot faster and how?
Why does it have to take this long?
Yeah, right.
And then that becomes an engineering question often.
And yes, I think when you get the ground truth of actually.
I actually, I remember one of the times I was hanging out with him,
he literally is going through the entire process of how to plug in cables into a rack.
He was working with an engineer on the ground that's doing that task,
and he's just trying to understand what does that process look like,
so it can be less air-prone.
And just building up that intuition from every single task involved in putting together the data center,
you start to immediately get a sense at the detailed scale,
and then at the broad system scale
of where the inefficientities are.
And so you can make it more and more and more efficient.
Plus you have the big hammer of being able to say,
let's do it totally different
and remove all possible blockers.
That's right.
Is there parallels in the Nvidia Extreme Systems Code Design approach
that you see in the way Elon approaches systems engineering?
Well, first of all,
code design is an ultimate systems engineering problem.
And so we approach the work that we do
from that principle.
The other thing that we do, and this is a philosophy that a thought, a state of mind, I guess, a method that I started 30 years ago, and it's called the speed of light.
A speed of light is not just about the speed.
A speed of light is my shorthand for what's the limit of what physics can do.
And so everything that we do is compared it against the speed of light.
memory speed, math speed, power, cost, time, effort, number of people, manufacturing cycle time.
And when you think about latency versus throughput, when you think about cost versus throughput,
cost versus capacity, all of these things, you test against the speed of light to achieve
all of these different constraints separately. And then when you consider it together, you know you have
to make compromises because a system that achieves extremely low latency versus a system that achieves
very high throughput are architected fundamentally differently. But you want to know what's the
speed of light of a system that achieves high throughput? What's the speed of light of a system that
achieves low latency.
And then when you think about the total system, you could make trade-offs.
And so I force everybody to think about what's the first principles, the limits, the physical
limits for everything before we, you know, before we do anything.
And we test everything against that.
And so that's a good frame of mind.
I don't love the other methods, which is continuous improvement.
The problem with continuous improvement,
first of all, you should engineer something
from first principles with speed of light thinking,
limited only by physical limits and physics limits.
And after that, of course, you would improve it over time.
But I don't like going into a problem and somebody says,
hey, it takes 74 days to do this today right now.
And we can do it for you.
you in 72 days. You know, I'd rather strip it all back to zero and say, first of all,
explain to me why it's 74 days in the first place. And let's know, let's think about what's
possible today. And if I were to build it completely from scratch, you know, how long would it
take? Oftentimes you'd be surprised and might come to six days. Now the rest of the six days to 74
could be very well-reasoned and compromises and, you know, cost reductions and all kinds of
different things, but at least you know what they are. And now that you know that six days
possible, then the conversation from 74 to 6, surprisingly much more effective.
And such incredibly complex systems that you're working with, is simplicity sometimes a good
heuristic to reach for? I mean, if I can just, I mean, the pod, the Vereruban pod that you
announce is just incredible. We're talking about seven chips, seven chip types.
five purpose built rack types, 40 racks,
1.2 quadrillion transistors,
nearly 20,000 Nvidia dies,
over 1,100 Ruben GPUs,
60 Xxaflops, 10 petabytes per second of scale,
bandwidth.
That's all just one.
That's just one pod.
That's just one pie.
Yeah, that's just one pie.
I mean, so you have the,
and then even the NVL 72 rack alone is 1.3 million components.
1,300 chips, 4,000 pounds crammed into a single 19-inch wide rack.
And Lex, we'll probably kind of crank out about 200 of these pods a week, just to put in perspective.
The amount of different components, I suppose simplicity is impossible.
But is that a metric that you kind of reach for and trying to design things?
You know, the phrase that I use most often is, we need things to be as complex as necessary, but as simple as possible.
And so the question is, is all that complexity there necessary?
And we ought to test for that.
And we ought to challenge that.
And then after that, everything else above it, you know, it's gratuitous.
But some of the most incredible semiconductor industry broadly,
but what Nvidia is doing,
some of the greatest engineering in history.
So these systems are just truly, truly marvels of engineering.
It is the most complex computer the world has ever made.
Yeah, the engineering teams, I mean, I don't know, it's not a competition, but I don't know, if it was like an Olympics of engineering teams, I mean, TSM does incredible engineering, like I said, ASML at every scale.
But Nvidia is going to give them a run for their money.
Yeah.
Just incredible, incredible teams.
Gold medalists in every single, every single sport all assembled right here and have to work together and report directly to you.
This is wonderful.
You've recently traveled to China.
So it's interesting to ask you, China has been incredibly successful in building up its technology sector.
What do you understand about how China is able to, over the past 10 years, build so many incredible world-class companies, world-class engineering teams, and just this technology ecosystem that produces so many incredible products?
A whole bunch of reasons.
Well, first of all, let's start with some fast.
50% of the world's AI researchers are Chinese, plus or minus.
And they're mostly in China, still.
We have many of them here, but there's amazing researchers still in China.
Their tech industry showed up at precisely the right time.
At the time of the mobile cloud era, their way of contributing with software,
and so this is a country's incredible science and math,
really well-educated kids.
Their tech industry was created during the era of software.
They're very comfortable with modern software.
China is not one giant economic country.
It's got many provinces and cities,
with mayors all competing with each other.
That's the reason why there's so many EV companies.
That's the reason why there's so many AI companies.
That's the reason why there's so many,
every company you could imagine,
they all create some of them.
And as a result, they have insane competition internally.
And what remains is an incredible company.
They also have a social culture
where it's family first, friend second, and company third.
And so the amount of conversation
that goes back and forth
between
they're essentially open source
all the time.
So the fact that they contribute
more to open source
is so sensible
because they're probably
what are we protecting?
You know?
My engineers,
their brothers are in that company,
their friends are in that company
and they're all schoolmates,
you know, the schoolmate concept.
It's a, you know,
one schoolmate,
your brother for life.
And so they
they share knowledge
very, very quickly.
And so there's no sense keeping technology hidden.
You might as well put it on open source.
And so the open source community then amplifies,
accelerates the innovation process.
So you get this rapid, incredible great talent,
rapid innovation because of open source and just the nature of friends.
And insane competition among the company,
what emerges is incredible stuff.
And so this is the fastest innovating country in the world today.
And this is something that has everything that I've just said is fundamental to just how the kids were grown, the fact that they have excellent education, the fact that parents want them to do well in school, the fact that their culture is that way.
These are, you know, these are just the thing about their country.
And they showed up at precisely the time when technology is going through that exponential.
Plus, culturally, it's pretty cool to be an engineer.
It connects to all the components that you're mentioning.
It's a builder nation.
It's a builder nation.
Yeah, it's a builder nation.
Our country's leaders, incredible, but they're mostly lawyers.
Their country's leaders, because they're trying to keep us safe, rule of law, governing,
their country was built out of poverty.
And so most of their leaders are incredible.
engineers, some of the brightest minds.
To take a small tangent, because you mentioned open source, I have to go to
Perplexity here, who you've been a fan of a long time.
I love it, yeah.
And thank you for releasing Open Source Nematron 3, Super, which you can also use
inside Perplexity to look stuff up, which is 120 billion parameter open weight
MOE model.
What's your vision with open source?
So you mentioned China with DeepSeek or Minimax
with all these companies
really pushing forward the open source
AI movement
and Nvidia is really leading the way
in close to state of the art
open source LMs.
What's your vision there?
First, if we're going to be a great AI computing company,
we have to understand how AI models are evolving.
One of the things that I love about Nemotron 3
is not just a pure transformer model,
it's transformer and SSMs.
And we were early in developing the conditional GANS,
which that progressive GANS,
which led step by step to diffusion.
And so the fact that we're doing basic research
in model architecture and in different domains
gives us visibility into what kind of computing systems
would do a good job for.
future models. And so it is part of our extreme co-design strategy. Second, I think we rightfully recognize
that on the one hand, we want world-class models as products, and they should be proprietary.
On the other hand, we also want AI to diffuse into every industry and every country, every researcher,
every student.
And if everything is proprietary,
it's hard to do research
and it's hard to innovate
on top of, around,
with.
And so,
open source is fundamentally necessary
for many industries
to join the AI revolution.
Nvidia has the scale
and we have the motives
to not only
skills, scale, and motivation
to build
and continue to build these AI models
for as long as we shall live.
And so therefore we ought to do that.
We can open up, we can activate every industry,
every researcher,
you know, every country to be able to join the AI revolution.
There's the third reason,
which is from that to recognizing
that AI is not just language.
These AIs will likely use tools
and models and subagents,
that were trained on other modalities of information.
Maybe it's biology or chemistry or laws of physics
or fluids and thermodynamics.
Not all of it is in language structure.
And so somebody has to go make sure
that whether prediction, biology, AI,
AI for biology, physical AI,
all of that stuff can be pushed to the limits
and push to the frontier.
We don't build cars,
but we want to make sure
every car company
has access to great models.
We don't discover drugs,
but I want to make sure
that Lilly has the world's best
biology, AI systems
so that they can go use it
for discovering drugs.
So these three fundamental reasons,
both in recognizing
that AI is not just language,
that AI is really broad,
that we want to engage
everybody into the world of AI
and then also co-design of AI.
Well, I have to say,
Once again, thank you for open sourcing.
It's really, truly open sourcing, Neutron 3.
Yeah, I appreciate you for saying that.
We open source the models.
We open source the weights.
We open source the data.
We open source how we created it.
Yeah, it's pretty amazing.
It's really incredible.
You're originally from Taiwan and have a close relationship with TSMC.
So I have to ask, TSMC, I think, also is a legendary company
in terms of the engineering teams,
in terms of the incredible engineering work that they do.
What do you understand about TSM culture
and their approach that explains how they're able
to achieve this singular unmatched success
in everything they're doing with semiconductors?
You know, first of all,
the deepest misunderstanding about TSM is that their technology
is all they have.
that somehow they have a really great transistor.
And if somebody shows up another transistor game over,
it's the technology, and of course, you know,
I don't mean just the transistor and metallization systems,
the packaging, the 3D packaging, the silicon photonics,
that, you know, all of the technology that they have.
That technology is really what makes the company special.
Their technology makes the company special.
but their ability to orchestrate the demands, the dynamic demands of hundreds of companies in the world,
as they're moving up, shifting out, you know, increasing, decreasing, pushing out, pushing out, pulling in,
changing from customer to customer, wafer starting, wafer stopping,
emergency wafer starts, you know, all of this dynamics of the world's complexity as the world is
shifting all the time.
And somehow they're running a factory with high throughput, high yields, really great costs,
excellent customer service.
They take their promises seriously.
When you're wafer, because they know that they're helping you run your company,
when the waferers were promised to show up, the waifers show up, you know, so that you
could run your company appropriately.
So their system, their manufacturing system is completely miraculous.
I would say then the second thing is their culture.
This culture is simultaneously technology focused on one hand, advancing technology,
simultaneously customer service oriented on the other hand.
A lot of companies are very customer service oriented, but they're not very technology.
excellent. They're not at the bleeding edge of technology or a lot of companies who are at the
bleeding edge of technology, but they're not the best customer service oriented company. And so
it just depends on somehow they've balanced these two and their world class of both.
And then probably the third thing is the technology that I most value in them that they created
this, you know, this intangible call trust. I trust them to put my company on top of them.
that's a very big deal.
But they trust, I mean, there's a really close relationship there that you have established
and that trust is established based on many years of performance,
but there's human relationships involved there as well.
Three decades, I don't know how many tens, hundreds of billions of dollars of business
we've done through them, and we don't have a contract.
That's pretty great.
Amazing.
Okay, there's a story that in 2013, the founders of TSMC and Morris Chang offered you
the chance to become TSM's chief executive.
And you said you already had a job.
Is this story true?
The story is true.
I didn't dismiss it.
But I was deeply honored.
And of course, of course, I knew then, as I know now,
TSM is one of the most consequential companies in history.
Yeah.
And Morris is one of the highest regarded executive
and business and personal.
friend that I've that I've had in my life. And for him to ask is, I was humbled and really honored.
But the work that I'm doing here is really important. And I've seen, you know, in my mind in
ways, in my mind's eye, what NVIDIA was going to be and what the impact that we could have.
And it was really important work. And it's my responsibility.
my sole responsibility to make this happen.
And so I declined it, you know, not because it wasn't an incredible offer.
It's an unbelievable offer, but I simply couldn't take it.
I think Nvidia, both Nvidia and TSMC are two of the greatest companies in the history of human civilization.
Running either one, I'm sure, is incredibly complicated effort.
Intakes, you have to truly be all in.
everybody at every scale
not just at the CEO level
everybody is really truly all in
to accomplish this kind of complexity
See now I can help both companies
Exactly
So Nvidia is now the most valuable
company in the world
I have to ask what is the
Nvidia's biggest moat
As the folks in the tech sector say
The edge you have that
protects you from the competition
Our single most important property as a company is the install base of our computing platform.
Our single most important thing today is the install base of Kuta.
Now, the reason why 20 years ago, of course, there was no installed base.
But what makes, and if somebody came up with a Gouda,
or Tudah, it wouldn't make any difference at all.
And the reason for that is because it's never been just about the technology.
The technology, of course, was incredible visionary.
But it's the fact that the company was dedicated to it, stuck with it, expanded its reach.
It wasn't three people that made CUDA successful.
It was 43,000 people that made CUDA successful.
And there's several million developers that believed in us.
that trusted that we were going to continue to make Kuta 1, 2, 3, 13,
that they decided to port and dedicate their software on top of it,
their mountain of software on top of it.
And so the install base is the number one most important advantage.
That install base, when you amplified with the velocity of our execution
at the scale that we're talking about,
no company in history had ever built systems of this complexity, period,
and then to build it once a year is impossible.
And that velocity combined with the installed base
in the developer's mind,
you just can now take a developer's mind.
From the developer's perspective,
if I support CUDA,
tomorrow it will be 10 times better.
I just have to wait six months on average.
Not only that, if I develop it on CUDA,
I reach a few hundred million people, computers.
I'm in every cloud.
I'm in every computer company.
I'm in every single industry.
I'm in every single country.
So if I created an open source package and I put it on CUDA first,
I get these both attributes simultaneously.
And not only that,
I trust 100% that Nvidia is going to keep CUDA
around and maintain it and improve it and keep optimizing the libraries for as long as they
shall live.
You could take that to the bank and that last part, trust.
You put all that stuff together.
If I were a developer today, I would target CUDA first.
I would target CUDA most.
And that's the reason that I think in the final analysis is our first, that's even our first
core advantage.
Our second one is our ecosystem.
The fact that we vertically integrated
this incredibly complex system,
but we integrated horizontally
into every single company's computers.
We're in the Google Cloud.
We're in Amazon. We're in Azure.
You know, we're ramping up AWS like crazy right now.
We're in new companies like CoreWeave and N-scale.
We're in supercomputers at Lilly.
We're in enterprise computers.
We're at the edge.
in radio base stations.
You know, it's just crazy.
One architecture is in all these different systems.
We're in cars, we're in robots.
We're in satellites.
We're out in space.
And so the fact that you have this one architecture
and the ecosystem is so broad.
It basically covers every single industry in the world.
Well, how does the Kuda install base
evolve into the future with AI factories as a moat?
Do you think it's possible that Nvidia of the future
is all about the AI factory?
Well, the unit of computing used to be GPU to us.
Then it became a computer.
Then it became a cluster.
Now it's an entire AI factory.
When I see a computer, when I see what Nvidia builds,
in the old days, I visualized the chip.
And then when I announced a new product,
you know, a new generation like,
ladies and gentlemen, we're announcing Ampure today.
I pick up the chip.
Yeah.
That was my mental model of what I was building.
Today, I went picking up.
the chip is kind of still adorable.
Yeah.
But it's adorable.
It's not my mental model of what I'm doing.
My mental model is this giant gigawatt thing that has power generation, is connected to the grid.
It's got cooling systems and networking of incredible monstrosity.
You know, 10,000 people are in there trying to install it.
Hundreds of networking engineers in there.
Thousands of engineers behind it trying to power it up, you know, powering up one of
those factories, as you know, it's not somebody going, it's on now.
Takes thousands of people to bring it up.
So mentally, you're actually, when you're thinking about a single unit of compute,
you're like literally, when you go to bed at night, you're thinking now about
collection of racks, so pods, not individual chips.
Entire infrastructure.
And I'm hoping my next click is when I'm thinking about building computers, it's, you know,
planetary scale.
That'll be the next click.
What do you think about the space angle that Elon,
has talked about doing compute in space for solving some of the, it makes some of the energy issues
in terms of scaling energy easier.
Cooling issues is not easy, yeah.
Cooling.
Well, there's a large number of engineering complexities involved with that.
So what, you know, Nvidia has also announced that you're already thinking about that.
Yeah, we're already there.
Nvidia GPUs are the first GPUs in space.
and I didn't realize it was so interesting to,
I would have declared it maybe.
We're in space.
You know, little astronaut suit on one of our GPUs.
But we've been in space.
It's the right place to do a lot of imaging, you know,
because those satellites have really high-resolution imaging systems,
and they're sweeping the Earth, you know, continuously now.
And you want, you know, centimeter scale,
imaging that is done continuously for the world so that you know you'll basically have real-time
telemetry of everything you don't want to beam that back down to earth it's just you know
petabytes and petabytes of data you got to just do AI right there at the edge throw away
everything you don't need you've seen before didn't change and then just keep the stuff that you need
and so AI ought to be done at the edge um obviously we have we have 24-7
solar if we put it at the polar's.
And, um, uh,
but, you know, there's no conduction, no convection.
And so, you know, you're pretty much just radiation.
And, um, uh, but, you know, space is big.
I guess, you know, we're just going to put big giant radiators out there.
How crazy of an idea I do you think it is?
Like, is this, is this five years out, 10 years out, 20 years out?
So, uh, we're talking about blockers for AI scaling.
You know, I'm just so much more practical.
I look for where my next bucket of opportunities are first.
Meanwhile, I'm cultivating space.
And so I send engineers to go work on the problem.
We're learning a lot about it.
How do we do radiation?
How do we do degrading performance?
How do we deal with continuous testing and attestation of defects?
And how do we deal with redundants?
and, you know, how do we deal with redundancy
and how do we degrade gracefully and things like that?
And so we could do, what about software?
How do you think about software and redundancy
and performance out in space?
Make it so that the computer never breaks.
It just gets slower, you know.
And so we could start doing a lot of engineering exploration up front.
But in the meantime, my favorite answer is eliminate waste.
You know, we've got all that idle power.
I want to evacuate it as fast as possible.
Yeah, there's a lot of low-hanging fruit here on Earth that we can utilize for the AI scaling.
Quick pause.
Quick 30-second thank you to our sponsors.
Check them out in the description.
It really is the best way to support this podcast.
Go to Lexfredman.com slash sponsors.
We got perplexity for curiosity-driven knowledge exploration.
Shopify for selling stuff online, element for electrolytes,
Finn for customer service AI agents, and quo for a phone system like calls, text, contacts for your business.
Choose wise to my friends.
And now back to my conversation with Jensen Kwong.
Do you think Nvidia may be worth $10 trillion at some point?
Let's ask it this way.
what does the future of the world look like where that's true i think that envydia's growth is is um uh extremely likely and in my mind
inevitable and let me explain why we're the largest computer company in history that alone should
beg the question why and the reason of course uh two reasons first two foundational technical reasons
The first reason is that computing went from being a retrieval-based file retrieval system.
Almost everything is a file.
We pre-write something.
We pre-record something.
We draw something.
We put it on the web.
We put it in a file.
And we use a recommender system, some smart filter, to figure out what to retrieve for you.
And so we were a pre-recording, human pre-recording, and file retrieving system.
That's what a computer is, largely.
to now AI computers are contextually aware,
which means that it has to process and generate tokens in real time.
So we went from a retrieval-based computing system
to a generative-based computing system.
We're going to need a lot more processing in this new world
than in the old world.
We need a lot of storage in the old world.
We need a lot of computation in this new world.
And so that's the first part of it.
we fundamentally changed computing in the way how computing is done.
The only thing that would cause it to go back is if this way of computation,
this way of computing generating information that's contextually relevant,
situationally aware that is grounded on new insight before it generates information,
this computation-intensive way of doing computing would only go back if it's not effective.
So for the last 10, 15 years while we're working on deep learning, if at any single moment,
I would have come to the conclusion that, that, you know what, this is not going to work out.
I think this is a dead end, or it's not going to scale, it's not going to solve this modality,
not going to be used in this application, then, of course, I would feel very differently about it.
But I think the last five years has given me more confidence than the last 10 years, the previous 10 years.
The second idea is computers, because it was a storage system, it was largely a warehouse.
We're now building factories.
Warehouses don't make much money.
Factories directly correlates with the company's revenues.
And so the computer did two things.
Not only did it change the way it did it, its purpose.
and the world changed.
It's no longer a computer.
It's a factory.
A factory is used for a generation of revenues.
We're now seeing,
not only is this factory generating products,
commodities that people want to consume,
we're seeing that the commodities are so interesting,
so valuable,
to so many different audiences,
that the tokens are starting to segment,
like iPhones.
You have free tokens, you have premium tokens, and you have several tokens in the middle.
And so intelligence, as it turns out, you know, it's a scalable product.
There's extremely high intelligence products, tokens that are used for specialized things.
People be willing to pay, you know, the idea that somebody's willing to pay $1,000 per million tokens is just around the corner.
It's not if, it's only when.
And so now we're seeing that the commodity that this factory makes is actually valuable.
And is revenue generating and profit generating?
Now the question is how many of these factories does the world need?
How many tokens does the world need?
And how much is society willing to pay for these tokens?
And what would happen to the world's economy?
If the productivity were to improve so substantially, what would happen?
Are we going to discover new drugs, new products, new services?
And so when you take these things in combination, I am absolutely certain that the world's GDP is going to accelerate in growth.
I'm absolutely certain the percentage of that GDP that will be used for computation will be 100 times more than the past.
because it's no longer a storage unit.
It's a product generation unit.
And so when you look at it in that context,
and then you back into what does Nvidia do
and how much of that new economics, new industry
would we have to benefit to address,
I think we're going to be a lot, lot bigger.
And then the rest of it, to me,
is, is it possible for
Nvidia to be a, you know,
$3 trillion revenues company in the near future?
The answer is, of course, yes.
And the reason for that
is because it's not limited by any physical limits.
There's nothing that I see that says,
you know, gosh, $3 trillion is not possible.
And as it turns out,
Nvidia supply chain is
the burden is shared by 200 companies.
and the fact that we scale out on the backs of with the partnership of this ecosystem,
the question is, do we have the energy to do so?
And surely we will have the energy to do so.
And so all of these things combined, that number is just a number.
And I still remember, Envidio was the first time we crossed a billion dollars.
I was reminded of a CEO who told me, you know, Jensen, it's theoretically impossible for a fabulous
semiconductor company to exceed a billion dollars. And, and I won't bore you with why, but, but
the, of course, it's illogical and there's a lot of evidence we're not. And then somebody told me,
you know, Jensen, you'll never be more than $25 billion because of some other company.
Somebody told me that you'll never be, you know, because. And then so, so, so, so,
those
those aren't
first principle
reason thinking
and the simple
way to think
about that is
what is it
that we make
and how large
is the opportunity
that we can create
now
Nvidia is not
in the market
share business
almost everything
that I just
talked about
don't exist
that's the part
that's hard
you know
if
Nvidia was a
was a
10 billion
company
trying to
take
invity a share, then it's easy to see for shareholders that, oh yeah, if they could just take
10% share, they could be this much larger. But it's hard for people to imagine how large we could be
because there's nobody I could take share from. You know? And so, so I think that that's one of the
challenges for the world is, is the imagination of the future. But I got plenty of time and I'll keep
reasoning about it and I'll keep talking about it. And every single GTC will begin.
become more and more real, you know,
and then more and more people to talk about
in one of these days, you know, we'll get there.
But 100% we'll get there.
Yeah, this view of, you know, token factories essentially,
this token per second per watt
and every token having value.
Like, it's an actual thing that brings value
and it brings different kinds of value,
different amounts of value to different people with value.
That's the actual product.
This really could be loosely thought of as the token.
And so you have a bunch of token factors
And it's very easy, first principles, to imagine a future, given all the potential things that AI can solve, that you're going to need an exponential number more of token factories.
Yeah.
And what's really interesting, the reason why I was so excited about it, the iPhone of tokens arrived.
Wait, are you saying Open Clause iPhone?
Yeah.
That's interesting.
Agents.
Yeah, agents.
True.
Agents in general.
The iPhone of tokens arrived.
It is the fastest growing application in history.
It went straight up.
Yeah.
Went straight up.
That says something.
Yep.
There's no question.
OpenClaw is the iPhone of tokens.
Yeah, there's something truly, as you know,
something truly special happening from about December.
Or people really woke up to the power of Clawed code, of Codex, of OpenClaw.
I mean, I've embarrassed to admit that in the way here in the airport,
I'm,
this is the first time I've done it's,
in public, I was programming
quote unquote by talking
my laptop. And I was
embarrassed because I was pretending like I'm talking
to a human colleague.
I'm not sure how I feel about the future where
everybody is walking
around, talking to their AI,
but it's such an efficient way
to get stuff done. And it's
more likely that your
AI is bothering you all the time.
And the reason for that is
because it's getting stuff done so fast.
It's reporting back to you. I got that
done. You know, what do you want me to do next? You know, that's the part that I think most people
don't realize is the person who's going to be chatting with them, texting them most, is their claws
or a lobster. What an incredible future. I read that you attribute a lot of your success to your
ability to work harder than anyone and withstand more suffering than anyone. So we can list
many of the things that entails. I mean, dealing with failure, the cost of engineering. The cost of
engineering problems we've talked about, the human problems, uncertainty, responsibility,
exhaustion, embarrassment, the near-death company moments that you've mentioned, but also the
pressure. Now as the CEO of this company that economies and nations strategize around, plan their
financial allocations around, plan their AI infrastructure around. How do you
deal with this much pressure? What gives you strength given how many nations and peoples depend on you?
I'm conscious about the fact that Nvidia's success is very important to the United States.
We generate enormous amounts of tax revenues. We establish technology leadership for our nation.
technology leadership is important for national security.
National security, not just in one aspect of national security,
all aspects of national security.
When our country is more prosperous,
we could do a better job with domestic policies
and helping social benefits.
Because we're generating so much reindustrialization in the United States,
we're creating mountains of jobs.
We're helping shift how we, how we,
how we build things back to the United States
in so many different plants,
chips, computers,
and of course these manufacturers.
I'm completely aware
that, and I have the benefit,
and this is a real,
a real gift
with mainstream investors,
teachers, policemen
who have somehow, for whatever reason,
invested in Nvidia or because they watch Jim Kramer,
bought some stock and now are millionaires.
And I am completely aware of that circumstance.
I'm aware of the circumstance that that Nvidia is central to a very large network of ecosystem
partners behind us and downstream from us.
And so the way I deal with that is exactly what I just did.
I reason about what it is.
what is it that we're doing?
What is it causing?
What's the impact that has other people
benefit, you know, positively or even
through great burden, for example, to supply chain?
And the question is,
therefore, what are you going to do about it?
In almost everything that I feel,
I break it down, I reason about, okay,
what's the circumstance?
What is changed?
What's hard?
And what am I going to do about it?
And I break it down, decompose the problem.
And the decomposition of these circumstances
turns it into manageable things that I can do.
And the only thing that I, after that I could do is,
did you do it?
Did you either do it or did you get somebody else to do it?
And if you didn't do it, you reasoned that you need to do it
and you didn't do it,
you didn't get anybody else to do it,
then stop crying about it,
you know?
And so,
and so,
I'm fairly,
I'm fairly,
uh,
tough on myself.
And,
but I also break things down so that,
so that,
um,
uh,
I don't panic.
Uh,
I can go to sleep because I've made the list of things that
needed to be done.
And I've made sure that everything that could put our company in harm's way,
could put my partners in harm's way,
put our industry in harm's way,
put our industry in harm's way.
harm's way, I've told somebody. Everything that I feel could put anybody in harm's way,
I've told someone. And I've told that someone who could do something about it. And so I've gotten
it off my chest or I'm doing something about it. And so after that, Lex, what else can you do?
So given all the insane, intense amount of suffering on the journey of building up in Vidaa,
you have you hit low point psychologically oh yeah oh yeah sure all the time all the time
and there you just break down the problem into pieces yeah see what you could do about it
and and part of you know lex part of it part of it is forgetting one of the most important
attributes of AI learning as you know is right systematic forgetting you need to know when to
forget some things. You can't memorize everything. You can't keep everything. And, you know,
you don't want to carry everything. One of the things that I do very quickly is I decompose the
problem, my reason about the problem, and I share the load with it. When I say I tell everybody,
I'm essentially sharing that burden as quickly as possible. Whatever worries me, tell somebody else.
Don't just keep it. You know, decompose, don't freak them out. Decompose the problem into smaller parts,
and get people to, and inspire them to be able to go do something about it.
But part of it is just forgetting.
You know, a lot of it is you've got to be tough on yourself.
You know, just come on, stop crying about it.
Let's get going.
You know, and then you get out of bed.
And then the other part is you're attracted to the next shiny light, the next future.
You know, the next opportunity, the next, okay, that's behind us.
What's next?
And it's a lot, I think, you know, you watch this with great athletes.
They just worry about the next point.
The last point is behind them.
The embarrassment, the, you know, the setback.
And because I do so much of my job publicly, you know, likes you do a fair amount of your job publicly, too.
And so I do a lot of my job publicly.
And so, you know, I say a lot of things that seem sensible at the time.
or funny at the time, mostly it's just because it's funny to me at the time.
And then, you know, you reflect on it. It's less funny.
But, but...
Yeah, trust me, I know.
But you basically allow yourself to be pulled by the light of the future.
Forget the past and just keep...
That's right.
Keep working towards that.
I mean, you did say, there's this kind of famous thing.
You said that if you knew how hard it would be to build Nvidia,
it turned out to be, what is it, a million,
times more hard than you anticipated,
that you wouldn't do it.
But isn't,
you know, when I hear that,
that's probably true about everything worth doing, right?
Exactly.
That is, by the way,
what I was trying to explain is that there's a,
there's an incredible superpower of being,
being, being,
have the mind of a child.
Yeah.
You know, and I say to myself,
oftentimes when I look at something,
and almost almost everything.
My first thought is, how hard can it be?
You know?
And so you get yourself into that mode.
How hard could it be?
And nobody's ever done it.
It looks gigantic.
It's going to cost hundreds of billions of dollars.
It's going to take, you know, all this.
And you just go, yeah, but how hard could it be?
You know?
Yeah.
How hard could be?
Yeah.
And so you got to get yourself into that state of mind.
You don't want to actually over-simulate everything and all the setbacks and all the trials and tribulations and all the disappointments.
You don't want to simulate all that in advance.
You don't want to know that.
You want to go into a new experience thinking it's going to be perfect.
It's going to be great.
It's going to be incredibly fun.
And then while you're there, you know, you need to have endurance.
You need to have grit so that when the setbacks actually happened.
And those setbacks are going to surprise you.
disappointments aren't going to surprise you. You know, the embarrassments are going to
surprise you. The humiliations are going to surprise you. You just can't, now you've got to turn on
the other bit, which is just forget about it. Move on. Keep moving. And to the extent that my
assumptions about the future and why the future is going to manifest, so long as those assumptions
and that input doesn't change,
or didn't change materially,
then I should expect that the output won't change.
And so my simulated output of the future
is still going to happen.
And if it's still going to happen,
I'm still going to go after it.
I believe it's going to, you know,
and so there's a combination of two or three
human characteristics,
the ability to go into an experience fresh-minded,
the ability to forget the setbacks,
the ability to believe in yourself,
you know, to believe what you believe and stay, stay true to that belief.
But you're constantly reevaluating.
This combination of three, four, five things, I think is really important for resilience.
And, you know, I'm fortunate that whatever, whatever life experiences led to this.
I've got kind of those four or five things.
You know, I'm always curious, always learning.
I'm always learning from everybody
you know
I'm always asking my
and because I'm humble about
about everything
I'm always thinking
gosh they did that so nicely
they did that so wonderfully
you know
I wonder what they're thinking through
how do they
you know so I'm simulating everybody
in a lot of ways
you know emulating almost everybody I watch
right you're empathetic
towards
towards everything that they do
that you're observing and respect
and so you're constantly learning
and you know.
You're now one of the wealthiest people on earth,
one of the most successful humans on earth.
Is it harder to be humble and to be able to,
do you feel the effect of money and power and fame
in making it harder for you to sort of be wrong in your own head
enough to hear out an opinion of somebody else
when it disagrees with you and learn from them,
those kinds of things?
Um, surprisingly no. And I would, I would actually go the other way. Because I do so much of my work publicly, when I'm wrong, pretty much everybody sees it.
You get humbled. Yeah. And, and when I'm wrong, when I'm wrong, or it didn't turn out that way or, you know, I mean, most of the things that that I say outside, I'm fairly certain about. And the reason for that is because, because it's going to impact some.
somebody else and I want to be quite concerned about that and quite circumspect about that.
For stuff that I'm reasoning about inside a meeting, you know, a lot of things could turn out
differently. But it doesn't ever stop me from reasoning. The way that I manage and lead,
I'm constantly reasoning in front of people. And even when I'm talking to you, you can kind of
see me kind of reasoning through things. And I want to make sure that you understand what I'm saying,
not because I told you, because I'm so humble about.
what I'm about to tell you,
I kind of show you the steps that I got there,
and then you could decide whether you believe what I said in the end.
And so I'm doing that all day long in meetings.
With all of my employees, I'm constantly reasoning through,
let me tell you how I see it, and I reason through it.
It gives everybody the opportunity to intercept and say,
I disagree with that part.
The nice thing about reasoning through things
and letting people interact with it
is that they don't have to disagree with your outcome.
they can disagree with your reasoning steps.
And they could pull me in different directions.
And then we can reason forward.
And so we're kind of, you know, collective path searching method.
And it's really fantastic.
Yeah, you have this way about you of when you're explaining stuff.
I can feel you actually reasoning on the spot about it with a constant open-mindedness
where you could, I could feel like I could steer your thinking.
Yeah.
And that's really beautiful that you've been able to maintain that after so many years of success and pain.
I think sometimes pain makes you close, closes you down a bit.
Yeah.
And I think to maintain.
Tolerance for embarrassment, I think is.
Yes, that's the tolerance.
I mean, that's a real thing.
Yeah.
There's many years of embarrassing yourself, even those meetings,
knowing that there's people around you where you declared.
one idea and it was shown that that idea was wrong and be able to admit that and to grow from
that. That's not, that's very difficult on a human level.
Yeah, well, you know, they knew I was, they knew that recently my first job was, was, you know,
cleaning toilets, so.
I'm glad you maintain that same spirit of Denny's, the work. I mean, that, that was beautiful.
Your whole journey from starting from Denny's is a beautiful one.
Let me ask you about video games, so I'm a big gaming fan.
So I have to say thank you to Nvidia
for many years of incredible graphics.
By the way, it is,
G-Force is our still, to this day,
our number one marketing strategy.
People learn about Nvidia
while they're in their teenage years.
And then they go to college
and they know who Nvidia is.
And then in the beginning, it's just, you know,
playing Call of Duty, you know, Fortnite.
And then later they're using Kuda.
And then later,
they're using in video and, you know,
blender and
just sew and
out of desk. I mean, I should say, I mentioned to a friend
that I'm talking with you.
He said, oh, they make great gaming
GPUs. Yeah, exactly.
Exactly.
You know, there's more to it.
But yeah, yeah, people really love the,
it really brought a lot of joy to a lot of people.
The hardware really brings these worlds to life.
there was some controversy around this
with DLSS-5.
Can you explain to me the drama around this?
I guess people,
gamers online are concerned that it makes games
look like AI slop.
Yeah.
What do you think of this drama?
Yeah.
I think their perspective makes sense
and I could see where they're coming from
because I don't love AI slop myself.
You know, all of the AI-generated
content increasingly
looks similar
and they're all beautiful
and I can, so I'm empathetic
towards what they're thinking.
That's just not what DLSS-5
is trying to do. I showed
several examples of it,
but DLSS-5
is 3D conditioned,
3D guided,
it's ground truth,
structured data guided.
So the artist determined the
geometry. We are completely
truthful.
to the geometry,
maintain so in every single frame.
It's conditioned by the textures,
the artistry of the artist.
And so every single frame
it enhances, but it doesn't change anything.
Now, the question is,
the question about enhancing,
DLSSF5 also lets,
because the system is open,
you could train your own models
to determine,
and you could even,
in the future prompt it, you know, I want it to be a tune shader. I wanted to look like this kind of, you know, so you can give it even an example, and it would generate in the style of that, all consistent with the artistry, you know, the style, the intent of the artist. And so all of that is done for the artist so that they can create something that is more beautiful, but still in the style that they want.
I think that they got the impression that the games are going to come out the way the games are, shipped the way they do, and then we're going to post-process it.
That's not what DLSS is intended to do.
DLSS is integrated with the artist.
And so it's about giving the artist the tool of AI, the tool of generative AI.
They could decide not to use it, you know.
I think people are very sensitive to human faces.
Yeah.
And we're now living in this moment, which I think is a beautiful.
one, which is people are sensitive to AI slop.
Yeah.
It puts a mirror to ourselves to help us realize that what we seek is imperfections,
what we seek is sometimes not perfect graphics.
It helps us understand what we find compelling in the worlds we create.
And that's beautiful.
And as long as it's tools that help us create those worlds.
Yeah, that's right.
That's right.
It's yet another tool.
And they want the generative models to generate the opposite of,
photo reel. Yeah, it'll do that too. And so it's just yet another tool. I think the
gamers might also appreciate that in the last couple years, we introduced skin shaders to the game
developers. And many of those games have skin shaders that include subsurface scattering
that make skin look more skin-like. And so the industry is,
game developers are looking for more and more and more tools to express their art.
And so this is just yet one more tool.
They get this, that way to use.
Ridiculous question.
What do you think is the greatest or most influential game ever made?
Maybe from Nvidia's perspective.
Doom.
Doom.
Unquestionably.
That was the start of the 3D.
I would say Doom from the intersection of the cultural implication as well as the industry,
turning a PC into a gaming device.
That was a very important moment.
Now, of course, flight simulation companies
were before it,
but they just didn't have the popularity
that Doom did to have made the industry
turned a PC from an office automation tool
into a personal computer
for families and gamers and things like that.
And so Doom was really impactful there.
From an actual game technology perspective,
I would say virtual fighter.
and so we're great friends
for both of them, you know.
And then there's games more recently,
I mean, Cyberpunk 2077,
really nice GPU
accelerated graphics.
Fully ray traced.
Fully ray traced.
Also, I like, I personally,
I'm a huge fan of Skyrim,
Elder Scrolls, and, you know,
it's been released a long, long time ago,
but people have released mods,
and they create these,
I mean, it's like,
a different game and it just allows me to replay the game over and over and it makes you realize
that you can re-experience in a totally new way the world you already love yeah so i do that all the time
one of my favorites is just walk around skyram we created this thing called rtx mod yeah it's a modding
tool awesome it allows it allows the community to inject the latest technology into an old game
of course like what makes a great video game is not just
graphics, it's also story and character development, but that's right.
Beautiful graphics can add to the immersion, the feeling like it's another place that you're
transported to.
What's, you said, I think accurately that the AGI timeline question rests on your definition
of AGI.
So let's, let me ask you about possible timelines here.
let's this ridiculous definition perhaps of what aGI is but an AI system that's able to essentially do your job
so run no start grow and run a successful technology company that's worth a good one or a one no more than a billion
more more than a billion dollars
So, you know, you know how hard it is to do all those components.
So how far are we away from that?
So we're talking about OpenClaw that does all the incredibly complex stuff that are required to,
to first of all, innovate, to find customers, to sell to them, to manage to build a team of some agents,
some humans, all that kind of stuff.
Is this 5, 10, 15, 20 years away?
I think it's now.
I think we've achieved AGI.
Do you think you can have a company run
by an AI system like this?
Possible.
And the reason for that is you said a billion
and you didn't say forever.
And so, for example,
it is not out of the question
that a claw was able to create
a web service,
some interesting little app
that all of a sudden, you know, a few billion people used for 50 cents.
And then it went out of business again shortly after.
Now, we saw a whole bunch of those type of companies during the Internet era,
and most of those websites were not anything more sophisticated
than what OpenClaug could generate today.
Interesting.
Achieve virality and monetize that virality.
Yeah.
It's just I don't know what it is, but I couldn't.
have predicted any of those companies at the time either, you know?
You're going to get a lot of people excited with that statement.
Yeah, I know.
It's like, what do you mean?
I could just launch an agent and make a lot of money.
Well, by the way, it's happening right now, right?
You know that when you go to China, you're going to see a whole bunch of people
teaching their, getting their claws to try to go out and look for jobs and, you know, do work, make
money.
And I'm not actually, I wouldn't be surprised if.
some social thing happened or somebody created a digital influencer super, super cute,
or some social application that, you know,
feed your little tomagachi or something like that.
And it becomes out of the blue an instant success.
A lot of people use it for a couple of months and it kind of dies away.
Now, the odds of, you know, 100,000 of those agents building Nvidia is zero percent.
and and then the one part that I won't do,
and I want to make sure we all do,
is to recognize that people are really worried about their jobs.
And I just want to remind them that the purpose of your job
and the tasks and the tools that you use to do your job
are related not the same.
I've been doing my job for 33 years.
I'm the longest running tech CEO in the world.
34 years.
And the tools that I've used to do my job has changed continuously in the last 34 years
and sometimes quite dramatically, you know, over the course of a couple to three years.
And the one story that I really want to make sure that everybody hears is the story,
the first job that computer scientists said, AI researchers said,
was going to go away, was radiology.
Because computer vision was going to a change.
achieve superhuman levels.
And it did.
Computer Vision was superhuman in 2019,
20, maybe a little bit later, 2020.
Okay?
And so it's been a long time since computer vision has been superhuman.
And so the prediction was radiologists would go away
because studying radiology scans was thing of the past.
AI will do that.
Well, they were absolutely right.
Computer vision is completely superhuman.
Every radiology platform and package today is driven by AI.
And yet, the number of radiologists grew.
And so the question is why, and we now have a shortage of radiologists in the world.
And so, one, the alarmist warning went too far and it scared people from doing this profession that is so important to society.
And so it did harm.
Now, why was it wrong?
The reason why is because the purpose of a radiologist, the purpose, is to diagnose disease and help patients and doctors.
Diagnosed disease.
And because we're able to study scans is so much faster now, you could study more scans, you could diagnose better, you could, you could impatient faster, we can see people more.
The hospitals are making more money.
You have more patients in the hospital.
You need more radiologists.
I mean, the amazing thing is it's so obvious this is what's going to happen.
The number of software engineers at Nvidia is going to grow, not decline.
And the reason for that is because the purpose of a software engineer
and the task of a software engineer of coding are related not the same.
I wanted my software engineers to solve problems.
I didn't care how many lines of code they wrote.
You know, but their job, their purpose of their job didn't change.
solving problems, working as a team, diagnosing problems, evaluating the result, looking for new problems
to solve innovation, connecting dots. You know, none of that stuff is going to go away.
So you think it's possible that let's even take coding, you think the number of programmers in the
world might increase? Yes. And the reason for that is this. What is the definition of coding?
I believe that the definition coding as of today is simply specific.
specifying, specification, and maybe if you want to be rather directive, you could even give it an
architecture of the software the year you wanted to write. So the question is, how many people
could do that? Describe a specification for a computer to go, telling the computer what to go build.
How many people? I think we just went from 30 million to probably one billion. And so every
every carpenter in the future will be a coder, except a carpenter with AI is also an architect.
They just increased the value that they could deliver to the customer. Their artistry just elevated
tremendously. I believe that every accountant is also your financial analyst, also your financial
advisor. So all of these professions have just been elevated. And if I were a
carpenter, I see AI, I would just completely go berserk. The services I can bring to my clients,
if I were a plumber, completely go berserk. And the people that are currently programmers and
software engineers, I think they're at the cutting edge of understanding intuitively how to communicate
with the agents using natural language in order to design the best kind of software. That's right.
So over time, they'll converge. But I think
there's still value in getting, I think,
learning how to program, like learning
what programming languages are,
the old kind of programming,
what are good practices for programming languages,
what are design principles for programming languages
for large software systems.
And the reason for that, Lex,
and you know that I'll just say for the audience,
I think the goal of specification,
the artistry of
specification, the goal and the artistry of it, it's going to depend on what problem you're
trying to solve. When I'm thinking about giving the company strategies and formulating
corporate directions and things that we should do, I describe it at a level that is sufficiently
specific that people generally understand the direction and it's actionable. It's
so specific enough that they can take action on it,
but I underspecify it on purpose
so that enable 43,000 amazing people
to make it even better than I imagined.
And so when I'm working with engineers,
when I'm working with people,
I think about what problem am I trying to solve,
who am I working with,
and the level of specification,
the level of architecture definition
relates to that.
And so everybody's going to have to learn
how where in the spectrum of coding they want to be.
Writing a specification is coding.
And so you might decide to be quite prescriptive
because there's a very specific outcome you're looking for.
You might decide that this is an area
you want to be much more exploratory.
And so you might underspecify
and enable you to go back and forth with the AI
to even push your own boundaries of creativity.
And so this artistry of where you are in the spectrum,
this is the future of coding.
But just to linger on it outside of coding,
I think a lot of people, rightfully so,
are worried about their jobs,
have a lot of anxiety about their jobs,
especially in the white collar sector.
I don't think any of us know what to do
with tumultuous times that always come,
when automations and new technology arrives.
And I just, first of all, I think we all need to have compassion
and the responsibility to feel sort of the burden
of what the actual suffering feels like for individual people
and families that lose their job.
I think whenever you have transformative technology,
like that's coming with artificial intelligence,
there's going to be a lot of pain.
And I don't know what to do about that pain.
hopefully it creates much more opportunities for those same people
for the same kind of job as the tooling evolves
and makes them more productive and makes them more fun.
Hopefully, as it does in the programming,
I've been having so much fun programming, I have to say.
Like, I've never had this much fun.
So hopefully it makes their job automates the boring parts
and makes the creative parts,
the ones that the human beings are responsible for.
But still there's going to be a lot of pain and stuff.
So my first recommendation before, and this is now how I deal with anxiety.
In fact, we just talked about it earlier.
Enormous anxiety about the future, enormous anxiety about the pressure, enormous anxiety about
uncertainty, I first break it down and then I'm going to tell myself, okay, there are some things
you can do something about, there are some things you can't do anything about, but for the stuff
that you can do something about, let's reason about it, and let's go do it.
If we were to hire a new college graduate today, and I have a choice between two, one that is no clue what AI is, and one that is expert in using AI, I would hire the one who's expert in using AI.
If I had an accountant, a marketing person, the one that is expert and using AI, supply chain, customer service, a salesperson, business development,
a lawyer, I would hire the one who is expert in using AI.
And so I would advise that every college student,
every teacher should encourage their student to go use AI.
Every college student should graduate and be an expert in AI.
And everybody, if you're a carpenter, if you're electrician,
go use AI.
Go see what it can do to transform your current job.
elevate yourself. If I were a farmer, I would absolutely use AI. If I were a pharmacist,
pharmacist, I would use AI. I want to see what it could do to elevate my job so that I could be the
innovator to revolutionize this industry myself. And so that would be the first thing that I would do.
And then I would also help them. It is the case that the technology will dislocate and will eliminate
many tasks
and because it will automate it,
if your job is the task,
if your job is the task,
then you're very highly
going to be disrupted.
If your
job's purpose includes
certain tasks, then it's
vital that you go learn how to use AI
to automate those tasks. And then there's
the world of spectrum in between.
And by the way, the beautiful thing about
AI, so the
the chat bot versions is you can break down,
you have anxiety and you can break down the problem by talking to it.
Like I've recently,
it's really just incredible how much you can think through your life's problems
and through,
and I don't mean like therapy problems.
I mean like very practically,
okay, I'm worried about my,
literally, I'm worried about my job,
what are the skills,
what are the steps I need to take,
how do I get better at AI?
Everything you just said,
you can literally ask and it's going to give you
A point by point of plan, I mean, it's just a great life coach, period.
I don't know how to use AI, and the AI goes, well, let me show you.
Exactly.
It's very meta, but it's kind of incredible.
So people definitely should.
You can't walk up to Excel and say, I don't know how to use Excel.
You're done.
I mean, that's really what AI has done for me in all walks of life,
is that initial friction of being a beginner of using a thing for the first time.
I can literally ask about any single thing,
what are the first steps I need to take?
That's right.
And that hand-holding that it does, removing the friction of all the experiences that the world offers, it's, you know, like I mentioned to you offline, you mentioned, I'm going to China and Taiwan.
So awesome.
I'm so excited for you.
Where do I, where do I, how do I, all of those questions immediately answered.
It is beautiful.
Well, when you go to Taiwan, just ask AI, what are Jensen's favorite restaurants in Taiwan?
Yeah.
And it will actually tell you.
Oh, yeah, yeah.
Is it accurate?
Yeah, yeah.
All right. It's all over Taiwan.
Well, you're a rock star over there.
And like we also mentioned offline, maybe our paths will cross, which would be really wonderful.
Computex. InvitaG.C. Taiwan.
Do you think there are some things about human nature, about human consciousness that is fundamentally non-computational?
Maybe something at chip, no matter how powerful, can never replicate?
I don't know if the chip will ever get nervous.
And that's the, you know, of course, the conditions by which that causes anxiety or nervousness or whatever emotion,
I believe that AI will be able to recognize those and understand those.
I don't think my chips will feel those.
And therefore, how that anxiety, how that feeling, how that excitement, how that,
that, how that, you know, all of those feelings manifest in human performance.
For example, extremely amazing human performance, athletic performance, you know,
average or lesser than average, that entire spectrum of human performance that comes out of
exactly the same circumstances for different people manifesting in different outcome,
manifesting in different performance, I don't think.
there's anything about anything that we're building that would suggest,
that two different computers being presented with all of exactly the same context,
of course it would produce statistically different outcomes,
but it's not because it felt different.
Yeah, the subjective, boy, there's something truly special about the subjective experience
that we humans feel.
Like I mentioned to you, I was pretty nervous talking to you.
I mentioned to you, that the hope, the fear, the anxiety, and just life itself, the richness of
life, how amazing everything is, how deeply we fall in love, how deeply a heart get broken,
how afraid we are of death and how much pain we feel when our loved ones pass away,
all of that, the whole thing. I know it's very hard to think AI being able to, a computational
device, being able to do that, but there's so many mysteries about this whole thing that we're
yet to uncover that I am open to be surprised.
I've been surprised a lot over the past few months and few years.
Scaling can create some incredible miracles in the space of intelligence.
Has been truly marvelous to watch, so I'm open to surprise.
And it's just really important to break down what is intelligence.
And the word, that word we use all the time.
It's not a mysterious word.
Intelligence has a meaning, you know?
And it's a system that, you know, it's something that we do that includes perception and understanding and reasoning and the ability to do plan.
And, you know, that loop, that loop is fundamentally what intelligence is.
Intelligence is not one word that is exactly equal to humanity.
And that's, I think, is really important to separate two.
We have two words for that.
I'm not, I don't over fantasize about,
and I don't over romanticize about intelligence.
Intelligence is, and people have heard me say it before,
I actually think intelligence is a commodity.
I'm surrounded by intelligent people,
and I'm surrounded by intelligent people,
more intelligent than I am in each one of the spaces that they're in.
And yet, I have a role in that circle.
it's actually kind of interesting
they're more educated than I am
they went to better schools than I did
they're deeper than in any
in the fields that they're in all of them
I have 60 of them
they're all superhuman to me
and somehow I'm sitting in the middle
orchestrating all 60 of them
and so you've got to ask yourself
what is it
about a dishwasher
that allows that dishwasher
to sit in the middle of superhumans
Does that make sense?
But that's my point.
My point is intelligence is a functional thing.
Humanity is not specified functionally.
It's a much, much bigger word.
And our life experience, our tolerance for pain, our determination, those are different words in intelligence.
And so the thing that I want to help the audience understand,
If I could give them one thing is intelligence is a word that we've elevated to very high form over time.
The word should really elevate is humanity.
Character, humanity, all of those things.
Compassion, generosity, all of the things that you say just now, I believe those are superhuman powers.
And that now intelligence is going to be commoditized because we've spoken about it.
The most important thing is your education.
Now, even when they said the most important things, your education, when you went to school, there's more than just knowledge that you gained.
But unfortunately, our society had put everything into one single word, and life is more than one word.
And I'm just telling you, my life would suggest that being lower on the intelligence curve than everybody around me doesn't change the fact I'm the most successful.
And so, and I think, I think that, that kind of is, I'm trying to hopefully to inspire everybody else that don't let this democratization of intelligence, this commoditization of intelligence, you know, cause you anxiety. You should be inspired by that.
Yeah, I think AI will help us celebrate humans more. And I'm certainly humanity and human first. And I think what makes this world incredible.
as humans, forever will be so.
And just AI is this incredible tool that makes us humans more powerful.
That's exactly right.
So much of the success of Nvidia and the lives of millions of people that I mentioned depend on you.
But you're just one human, like we mentioned.
Mortal, like all of us.
Do you think about your mortality?
Are you afraid of death?
I really don't want to die.
I have a great life. I've great family. I've really important work.
This is not a once-in-a-once-in-a-lifetime experience suggests that it has been experienced by many people, just not one person.
This is a once-in-a-humanity experience what I'm going through.
Nvidia is one of the most consequential technology companies in history.
We're doing very important work.
I take it very seriously.
And so some of the things that, of course, are practical things.
Like, how do we think about succession planning?
And I'm famous in saying that I don't believe in succession planning.
And the reason for that isn't because I'm in more.
mortal. The reason for that is because if you're worried about succession planning, if you're
worried all that anxiety of succession planning, then what should you do about it? Then you break it all
way back down. The most important thing you should do today, if you care about the future of your
company, post you, is to pass on knowledge, information, insight, skills, experience as often and
continuously as you can, which is the reason why continuously reason about.
everything in front of my team.
Every single meeting is about a reasoning meeting.
Every moment I spend inside a company,
outside a company,
is about passing on knowledge to people as fast as I can.
Nothing I learn ever sits on my desk
longer than, you know, a fraction of a second.
I'm passing that information, that knowledge.
Oh my gosh, this is cool.
Before I even finished learning all of it myself,
I've already pointing it to somebody else.
Get on this.
This is so cool.
You're going to want to learn this.
And so I'm constantly passing knowledge, empowering people,
elevating the capability of everybody around me,
so that the outcome that I seek, that I hope for,
is that I die on the job, you know,
and hopefully I die on the job instantaneously.
And there's no long periods of suffering, you know.
Well, from a fan perspective, given your, your,
extremely
your enormous positive impact
on civilization, of course, I hope
you keep going. But also it's just
fun to watch. What is doing
in your keynote, it's just the rate
of innovation. And I'm a huge fan
of engineering. It's so much incredible
engineering is continuously being
done by Nvidia. It's just fun to watch.
It's a celebration of humanity, it's a celebration
of great builders, a celebration of
great engineering. So it represents
something special.
So I hope you and NVIDIA keep going.
What gives you hope about this whole thing we've got going on?
About humanity, about the future of humanity.
When you look out, and you think about the future quite a bit,
when you look out 10, 20, 50, 100 years from now,
what gives you hope?
I've always had great confidence
in the kindness, the generosity,
the compassion, the human capacity.
I've always been extremely confident of that.
Sometimes more so than I should.
And I get taken advantage of,
but it doesn't ever cause me not to.
I start with always that people want to do good.
people want to help others.
And vastly I am proven right, constantly proven right, and often exceeds my expectations.
And so I have complete confidence in the human capacity.
I think the things that give me incredible hope is what I say.
see as I extrapolate, as what I see now is possible, and as I extrapolate based on the things
that we're doing, what will very likely happen. And that there's so many things that we want to
solve, there's so many problems we want to solve, there's so many things that we want to build,
there's so many good things that we want to do that are now within our reach and within the
reach of my my lifetime, you just can't possibly not be romantic about that. You know what I'm saying?
Yeah. What an exciting time to be alive. Yeah. Like truly, truly so. How can you not be romantic about
that? The fact that there is a, it's a reasonable thing to expect the end of disease. It's a
reasonable thing to expect. It's a reasonable thing to expect that pollution will be drastically
reduced. It's a reasonable thing to expect that traveling at the speed of light is actually in our
future. And then, you know, not for long distances, but short distances. You know, people ask me how.
Well, first of all, very soon, I'm going to put a humanoid on a spaceship, and it's going to be, you know,
my humanoid. And we're going to send it out, you know, a soon, you know, a
soon as possible. And it's going to keep improving and enhancing along the flight. And then when
it's time, all of the, all of my consciousness has already been, you know, so much of my life has
been uploaded in the internet, take all my inbox, take everything I've done, everything I've said.
You know, it's been becoming my AI. And I'm just, you know, when the time comes, you know,
we'll just send that at the speed light, catch up with my robot.
Oh, that's brilliant. I mean, but for me, that's,
sort of application focus.
But also for me, the curiosity, maxing perspective,
I just, all of those mysteries,
it's so much fascinating scientific questions there.
Understanding the biological machine is right around the corner.
It's not 10 years.
It's five years probably.
And then your biological machine,
the human mind and cracking physics,
theoretical physics open.
It's so exciting.
Explaining consciousness.
That one would be awesome.
And it's all within our reach.
Yeah.
Jensen, thank you so much for everything you've done over the years.
Thank you for everything you're doing for the world.
Thank you for being who you are.
I can tell you're a great human being.
And I wish you incredible success this year.
I can't wait.
As a fan, I can't wait to see what you do next.
And hopefully I'll see you in Taiwan.
And thank you so much for talking today.
Thank you, Lex.
I had a great time.
And also, if I could just say one more thing.
Yes.
And thank you for all the interviews that you do,
the depth, the respect that you go through with
and the research that you do to reveal for all of us,
the amazing people that you've interviewed over the years.
I've enjoyed them immensely.
And as an innovator, to have created this long form, unbelievable,
and yet, you know, it's just captivating.
So anyways, thank you for everything you do.
It means the world. Thank you, Jess.
Thank you, Lix.
Thank you for listening to this conversation with Jensen Huang.
To support this podcast, please check on our sponsors in the description,
where you can also find links to contact me, ask questions, give feedback, and so on.
And now, let me leave you with the words from Alan Kaye.
The best way to predict the future is to invent it.
Thank you for listening, and I hope to see you next time.
