Moonshots with Peter Diamandis - Ex-Google China President on How China Is Shaping the Future of AI w/ Kai-Fu Lee | EP #134
Episode Date: December 5, 2024In this episode, Kai-Fu and Peter discuss 01.AI’s growth, Chinese entrepreneurship, and how open-source AI can impact the world.  Recorded on Oct 19th, 2024 Views are my own thoughts; not Financ...ial, Medical, or Legal Advice. Kai-Fu Lee is the Chairman and CEO of Sinovation Ventures, a venture capital firm he founded in 2009 that manages over $2 billion in assets and focuses on fostering the next generation of Chinese high-tech companies. In 2023, Lee launched 01.AI, a startup that built AI applications tailored for China, including Wanzhi, a productivity assistant similar to Microsoft Office 365 Copilot. As a leading figure in artificial intelligence, Lee continues to shape the tech landscape in China, where he recently noted that Chinese AI models are only 6 to 9 months behind their U.S. counterparts. He has authored influential books such as AI Superpowers (2018) and AI 2041 (2021) and was named one of Time Magazine’s 100 most influential people in 2013. Earlier in his career, Lee held prominent positions in tech, including Vice President at Google, President of Google China (2005-2009), and Corporate Vice President at Microsoft (2000- 2005). He also founded and led Microsoft Research Asia from 1998 to 2000. Lee remains a highly respected thought leader in AI and continues to drive innovation in the field. Beago: https://www.beago.ai/ 01.AI: https://www.01.ai/ Kai-Fu’s X: https://x.com/kaifulee Kai-Fu’s LinkedIn: https://www.linkedin.com/in/kaifulee/ Pre-Order my Longevity Guidebook here: https://longevityguidebook.com/ ____________ I only endorse products and services I personally use. To see what they are, please support this podcast by checking out our sponsors: Get started with Fountain Life and become the CEO of your health: https://fountainlife.com/peter/ AI-powered precision diagnosis you NEED for a healthy gut: https://www.viome.com/peter Get 15% off OneSkin with the code PETER at  https://www.oneskin.co/ #oneskinpod Get real-time feedback on how diet impacts your health with https://join.levelshealth.com/peter/ _____________ I send weekly emails with the latest insights and trends on today’s and tomorrow’s exponential technologies. Stay ahead of the curve, and sign up now: Blog _____________ Connect With Peter: Twitter Instagram Youtube Moonshots
Transcript
Discussion (0)
I always remembered when I joined Google, Larry Page came and talked to us and he said
the ultimate search engine should be one where you ask a question and get a single correct
answer.
You've been at Apple, at Microsoft, president of Google China.
I love Google, but they're an engine that has been powered by advertising.
How long is that going to last?
Do you think that's going to survive?
I think there needs to be a business model flip at some point and Google will fail to
do that just as any innovator facing innovator's dilemma.
Your last book, you said something like the US will lead in breakthrough innovations,
but China is better in execution.
What does that mean?
The major technology breakthroughs were almost invariably
invented by Americans. Now when it comes to execution, it requires additional capabilities.
Before we get started, I want to share with you the fact that there are incredible breakthroughs
coming on the health span and longevity front. These technologies are enabling us
to extend how long we live, how long we're healthy. The truth is, a lot of the approaches
are in fact cheap or even free. And I want to share this with you. I just wrote a book
called Longevity Guidebook that outlines what I've been doing to reverse my biological
age, what I've been doing to increase my health, my strength, my energy. And I want to make
this available to my community at cost.
So, LongevityGuidebook.com.
You can get information or check out the link below.
All right, let's jump into this episode.
Hey, Kaifu. Good morning to you.
Hi, Peter. Good to be back.
Yeah, it's great to see you, my friend.
We're on the flip sides of the planet.
I can't wait till we're having this podcast and we're in different parts of the solar system
That'll be that'll be fun. Yeah, we need faster than light travel
you know, I I
have the fondest memories of
Coming and visiting you in in China in your different locations. And I have to say, you know, my takeaway from,
I used to come to China every year,
you would host a number of the Abundance 360 members
I'd bring with me, super gracious.
And I remember my takeaways were,
that number one, there was incredible work ethic from Chinese entrepreneurs.
And I remember you describing it, the work ethic, as 996.
Is that still the saying there?
Yes, yes, definitely.
Yeah, a good job was 9 a.m. to 9 p.m. six days a week. That was a good job was 9 AM to 9 PM, six days a week.
That was a good balance of life.
And the second thing I remember as a key takeaway
was, at least this was a decade ago,
and it's been some time, but that in the US,
entrepreneurs see the marketplace as the US, Decade ago and it's been some time but that you know in the u.s
Entrepreneurs see there the marketplace as the u.s
maybe Europe in in China the entrepreneurs saw the marketplace as
China and Europe and the u.s. There was a much more global view and I am curious if that's still
the the view in in entrepreneurial world in China today because I've heard you know and I'm seeing comments where
you know we're sort of like going into two parallel universes where products
develop in China or staying in China and products developed in the US or staying in the US.
How do you see that?
I'm curious.
Yeah, I think a lot of the B2B
is becoming very much a parallel universe.
It's hard to sell B2B,
especially given export control and geopolitical issues,
especially in the deep tech areas,
which you and I care deeply about.
B2C areas are much easier. Americans use
Shien and Temu and TikTok. And of course, Chinese use a lot of American products,
Mac, Apple, Windows, and so on. So that hasn't been as effective. I would also say the in
pursuit of scaling law, AGI, Gen.AI, while the efforts are separate, the
collaboration or at least the sharing of ideas are pretty strong in paper
publishing, open source, of course with the notable exception of open AI and
now Google who don't publish, but they don't do it for geopolitical reasons.
They don't want the competitor in seat. Yeah that is fascinating. We'll get
into that because you've taken very much an open source focused mindset and
there have been many that do.
I have to ask a question.
So you've seen, you've been at Apple, at Microsoft,
president of Google China.
You've seen so much and in innovations,
I mean, you're managing what, like three billion
in investments thereabouts.
So you've seen it all.
I mean, and of course, your two excellent books,
which we've discussed on my stages before.
I am curious about something,
and I'd love your opinion if you're willing,
which is, I love Google.
I love Google for many reasons,
what they've done, their investments investments their mindset of driving breakthroughs
But they're an engine that has been powered by advertising
And they've been able to reinvest that but what happens now when?
Gen AI is giving single solutions and the ad powered models
Is yeah, how long is that going to last?
Do you think that's going to survive?
Or is there going to have to be a business model flip for Google?
I think there needs to be a business model flip at some point and Google will fail to
do that just as any innovator facing innovator's dilemma. Because Google is critically dependent
on the advertising revenue.
And to do the flip would require going away,
losing all the revenue coming in,
going to an at best break even value proposition
of a single answer search engine,
and then rebuilding up the new business model,
whether it's
subscription or advertising and that's going to cause a roller coaster ride
mostly downwards for the stock price and that's not something that a publicly
listed company can do. It's kind of sad to see because Google is clearly in the
best position to reinvent search. It's handcuffs, right? It's handcuffs. Yeah. Your quarterly earnings reports are handcuffs.
Your stock market, your stockholders aren't gonna let you sacrifice or take
the risks and that's why a lot of companies that should have
jumped to the next generation of technology never made it. Right, right.
It's such a pity because Google clearly has
one of the world's top two AI engines and by far the world's number one
search engine and now we're talking about merging
two areas in which they're the best yet they can't win because of this
innovator's dilemma. It's unfortunate. Yeah.
I want to dive in during our conversation
into what you're doing now. So for the better part of 30 years, you were one of the lead investors
in technology in China. But you also invested across around the world, but typically in Chinese
markets. And you've flipped over from being an investor now
to being an entrepreneur.
And building O1.ai, what was the causative moment?
I mean, cause I am curious.
I mean, you've seen so many entrepreneurs
and so many deals.
I mean, just to, you you know what I wrote down here was
you know you've been investing in NLP tech, enterprise AI, AI driven financial solutions,
autonomous vehicles, autonomous software, a lot. But there was a moment in which you said okay
I need to go and build a company. Why? Yeah right. By the way let's call it zero one
dot AI. Okay. We were flexible before but now that OpenAI has taken the O one name.
We'll let them have it. We'll just be zero one. Zero one dot AI it is. Yeah yeah it really means
recreating the world with zero one using AI technologies. So yeah, I was
content doing investment in the early days of AI in the days of
deep learning, computer vision, convolution neural networks, I
was super excited because in my 40 year career in AI, I
basically saw two AI winters. And even the non winter days
were not that shiny. So
finally I saw wow this AI is becoming mature so I was very excited. It's not a
fad. No, no right and I was in the position of being a venture capitalist
so I figured the role I should play is invest because I'm you know older
hopefully in some ways wiser and experience and those technology and
those business.
So I invested in about 50 AI companies,
mostly in China, but some in the U S and they did well.
We now have 12 AI unicorns.
We'll soon have half a dozen IPOs from just from the AI companies,
being the first investor, which is pretty rare. So I kind of got on the
ride and enjoyed watching from the backseat, the excitement that my
entrepreneurs went through. So that was good. But then, Gen.ai came about. We
all saw and understood Gen.ai, but we didn't see how big it would be until OpenAI showed
us with Chagypti.
And at that moment, I realized that I could invest in the area, in China and elsewhere.
I looked at a bunch of GEN.ai companies, but then I realized that to start one that
late, to start a J&AI company after Chagy BD had taken over World by Storm,
would really be very very hard for any entrepreneur because you have a
you're behind by six or seven years and if you don't already have a team or
product or technologies, how could I fund these people?
Because China did not have really a lot of J&AI companies.
There was one or two at most.
And I just thought, hey, if anyone could do it,
maybe I could do it.
It would still be a long shot.
But given my years of experience and people
network and understanding of the technology and business, let's give
it a shot.
It may be a long shot, but I feel that when I'm really, really old, I'm old now, but when
I'm really, really old and look back-
Don't call yourself old, my friend.
You're still young and vibrant.
Okay.
When I'm 80, I would not want to look back and saying,
hey, I just had a cold fee, and I decided to invest. And even if I won with a great
investor building China's open AI, and I were an investor, I would still have regrets because
how could I my my love of my life not to participate in it this time.
And also I saw that if I did it, it really could work.
I would have a shot.
Others may have a shot too,
but I thought I would have a better shot
because I could pull a great team that have the right ideas.
And also I saw the world kind of dividing up into
parallel universes and that someone needed to do a GNI for China otherwise
the Chinese businesses and people will fall way behind and all the work that
Deng Xiaoping did to bring China forward could be lost if the world had GNI but
China didn't so I thought I would do it. It's interesting right because
If well, here's the question right if open AI and all the other
LLM systems had been equally available in China as they are in other parts of the world
Would you still have done it?
Good chance. I might. I would then have to think about the likelihood of success, right?
That becomes the main factor, not as much as
to helping the Chinese people in business to have a
solution even if it's not as good as OpenAI. So it might be 50-50
in that case,
but OpenAI decided not to make it available to China. So that was tough.
I mean, I think everyone would agree every country and every human is going
to need to have access to this infrastructure called GenAI. It's going to
be your consultant, your doctor,
your educator, your everything.
And it would be like denying a country access
to oxygen or electricity.
Yeah, that's why when we found the 01.ai,
our vision statement was to make AGI
beneficial and accessible and
That's very similar to open AI's
vision make make gen AI
Beneficial to humans, but we added accessible we wanted to
Stress the point that we want everyone to access it no matter where they live
What their nationality their income level etc. It is quite interesting that you've gone the open source road.
Can you speak to that a little bit? I mean, I just had Ray on the podcast and we're talking about open source and I had the Mozilla Foundation CEO on the podcast talking
about open source. Why aren't all companies going open source and what was your
motivation for open source? Right, well I think a smaller company, a newcomer,
really needs the open source community because having a hundred
people trying to compete with a thousand people at you know Google and starting
ten years late is a lose definite lose proposition if you don't somehow work
with the open source community to help each other make progress so that's just
out of a practical consideration.
Secondly, we saw a lot of good work in open source,
from universities, from Meta, from Microsoft, from NVIDIA.
And we couldn't start our company without these,
especially NVIDIA's Megatron, Microsoft's DeepSpeed.
Without these, it would have taken us much longer to start the company. especially NVIDIA's Megatron, Microsoft's Deep Speed.
Without these, it would have taken us much longer
to start the company.
So we said, well, if we're going to take from
the open source community,
well, we should rightfully give back.
Every model we make, except the most frontier model.
So we would keep closed source,
the very, very best model that we make,
everything else would become open source. And that is a way of giving back.
I know some companies open source everything, but we can't do that.
We do need a business model and some commercial advantage.
And also we decided the way we would do open source is through the Apache license. We would
not be asking people, you know, to get our approval for commercialization nor
would we put a limit that if you started making too much money or have too many
users, we have to sit down, talk commercial terms. We want everyone to
have what we have, just like we took from NVIDIA and Microsoft.
They're open source.
They didn't ask for anything back.
And we thought we also should not.
So you're putting everything up on a hugging face.
Yes.
For access and GitHub.
Did you see the movie Oppenheimer?
If you did, did you know that besides building
the atomic bomb at Los Alamos National Labs,
that they spent billions on bio-defense weapons, the ability to accurately detect viruses and microbes by reading their RNA?
Well, a company called Viome exclusively licensed the technology from Los Alamos Labs
to build a platform that can measure your microbiome and the RNA in
your blood. Now, Viome has a product that I've personally used for years called
Full Body Intelligence, which collects a few drops of your blood, spit, and stool
and can tell you so much about your health. They've tested over 700,000
individuals and used their AI models to deliver members critical health guidance
like what foods you should eat, what foods you shouldn't eat,
as well as your supplements and probiotics,
your biological age and other deep health insights.
And the results of the recommendations
are nothing short of stellar.
As reported in the American Journal of Lifestyle Medicine,
after just six months of following Biome's recommendations,
members reported the following,
a 36% reduction in depression, a 40% reduction in anxiety, a 30% reduction in diabetes,
and a 48% reduction in IBS. Listen, I've been using Viome for three years. I know that my oral
and gut health is one of my highest priorities. Best of all, Viome is affordable, which is part of my mission to
democratize health. If you want to join me on this journey, go to viome.com slash peter. I've asked
Naveen Jain, a friend of mine who's the founder and CEO of Viome, to give my listeners a special
discount. You'll find it at viome.com slash peter. I'd like to, you provided me a few charts, I want
to share one or two of these with the
audience at this point.
There's one in particular that talks about the impact of GDP of the PC era, the mobile
era and the AI era.
If we can put that up, let's talk about what that means.
How do you interpret this?
Yeah, I think, you know, if we look at the global GDP, it's interesting to note that the PC era brought about an uplift of the global GDP.
Then it kind of saturated, then mobile brought another, then it kind of saturated.
Now, there are many factors to the GDP. I don't claim PC and mobile
were the only factors, but they were clearly major factors that greatly enhanced productivity
and changed the way we worked. We as humans do more or less the same things for thousands of years.
We work, we play, we communicate, we learn, but the way in which we do them change from PC to mobile.
And I would say with AI, it would be in some sense
a similar change.
It would be a new platform that rather than infusing
a computer on every desktop or allowing anywhere,
anytime mobile access, we would make super intelligent AI
in every app.
And we would have apps would make super intelligent AI in every app and we would have apps
that are super intelligent that could do work for us, that could give us answers
and I think that is clear that this is not only the third platform revolution,
third productivity revolution, but by far the largest one because of how much
value it adds.
There was something you've said that you wrote about in your last book I
thought was fascinating. It's an approximate quote but you said something
like the US will lead in breakthrough innovations but China is better in
execution. Yes. Fascinating about that. Please elaborate. What does that mean for entrepreneurs here in the US,
entrepreneurs in China? Yeah, I think, you know, we've seen this through mobile revolution and
through the early days of deep learning and computer vision, AI revolution, that the major
technology breakthroughs were almost invariably invented by Americans. And that's because of the great university system,
research labs, and the culture that that encourages and rewards
risk taking and innovation, and and the amazing early stage
venture community that allows new ideas to be funded. And also the patent system, all of that,
basically started in the US and no wonder
that US is best at discovering new technologies
in the phases where new ideas were coming out.
Now, when it comes to execution,
it requires additional capabilities. I think the breakthrough innovation is less
important, but more important would be figuring out what to build and being focused on building
it and ask no questions and execute and work incredibly hard. In particular, asking really,
really smart people to say,
well, you're not writing papers, you're writing code to get this out there.
And to view success as a success of a product or a business, not as
success of a paper or an award.
So it's, it's the notion that a lot of AI researchers today are more
concerned about the sightings they get on their paper
versus making something that is generating revenue and users. And I see that. It was fascinating,
of course, that OpenAI turned on chat GPT and made a very successful first user product.
But I've criticized and many have Google for not having an app, right?
You have to do a lot of steps between something you're doing to get to a Gemini search.
So you think that the Chinese entrepreneurs
are better at execution and better at creating
something that's a beautiful user interface?
Is that the primary?
Yeah, but that interface is not just the artistic beauty,
but rather using all the principles,
again, invented in Silicon Valley,
the Lean Startup zero to one 1 that is building the MVP,
doing A-B tests and tweaking. And really it's the availability of the internet as
an instrumentation that allows entrepreneurs to no longer have to be
Steve Jobs. You don't have to know what the user is thinking. You just test it and
tweak it and if you work hard around the clock and measure the right things, improve
the right things, you will evolve to the right user interface. And that's where
hard work becomes the oil that makes the engine work and building a
good app. It's not just brilliance and insight and brilliance
insight would have favored the American entrepreneur. I mean there's another
thing going on with a lot of US AI companies which are called the race to
AGI and I'd like to show a short video of a statement by Sam Altman.
Let's go ahead and pull that up
and what you think about this.
Whether we burn 500 million a year
or 5 billion or 50 billion a year, I don't care.
I genuinely don't.
As long as we can, I think, stay on a trajectory
where eventually we create way more value
for society than that, and as long as we can figure out a way to pay the bills like we're making a GI
It's gonna be expensive. It's totally worth it. So what's your reaction to that?
How do you think about that?
Well, I think we all aspire to build a GI I know you do
I've wanted I've wanted it for 40 plus years that I've been in AI.
And we're very lucky to be at a point where scaling law appears to still be working,
meaning that if you throw 10 times more computing at the AGI problem, it gets smarter.
So it's tempting and logical to want to keep throwing 10 times more compute every one and a half years or so.
Of course, where this runs into some issues is, you know, is it a good investment once you're
putting 50 billion dollars into it? Are you sure there would not be diminishing returns?
And also, are you too focused on the breakthrough AGI and not enough on the application ecosystem.
Yeah, is it a bunch of researchers geeking out versus people building
businesses? You know, there's another chart here I want to bring up
and the question is as I watch the cost of models, in particular, inference models and such,
plummeting in price.
And the question is, is it a race to the bottom?
Is there, I mean, it's fascinating
that the single most powerful technology in the world
is effectively free.
So this is your chart, Kaifu.
Tell me what this means for you.
Yeah, actually, I wouldn't quite draw that conclusion about effectively free. It's eventually free.
So given a particular technology, let's say GPD-4 in this chart, it started, it was launched in May May 2023 at $75 per million tokens. And today it's at only $4.40.
And it's using a better version, GPD 4.0,
which is smaller, faster, better, and much cheaper,
roughly coming down 10 times a year.
And this is a good thing.
This is the market leader reducing price.
And it's reflecting the lower costs that they've accomplished
because GPU costs have come down, it's reflecting better technologies because you can get better
performance with a smaller model and just as we had Moore's law, I think the scaling law is a law because we do seem to see every year and a half or so it gets better and
also the cost comes down 10x per year so so I I would see a conclusion of wow
this is going great we just can sit around and then all the things we want
that are too expensive will become cheap.
But I would also have a word of caution because we are basically in the stratosphere,
going at turbo speed.
So one year is a really, really long time.
Just think, two, one years ago, we had no idea any of this was happening, right?
One year ago, we were still complaining,
chat GPT was hallucinating,
didn't know anything that's recent,
and all that's been changed.
So this industry is moving in one year,
what mobile probably would have taken
seven or 10 years to do.
So a year is a long time.
And I would also argue that even at $4.40,
GPT-4.0 is way too expensive for applications.
For example, let's take a look at, let's say AI Search, the example we talked about earlier.
I think if you took GPD-4.0 and used it to build an AI search, you would end up basically paying something like 10 cents or more per
search query and Google only makes 1.6 cents of revenue per search query so
you'd be on a fast road to bankruptcy due to the cost of GPD 4.0 and that's
not even counting you have to build a search infrastructure that's not even counting. You have to build a search infrastructure. That's just the LLM costs. It's just way way way way too high. I think more more precisely.
It is yeah around 10 cents. So to summarize you find yourself a situation where chat GPT is not available to China and
not available to China and the people of China need generative AI and you look around and say who better to do this than me?
And you've got a huge amount of experience so you jump in and you create 01.ai.
So what's the background there?
And one of the products you showed me was Bego, which is beautiful and fast and apparently cheap.
So let's talk about the history real quick of zero one dot AI and then let's jump into Bego.
Yeah, in zero one dot AI, we realized we were way behind open AI. We were perhaps seven years behind
when we founded it 17 months ago. It was only 17 months ago. I didn't have a team of engineers I had to use the first four or five months to hire people
But but even with that
Basically the playbook that I took was from my own book AI superpowers
We said we're not going to beat open AI at their own game. Can we build things very quickly?
sometimes the most challenging part of building something is proving
an unknown idea to be feasible, which ChaiGBD have done, which GBD4 and GBD40 and now GBD01 have done,
and that with the leaders in research demonstrating that something is feasible. That is all we need to know because when someone builds a nuclear bomb or puts a man on the moon,
for others to do it is much much easier because empirically it had been demonstrated.
So we were just saying now we just have to be more diligent, read more papers,
and work harder around the clock,
and leverage the strength that we have
as Chinese entrepreneurs and engineers,
and just go 996 or longer if needed,
until we get to products that are competitive
and efficient, right?
Because we probably can't win on accuracy,
but can we make the equally
accurate product much cheaper? Cheaper to train, cheaper to inference, cheaper to
train because we're poor, we don't have the 50 billion or 5 billion dollars Sam
Altman talked about, and cheap to inference because we want apps to run
lightning fast and that is what it would take for adoption
because you want fast and low cost of inference and in the last 17 months we
achieved all that. So you just said something that's fascinating which is
access to compute. I mean I think everybody imagines China has huge
infinite resources but it hasn't been the case in terms of GPUs.
And does that scarcity of resources cause you to think differently and not be lazy or
to be more innovative?
Yes.
One is just the difficulty of acquiring GPUs given the US restrictions.
But also we only raised a small amount of money so we couldn't afford 10,000 GPUs anyway.
And basically everything we've done, we did production runs on only 2,000 GPUs,
which is a small fraction of what the US companies are using Elon Musk
just put together a hundred thousand H100s and open AI, Google have even more.
It's impressive but we have basically you know less than 2% of their
compute but I'm a deep believer in efficiency, power of engineering, small teams working together,
vertical integration.
And I'm a strong believer that necessity is the mother of innovation.
So I have a team, I told them all we got is 2000 GPUs, we don't have 100,000.
I don't need you to, you know, invent the GPT-5. I want you to take a look at GPT-4, GPT-4.0.
And can we match that in 12 months, in five months?
And in the process of making it, all we have is 2000 GPUs.
You don't have a lot of compute. And when you make it, by the way,
if you train it in it
very efficiently we can also can we also have an inference that's very efficient
calling costing only a few percent to run it in apps so talk to me about your
product Beagle by the way I asked you earlier how to pronounce it and where
the name came from and I think it's worth repeating so people remember it better.
Beagle and Golden Retriever, right? It's a dog name.
Yeah. Yeah. Yeah. I mean, we all know the name Beagle B E A G L E.
It's a cute little dog and Golden Retriever.
But when the two of them make a little puppy, that puppy is called Beagle.
B E A G O. And okay. Beagle, Beagle, But when the two of them make a little puppy that puppy is called be go be a go and
Okay, be gold be gold be a GLE is very good at hunting and
Golden retrievers good at retrieving so it's an apt name for an AI search engine
But I should also point out that beagle is not zero one dot AI products
It is a product that I did venture build
and it's actually an American company and it uses a model very similar to the
model that zero one dot AI has built super fast super cheap thereby thinking
that AI search could be reinvented. So do you want to talk about zero one or you
want to jump into a little bit about Bego?
Which you prefer?
Oh, yeah, let me start with Zero One then we go into Bego.
So yeah, Zero One, I think, you know, this May we came up
with a very good model called eLarge and eLarge was a bit
behind GPD 4.0 which came out one day later. And we we at the
time were ranked number seven, and which is great number seven model number four
company just behind opening at Google and anthropic something to be really
let's pull up that chart one second. Good. So that's the May chart. But I want to
talk about what just happened in October,
because over between May and October, lots of models emerged.
E-Large was no longer competitive.
And but we had been working based on what I described as working super hard building to match GPD 4.0 and maybe even be faster. And that was accomplished in October
that we kind of took revenge on the GPD 4.0 May version,
which you can now see on this chart as just below us,
as number seven in the world.
We just beat them by a little bit.
So this is a case in point where we saw GPD 4.0,
we saw what it could do, and we knew it could be done,
and we said let's go do it, and basically with no hint on how it's done, we figured
out how to do it ourselves. I'm sure the methods are different, but we did match
their performance in just five months. Of course, in these past five months, other
great models came out, including new version
of GPT-4-0 and GROK and others.
So we came out in October with a tiny model called eLightning because we wanted to be
lightning fast.
This is a much smaller, much faster model, but it became number six model in the world
in number three company.
And we also surpassed Anthropic this time.
How big was your team building this?
The pre-training team is basically three or four people.
The post-training team was maybe 10 people.
The infrastructure team, maybe another 10 people.
So it's a 20 to 30 person project.
It's pretty small. And the thing
we're most excited about and proud and unique about is that we train this model,
the pre train only costs a little over $3 million. And this is this is 3% of what
GBT four cost to train. And we actually beat GBD4 in performance and the
inference cost is very very low it's around 10 cents per million tokens and
we sell it. Let's go to that chart there's a chart here it looks at
inference cost over time which is super impressive as well. So explain this chart here, please.
Yeah, so back in June, we were $1.40 per million tokens cost,
which was a lot lower than GPT-4.0 at the time,
which was I think about $10 price.
And then by September, we came up with a number of breakthroughs,
including new ways of
doing a mixture of experts and better inference and ideas of KV Cash management, etc. So we had
really a big breakthrough not in launching the e-lightning, because e-lightning was 1.14, the
cost of the previous e-large model and at 10 cents per million tokens
and and GPT-4.0 had also come down in price but it was $4.40
cents so to a developer... Yeah just to just to give folks who are just listening
and watching this back in June e-lightning was $1.40 per million tokens.
Today it's at $0.10 per million tokens.
And next June, it's expected to be $0.03 per million tokens.
It's a 50-fold decrease.
And comparing to GPT-4.0, which was at $4 four and a half bucks per million tokens so I mean we
are seeing this precipitous efficiency gains over time. Yes so you know another
way to look at it is GPT-4.0 dropped 10x in one year. We actually
succeeded in dropping 50x in the past year. So we feel we now have the most competitive, lowest price Lightning engine.
Our cost is 10 cents per million tokens.
Our price is only 14 cents per million tokens.
So we're also not taking a big margin.
So if you look at the performance of GPD 4.0 and Feed Lightning, their new version is a little better,
not a lot, only a little better, but they're 440 and we're 14 cents. Incredible. Is it all algorithmic gains?
There are a number of differences. I think we actually, I'm sure we use different
algorithms because we don't know what they use. We came up with our own.
But I'm saying the improvements you're making over time, are they algorithmic gains there?
Yeah, our performance gains going from eLarge to eLightning are using a new mixture of experts model and new ways of modeling and also getting more high quality
diverse data and also having super fast infrastructure so we can train multiple times
to learn more and to do research. The efficiency gains were also mostly by the super fast mixture of expert model,
but also by some inference advancements in terms of KVCache memory management.
As an example, the way we do our next-gen model design is not go invent a bunch of new things and go make them fast,
but it's from the get-go they have to be fast. So we would first ask the question, where do we project in
in four months, which is our product cycle, the best chips might be and
how do we get those chips to inference really fast? And oh these chips with a
lot of HBM, which is high bandwidth memories coming out.
And can we turn the inference problem
from a compute problem to more of a memory bound problem?
Then should we rewrite our inference engine?
Then how much RAM can we put out as a second layer memory?
How much SSD can we put on?
Can we construct a computer four months from now
that is super fast. I mean, it's made out of standard parts, but still it has a lot
of memory. Then we put the memory bound inference engine on top. Then we asked the modeling
team, in four months, what model can you build that fits perfectly into this box? Not too
large, not too small, use up all the memory but don't go too far
and use power of two.
So a lot of constraints for the researchers
which some companies might face reluctant researchers.
But in our case, we are all building a product.
We're one team in one direction.
So we all took, each team took the order
and then marched ahead and
Outcame a very accurate and a super fast model thanks to this vertical integration
From model to inference engine down to the hardware and memory. I
Love the old saying
You know innovation comes from thinking in a smaller and smaller box
When you put constraints on yourself.
Everybody, I want to take a short break from our episode to talk about a company that's
very important to me and could actually save your life or the life of someone that you
love.
The company is called Fountain Life and it's a company I started years ago with Tony Robbins
and a group of very talented physicians.
Most of us don't actually know what's going on inside our body.
We're all optimists.
Until that day when you have a pain in your side,
you go to the physician in the emergency room
and they say, listen, I'm sorry to tell you this,
but you have this stage three or four going on.
And you know, it didn't start that morning.
It probably was a problem that's been going on
for some time, but because we never look, we don't find out. So what we built at Fountain Life was
the world's most advanced diagnostic centers. We have four across the US today
and we're building 20 around the world. These centers give you a full-body MRI, a
brain, a brain vasculature, an AI enabled coronary CT looking for soft plaque,
DEXA scan, a grail blood cancer test, a full executive blood workup.
It's the most advanced workup you'll ever receive.
150 gigabytes of data that then go to our AIs and our physicians to find any disease
at the very beginning when it's solvable.
You're gonna find out eventually.
Might as well find out when you can take action.
Fountain Life also has an entire side of therapeutics.
We look around the world for the most advanced therapeutics
that can add 10, 20 healthy years to your life,
and we provide them to you at our centers.
So if this is of interest to you,
please go and check it out. Go to
fountainlife.com backslash Peter. When Tony and I wrote our New York Times
bestseller Life Force, we had 30,000 people reached out to us for Fountain
Life memberships. If you go to fountainlife.com backslash Peter, we'll put you to the top
of the list. Really it's something that is for me, one of the most important things I offer my entire family,
the CEOs of my companies, my friends.
It's a chance to really add decades onto our healthy lifespans.
Go to fountainlife.com backslash Peter.
It's one of the most important things I can offer to you as one of my listeners.
All right, let's go back to our episode.
One of the conversations over the last year is we're running out of data to really improve
the models.
What do you think about that?
Do you believe that to be the case?
I believe it has a bit of a dampening effect on how much we can expect scaling law to continue.
But I do think we have ways of getting more data,
just not as easily as it used to be.
Because the fact is that humans were smart
to create language as something that could be passed on
over millennia.
And we have, so once we start doing Gen.AI,
we took all the language data and put it up. Now every year we're
generating more language data, but clearly way less than the
total collection. So that is incrementally much, much slower.
But on the other hand, we have video data, we have audio data.
And also, we're going to have embodied AI gathering spatial
data. So those and also we have embodied AI gathering spatial data.
So those and also we have ways of creating synthetic data,
which is not as good but better than not using it.
So these are the ways I think we're trying to compensate for the fact that most actual data has been used.
If the outcome is, I think we will still get more data benefit, just not as much as we used to get.
Yep. Let's jump over to Tobigo. So it's a US company. Right. And why was it started in the US?
Is this something that was funded out of innovations? What's its mission? Talk to us about it.
Right, as I was building up zero one dot AI,
I ran into a lot of brilliant American engineers
and researchers.
They want to stay in America, but they liked my vision.
So I said, why don't I help you guys venture build a company?
So they build a company called Rimes Technologies,
and they build an excellent model, very similar in approach to the model in 0.1.AI.
And on that model, they added a lot of their unique multimodal
and launched Aria, which is an open source multimodal engine,
which is one of the best in the world, but only 3.5 feet.
So continuing
the tradition that companies that I helped build are very committed to open
source and on an advanced version of Aria they built an AI search engine. So I
was pleasantly surprised when they showed it to me. In fact I was blown away
by how good it was already and also how really really fast it was.
What's your hope with Bego? Do you, I mean, to come in and through an app become
the dominant search player? Yeah, I always remembered when I joined Google, Larry
Page came and talked to us and he said Google in this current form
is not the ultimate form. The ultimate search engine should be one where you
ask a question and get a single correct answer. That always kind of stuck with me
and when I venture built the Rhymes and Beagle team in the US we talked about
it and we feel that the time has come.
And in building such an engine, we also consider, well, first, on the mobile phone is a very
small screen, so you can't have all the tabs.
So doing a research-oriented, multi-link search exploration is very, very awkward.
And a single answer just makes so much sense.
But of course, the first issue with a single answer
is whether it is not correct, whether there's hallucination
or some errors.
So we work very hard to maximize factuality.
And Beagle is actually better in factuality
than a lot of the other AI engines measured
by objective third party queries.
So I think those really bring us one step closer to Larry Page's dream.
I think right now the team just wants more people to try it and they want, you know,
really knowledgeable, caring, smart people to try it first and give them the most feedback.
And it's great that I can be on your abundance program because those are the types of users your readers are.
Yeah and how do you possibly compete against you know companies who've got
billions of dollars in this field? Is it just that much better in
implementation that much better and alternate?
Yeah, it's a tough challenge. That's why very few companies go after this space.
The fact perplexity gained some ground is an indication people want something refreshing.
And also I think we're confident about Beagle's factuality and engagement.
It also has pictures inside the search making it more engaging and entertaining.
But also I think just the search players, particularly Google, and to some extent Bing,
will be hesitant to replace their search engine with the one answer engine because with one answer people don't look at ads and the ad revenue will
Yeah, they they will be I have to ask the question that probably a lot of people are thinking is this
Another tick-tock where it's a Chinese owned company and it's a way for you know, what people's imaginations that it's just a
way to get US data into this is a US based company a US owned company yes it's
actually both US based and US owned so it's not its employees are Americans
Singaporeans Taiwanese I myself I'm Taiwanese so it's it's not a Chinese home. It's quite
different from TikTok. Yeah I get that. I've been a fan of your work, Kaifu, for a
decade now and I've had the pleasure of calling you a friend. I you know I had
Elon Musk and Jeffrey Hinton and Ray Kurzweil, you know them all, on my stage at
Abundance 360 last year and it was a fascinating question that came up.
The probability that AI will be the greatest invention versus the probability that it will
destroy humanity, to put it very bluntly and I think Hinton and congrats to him on his Nobel and
Elon said yeah 80% it's good 20% were screwed.
Where do you come out on that?
Do you have an opinion on this and then how do we protect the downside in your mind?
Yeah, so well if we assume it's like a 10-year horizon, is that reasonable?
Yeah, I think all of it's going to play out in the next five to ten years.
I think if we get through the next, my belief, I don't know if you agree
with me, if we get through the next 10 years we're fine.
I totally agree, that's why I asked the question. Okay, so in the 10-year horizon, I would say we have a
5% chance of a disaster caused by AI and
35% chance of a disaster caused by humans using AI and 60% were good
Okay, so
60% were good. OK.
So now you've written an entire book on this,
but I'm going to ask you to provide some summarization.
What do we do?
How do we protect our downside?
The upside's fantastic.
Do you have any advice for parents, entrepreneurs,
leaders here?
How should they think about protecting our downside?
What would you, if you're head of, you know,
head of the world here, what do you do?
What do you think?
Well, I think a lot of technological risks
are best addressed by technologies.
Like when electricity went out,
the invention of circuit breakers,
when the internet went out, the antivirus.
So technologies are the best likely savior
to technological problems.
So I would encourage more computer scientists, AI people,
to not just work on the biggest, next big model,
or AI applications, or AI inference, or whatever,
but some percentage of them, the ones who feel a responsibility
and they're conscious asking them questions,
then they should jump into AI safety.
They define the various types of safeguards
and guide rails that will protect us.
I think to me, that's the most important thing. Regulation comes second. I would actually feel general AI only regulation to be
in unchartered territory and potentially not constructive. I think it would be
better to take existing laws, let's say laws about fraud, laws about other blackmail, and then apply it,
apply the use of AI to achieve those things. Laws about slander, laws about
theft. So we have lots of those laws. Those laws are effective, understood, so
apply them to people who use AI to break those laws, make sure the punishments
are equally if not more severe.
That would create some deterrence.
To start to regulate AI before it matures
and while it's changing,
by governments that are slow moving,
seems like a futile exercise.
Yeah, governments are linear or sub-linear at best.
You're gonna be joining me on stage in one of my panels
in Saudi Arabia in just 10 days or so.
Excited to see you there at the FII summit.
Yeah.
Yeah.
That'd be great.
One of the conversations we're going to have
is around the potential dangers of ASI,
artificial super intelligence. but before I go there
you know I would argue that we passed the Turing test many years ago and no
one really noticed it just just you know it's coming gone Will we know when we get to a GI I don't know that there's a good definition of a GI
And I don't even know if there's a good definition of digital super intelligence
I mean these are these are challenges when we talk about these words. Do you agree with that?
Yeah, I think a GI was created to mean that AI could do absolutely everything humans do
created to mean that AI could do absolutely everything humans do.
And that may not be the right definition because we can't yet project when AI will have love or, uh, or even when AI will be viewed as having love.
Those are still some distance away. Um,
but I think thinking generally that AI just means something overwhelms us that
does almost everything we do so much better,
even the most challenging intellectual tasks like inventing a new theorem or something in physics or
chemistry. So if we kind of extrapolate that to be the ASI or AGI, then I think it's highly likely that it will arrive in the next five to ten years
and we do need to put in the safeguards.
I imagine, we just saw a number of Nobel Prizes related to AI.
I have to imagine that in the very near future,
every single breakthrough in physics and math and chemistry is going to be
enabled or driven or connected to some AI models doing the work? Yes, absolutely.
Yeah, I met an economic professor on my recent trip to the Bay Area and he said he already treats GPT-01 as a graduate student.
One that's able to challenge him and find mistakes and he will teach and guide the student
and the student learns and the two of them are great partners in inventing new things.
So it's already happening and so I would add also economics to be perhaps in another area to what you listed
Do you say please and thank you to your AI when you're speaking to it? I
Do say please I don't say thank you
I'm not sure why
It's interesting, right? I find myself saying please sometimes. Thank you
for it.
But it is interesting to get how it's just a very small step
away from being a part of every aspect of our lives.
People are worried deeply about jobs.
You've made some prediction about loss
of white collar jobs. And of course course all the multimodal AI systems that you're speaking about are being embodied in robots
there is a number of
Fantastic humanoid robot companies coming out of China China needs robotics
for its aging population, right the one child policy has
significant implications for an aging population.
So talk to me about your prediction about jobs, your advice. I know you wrote about
this as well in your previous bestselling books.
But if you just have a few,
how should people think about jobs,
white collar jobs and then labor
with humanoid robots coming and what should,
what do they tell their kids?
What do they do for themselves?
How do you think about that?
Well, I think the fact is that the white collar jobs are going
to be the first set most challenged by AI because just software can replace a
lot of the routine and even non routine work and they will do so very rapidly in
the next five years. That's I think now universally believed when I wrote about it in my earlier books.
It was met with a mixed reaction.
So everybody expected it was going to be blue collar work leaving first.
Yes, yes, right.
Because it seems, you know, having intelligence in a white collar job is harder to replace, but it turns out dexterity is
harder to replace because that's not necessarily solved by the gene AI
technologies. So blue collar work, I think going from factory to the type of
caring work you talked about for elderly is going to happen as the next
wave. I'm among the more conservative on how fast that
will happen because I think these technologies are very expensive. Not only do you need the LLM
expense but also these robots that Elon Musk has shown are way out of any consumer's price range.
So they're kind of gonna be a while
because before the kinks are worked out,
before people accept them into their families
and lives and offices,
and the costs have to come down.
So I would project it will not be that soon.
You say that, but I'm an investor in Figure AI,
Brett Atcock's company, and Figure and Tesla both are projecting
around a $30,000 price tag.
Let's say it's $40,000 price tag.
If you could lease that, right,
and you lease it at $300 or $400 per month,
having a 24-7 employee for $400 a month
is pretty affordable.
I can see your point. I can see your point.
I'm still a little more cautious because you know especially used around the home.
You know the just clean my room is one thing that I would predict in three years.
This robot cannot even begin to do because every room is different.
Every definition of clean is different
and every family home is different.
But there are many other things,
like talking to the kids
or doing more household repetitive work,
that can be done.
30,000 I think is probably a reasonable price point for
Middle-class America but for China India other countries is still way too expensive real quick
I've been getting the most unusual compliments lately on my skin
Truth is I use a lotion every morning and every night religiously called One Skin. It was developed by four PhD women who determined a 10 amino acid sequence
that is a synolytic that kills senile cells in your skin.
And this literally reverses the age of your skin.
And I think it's one of the most incredible products.
I use it all the time.
If you're interested, check out the show notes.
I've asked my team to link to it below. All right, let's get back to the time. If you're interested, check out the show notes. I've asked my team to link to it below.
All right, let's get back to the episode. Kaifu, looking forward into 2025-2026,
would you mind sharing some predictions of what you might see in the AI world coming that would be surprising?
surprising? Sure, I think from the product side we will see basically every category of mobile app incorporate AI with many of them showing AI first disruptive types
of changes. So in other words in in about two years, every app we use
will be replaced by another app or super upgraded by the same app. I think agents will be a major
technology where we delegate what we want rather than just get an answer. I think multimodal will
not only be dramatically better because a lot of the super smart people are working on it
And we're seeing text to video something that would be fairy tale in my youth is now starting to happen
And I would also project that these great technology advancements will find real uses in applications
So it's not just You know, we just, I don't know if you saw this, we just mapped the connectome of the
Drosophila.
Did you see that?
There was a, they were able to map 50 million synaptic connections.
And it's a step away from a mammal.
Let's see the mouse.
I think we're going to probably see the connectome of a mouse done in the next year or two.
Where do you come out on the whole brain-computer interface world?
I'm still a little bit more, I would say I'm a bit more cautious about it. I think this is one of the areas where there's major
disagreement on how fast this is moving and what dangers it might
provide. And I think we need to be cautious, because it is
intrusive to our bodies. And it's a kind of a potential, a potentially slippery slope, right?
I think people can all get on board with treatments using interfaces,
and get on board with non-intrusive kinds of BCI.
But as we go deeper and deeper into reading our mind, creating scar tissues. And I think we just have to make sure that people
who are being experimented with are aware
of what kind of risks they have
and that the downside doesn't outweigh the upside.
Let's wrap up with a quick look at something I've heard you speak about, which is you were
there at the PC revolution, the mobile phone revolution, and the AI revolution.
And you've seen those progressions.
And I think you've modeled what the progression will be for AI.
Can you give me that summary because I think it's super useful for entrepreneurs listening if you're looking at starting a company in the AI world.
I mean, there's a lot of lessons learned from the PC in the mobile phone world, yes? Yeah, I think applications always follow a reasonable pattern of being replaced
because when a new technology, new form of content comes out, you as users have to first browse them,
then you make the content, then you search and organize the content,
then the content gets richer into multimedia, multimodal, then you can transact on the content, whether it's by payment, advertising,
e-commerce, or online to offline. Because these are the fundamental needs of
people and the progression of apps that I talked about go from fewer users to
more users, a small amount of usage to more usage, simple usage to complex usage.
So it's a really exciting iteration of better technology enables the next step on the trajectory.
More people use it, more money is made, more entrepreneurs, more funding, more GPUs, more
products, more models. So the virtuous cycle goes on and the most most exciting thing is
it took PC ecosystem easily 30 years to play out. It took mobile maybe 15 but
we're gonna see AI play out in the next three years or so. So if you jump in to
start the company this is the biggest roller coaster ride you can ever imagine.
What's your advice to the entrepreneur jumping in to start a company in the AI
space? What should they do, what should they
not do? Right, yeah I use the roller coaster as a
metaphor because I don't see it as a rocket ship purely upside.
There are a lot of challenges and traps. I would be cautious to
probably look at an app company
because that's the biggest space
with the most entrepreneurs
and the inference costs are coming down.
But when you think about starting an AI app company,
be cautious first about can you handle the inference costs
because those are too expensive.
Don't run out of money because inference cost is coming down.
So time your launches, time your product design
according to the technology you need
and when that technology will be low enough
in inference costs.
Secondly, be careful of the modeling companies
because we've seen companies like Jasper
who build great apps, but then the model
sucked all their know-how because they saw the data
So fine ensure that you don't do that
The last advice is all the models are getting better one day. They'll be close to a GI. Does that mean my app will be eventually?
Limited to a veneer and with very limited value
I don't necessarily think so because historically we've seen great platforms emerge,
but other apps can often build emotes.
The mode that TikTok, Instagram, and others have,
their value were not taken away
by the lower level transaction layers
or operating system layers or browser layers.
So the key is when you build an
app and gain some edge and don't sit on your laurels, think about how to build a
moat. That moat could be your brand, your user loyalty, user data or social graph,
things that we have seen or maybe new things in the AI era. Yeah, I like to
think about it and I'm curious if you agree,
when I'm evaluating an AI company to invest in it,
I'm looking at what unique data do they have
and what customers do they have a very close relationship with.
And everything else in the middle
will get demonetized and replaced over time. Yeah
Yeah, I think your advice is great for B2B. I was thinking more B2C. The two are definitely in concert
Yeah, what should people know about you know, let's let's turn to the last question which is
We live on one planet
and we've got sort of this
bipolar We live on one planet and we've got sort of this bipolar element of China versus the US and AI. And we have this split universe.
It would always be better to have alignment and everybody working well together.
But what advice do you have there? How should people think about this?
I mean, because it's a complicated way above my pay
grade.
And I don't want to put you in a situation where you're
talking about anything that you don't feel comfortable about,
but I can't not have the conversation of I see a lot
of people feeling like China is the enemy there,
or US is being monopolistic there.
How do we navigate the next 5, 10 years,
which are the most critical?
Yeah, there are some things that we're just not able to change.
They are what they are.
But I think each of us can make our own judgments and decide where we can reduce the
impact of this unfortunate geopolitical situation. For example, in the open source is one area where
all the countries collaborate equally and generously. Academic collaborations continue on.
Areas of collaborations not involved in the sensitive model
or semiconductor can still go on.
And I think connectivity in the world,
working people to people, business to business,
needs to go on, it has to be good.
Globalism has to be right.
Differences between governments is kind of like, you know, when our parents
have fights with another parents, we kids can still get along and, and do something
interesting and fun, right?
Agreed. You know, I like to say we all have the same biology. So a breakthrough in medicine
in China is the equivalent of a breakthrough in the Bronx. And we all share 24 hours in a day and seven days in a week.
That's something every single human has.
And so anything that gains us time efficiency in one place, gains time efficiency in another.
We share the same planet.
Yeah.
Which is facing its own challenges.
Yes.
Thank you for sharing time.
Super excited about the performance I've seen in Bego
and look forward to playing with it.
And thank you for joining me on Moonshots
to talk about your passion, your vision,
and congrats on going from the guy behind the curtain
to the guy in front of the company.
Thank you. Thank you so much.
Be well.
See you soon, my friend.
Bye.