a16z Podcast - The State of AI with Marc & Ben
Episode Date: June 14, 2024In this latest episode on the State of AI, Ben and Marc discuss how small AI startups can compete with Big Tech’s massive compute and data scale advantages, reveal why data is overrated as a sellabl...e asset, and unpack all the ways the AI boom compares to the internet boom. Subscribe to the Ben & Marc podcast: https://link.chtbl.com/benandmarc Stay Updated: Let us know what you think: https://ratethispodcast.com/a16zFind a16z on Twitter: https://twitter.com/a16zFind a16z on LinkedIn: https://www.linkedin.com/company/a16zSubscribe on your favorite podcast app: https://a16z.simplecast.com/Follow our host: https://twitter.com/stephsmithioPlease note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details please see a16z.com/disclosures.
Transcript
Discussion (0)
Most of the content created on the internet is created by average people.
And so kind of the content on average, as a whole, on average, is average.
The test for whether your idea is good is how much can you charge for it?
Can you charge the value?
Or are you just charging the amount of work it's going to take the customer to put their own wrapper on top of open AI?
The paradox here would be the cost of developing any given piece of software falls.
But the reaction to that is a massive surge of demand for software.
offer capabilities.
And I think this is one of the things that's always been underestimated about humans is our ability
to come up with new things we need.
There's no large marketplace for data.
In fact, what there are is there are very small markets for data.
In this wave of AI, big tech has a big compute and data advantage.
But is that advantage big enough to drown out all the other startups trying to rise up?
Well, in this episode, A16C co-founders, Mark and Dries
and Ben Horowitz, who both, by the way, had a front row seat to several prior tech waves,
tackle the state of AI.
So what are the characteristics that will define successful AI companies?
And is proprietary data the new oil, or how much is it really worth?
How good are these models realistically going to get?
And what would it take to get 100 times better?
Mark and Ben discuss all this and more,
including whether the venture capital model needs a refresh to match the rate of change
happening all around it.
And of course, if you want to hear more from Ben and Mark, make sure to subscribe to
the Ben and Mark podcast.
All right, let's get started.
It is kind of the darkest side of capitalism when a company is so greedy, they're willing
to destroy the country and maybe the world to like just get a little extra profit.
And they do it like the really kind of nasty thing is they claim, oh, it's for safety.
You know, we've created an alien that we can't control.
But we're not going to stop working on it.
We're going to keep building it as fast as we can,
and we're going to buy every freaking GPU on the planet.
But we need the government to come in and stop it from being open.
This is literally the current position of Google and Microsoft right now.
It's crazy.
The content here is for informational purposes only,
should not be taken as legal, business, tax, or investment advice,
or be used to evaluate any investment or security
and is not directed at any investor or potential investment.
investors in any A16Z fund.
Please note that A16Z and its affiliates may maintain investments in the companies discussed in this podcast.
For more details, including a link to our investments, please see A16Z.com slash disclosures.
Hey, folks, welcome back.
We have an exciting show today.
We are going to be discussing the very hot topic of AI.
We are going to focus on the state of AI as it exists right now in April of 2024.
And we are focusing specifically on the intersection of AI and company building.
So hopefully this will be relevant to anybody working on a startup or anybody at a larger company.
We have as usual solicited questions on X, formerly known as Twitter,
and the questions have been fantastic.
So we have a full lineup of listener questions, and we will dive right in.
So first questions, so three questions on the same topic.
So Michael asks, in anticipation of upcoming AI capabilities,
what should founders be focusing on building right now?
Gwen asks, how can small AI startups compete with established players with massive compute and data scale advantages?
And Alastair MacLay asks, for startups building on top of Open AI, etc., what are the key characteristics of those companies that will benefit from future exponential improvements in the base models versus those that will get killed by them?
So let me start with one point, Ben, and then we'll jump right to you.
So Sam Holtman recently gave an interview, I think maybe Lex Friedman or one of the podcasts.
And he actually said something I thought was actually quite helpful.
Let's see, Ben, if you agree with it.
He said something along the lines of, you want to assume that the big foundation models coming out of the big AI companies are going to get,
a lot better. So you want to assume they're going to get like a hundred times better. And as a
startup founder, you wanted them think, okay, if this current foundation models get 100 times better
is my reaction, oh, that's great for me and for my startup, because I'm much better off as a
result. Or is your reaction the opposite? Is it, oh, shit? I'm in real trouble. So let me just
stop right there, Ben, and see what you think of that as general advice. Well, I think generally
that's right, but there's some nuances to it, right? So I think that from Sam's perspective,
if he was probably discouraging people from building foundation models,
which I don't know that I would entirely agree with that
and that a lot of the startups building foundation models are doing very well.
And there's many reasons for that.
One is there are architectural differences,
which lead to how smart is a model,
there's how fast is a model,
there's how good is a model in a domain.
And that goes for not just text models,
but image models as well.
There are different domains, different kinds.
kinds of images that response to prompts differently.
If you ask Mid Journey and ideogram, the same question, they react very differently,
you know, depending on the use cases that they're tuned for.
And then there's this whole field of distillation where, you know, Sam can go build the
biggest, smartest model in the world, and then you can walk up as a startup and kind of do
a distilled version of it and get a model very, very smart at a lot less cost.
So there are things that, yes, the big company models are going to get way better,
kind of way better at what they are.
So you need to deal with that.
So if you're trying to go head-to-head full frontal assault, you probably have a real problem,
just because they have so much money.
But if you're doing something that's different enough or like, you know, different domain and so forth,
for example, you know, at Databricks, they've got a,
foundation model, but they're using in a very specific way in conjunction with their kind of
leading data platform. So, okay, now if you're an enterprise and you need a model that knows
all the nuances of how your enterprise data model works and what things mean and needs
access control and what needs to use your specific data and domain knowledge and so forth,
then it doesn't really hurt them if Sam's model gets way better. Similarly,
11 Labs with their voice model has kind of embedded into everybody.
Everybody uses it as part of kind of the AI stack.
And so it's got kind of a developer hook into it.
And then, you know, they're going very, very fast to what they do
and really being very focused in their area.
So there are things that I would say like extremely promising
that are kind of ostensibly,
but not really competing with OpenAI or Google or Microsoft.
So I think it sounds a little more coarse-grained than I would interpret it if I was building a startup.
Right.
Let's stick into this a little bit more.
So let's start with the question of, do we think the big models, the God models, are going to get 100 times better?
I kind of think so.
And then I'm not sure.
So if you think about the language models, let's do those because those are probably ones that people are most familiar with.
I think if you look at the very top models, you know, Claude and Open AI and Mistral and Lama,
The only people who I feel like really can tell the difference as users amongst those models are the people who study them.
You know, like they're getting pretty close.
So, you know, you would expect if we're talking 100x better, that one of them might be separating from each other a lot more.
But the improvement, so 100% better in what way?
Like for the normal person using it in a normal way, like asking it questions and finding out stuff?
Well, let's say some combination of just, like, breadth of knowledge and capability.
Yeah, like I think in some of them, they are, yeah.
Right, but then also just combined with, like, sophistication of the answers,
you know, sophistication of the output, the quality of the output,
sophistication of the output, you know, lack of hallucination, factual grounding.
Well, that, I think, is for sure going to get 100 times better.
Like that, yeah, I mean, they're on a path for that.
The things that are, so against that, right, the alignment problem where, okay,
Yeah, they're getting smarter, but they're not allowed to say what they know.
And then that alignment also kind of makes them dumber in other ways.
And so you do have that thing.
The other kind of question that's come up lately, which is kind of do we need a breakthrough
to go from what we have now, which I would categorize is artificial human intelligence
as opposed to artificial general intelligence, meaning it's kind of the artificial version
of us.
we've structured the world in a certain way
using our language and our ideas and our stuff
and it's learned that very well, amazing.
And it can do kind of a lot of the step that we can do
but are we then the asymptote
or you need a breakthrough to get to some kind of higher intelligence
or general intelligence.
And I think if we're the asymptote,
then in some ways it won't get 100 times better
because it's already, like, pretty good relative to us.
But, yeah, like, it'll know more things.
It'll hallucinate less.
On all those dimensions, it'll be 100 times better, I think.
You know, there's this graph floating around.
I forget exactly what the axes are,
but it's basically shows the improvement across the different models.
To your point, it shows an asymptote against the current tests
that people are using that's sort of like at or slightly above human levels,
which is what you would think if you're being trained on an entirely human data.
Now, the counterargument on that is, are the tests just too simple, right?
It's a little bit like the question people have run the SAT,
is if you have a lot of people getting 800s,
you know, on both math and verbal on the SAT,
is the scale too constrained.
Do you need a test?
They can actually test for Einstein.
Right, right.
It's memorized the tests that we have.
And it's great of them.
Right.
But you can imagine SAT that like really can detect gradations of people
who have like ultra high IQs who are ultra good at math or something.
You could imagine test for AI.
You know, you can imagine tests the test for reasoning above human levels when it assumes.
Yeah, well, maybe the AI needs to write the test.
Yeah, and then there's a related question that comes up a lot.
It's an argument we've been having internally, which is also I'll start to take some sort of
more provocative and probably more bullish, or as you would put it, sort of science fictiony
predictions on some of this stuff.
So there's this question that comes up, which is like, okay, you take an LLM, you train it on
the internet.
What is the internet data?
What is the internet data corpus?
It's an average of everything, right?
It's a representation of human activity.
Representation of human activity is going to kind of, you know, because of the sort of
distribution of intelligence of the population, you know, most of it's somewhere in the
middle.
And so the data set on average sort of represents the average human.
You're teaching it to be very average, yeah.
Yeah, you're teaching to be very average.
It's just because most of the content create on the internet is created by average people.
And so kind of the content on average, you know, as a whole on average is average is average.
And so therefore, the answer is our average, right?
You're going to get back an answer that sort of represents the kind of thing that an average 100 IQ,
you know, kind of by definition, the average human is 100 IQ.
It's IQ's index to 100 at the center of the bell curve.
And so by definition, you're kind of getting back the average.
I actually argue like that may be the case for the default prompt today.
Like you just ask the thing, does the Earth revolve around the sun or something?
You get like the average answer to that and maybe that's fine.
This gets to the point as well, okay, the average.
coverage data might be of an average person, but the data set also contains all of the things
written and thought by all the really smart people. All that stuff is in there, right? And all the
current people who are like that, their stuff is in there. And so then it's sort of like a
prompting question, which is like how do you prompt it in order to get basically, in order to basically
navigate to a different part of what they call the latent space to navigate to a different
part of the data set that basically is like the super genius part. And, you know, the way these things
work is if you craft the prompt in a different way, it actually leads it down a different path
inside the data set, gives you a different kind of answer. And here's another example of this.
If you ask it, write code to do X, write code to sort of list or whatever, render an image,
it will give you average code to do that.
If you say, write me secure code to do that, it will actually write better code with fewer
security holes, which is very interesting, right?
Because it's accessing a different purpose of training data, which is secure code.
And if you ask, you know, write this image generation thing the way John Carmack would write it,
you get a much better result because it's tapping into the part of the latent space
represented by John Carmack's code, who's the best graphics programmer in the world.
And so you can imagine prompting crafts in many different domains such that you're kind of unlocking
the latent super genius, even if that's not the default answer?
Yeah, now, so I think that's correct.
I think there's still a potential limit to its smartness in that.
So we had this conversation in the firm the other day where you have, there's the world,
which is very complex.
And intelligence kind of is, you know, how well can you understand, describe, or present
the world?
But our current iteration of artificial intelligence consists of humans structuring the world
and then feeding that structure that we've come up with into the AI.
And so the AI kind of is good at predicting how humans have structured the world
as opposed to how the world actually is, which is something more probably complicated,
maybe irreducible or what have you.
So do we just get to a limit where, like, it can be really smart,
but its limit is going to be the smartest humans as opposed to smarter than the smartest humans.
And then kind of related, is it going to be able to figure out brand new things,
you know, new laws of physics and so forth?
Now, of course, there are like one in three billion humans that can do that or whatever.
That's a very rare kind of intelligence.
So it still makes the AI extremely useful.
But they play a different role if they're kind of artificial humans
than if they're like artificial, you know, super duper mega humans.
Yeah.
So let me make the sort of extreme bull case for the hundred.
Because, okay, so the cynic would say,
the Sam Malman would be saying they're going to get 100 times better
precisely if they're not going to.
Yeah, yeah, yeah, yeah.
Right?
Because he'd be saying that basically in order to scare people into not competing.
Well, I think that whether or not they are going to get 100,
times better, Sam would be very likely to say that. Like, Sam, for those you don't know him,
is he's a very smart guy, but for sure he's a competitive genius. There's no question about that.
So you have to take that account. Right. So if they weren't going to get a lot better,
he would say that. But of course, if they were going to get a lot better, to your point,
he would also say that. Yes. Why not? Right. And so let me make the bull case that they are going
to get 100 times better or maybe even, you know, on an upper curve for a long time. And there's like
enormous controversy, I think, on every one of the things I'm about to say, but you can find
very smart people in the space who believe basically everything I'm about to say. So one is,
there is generalized learning happening inside the neural networks. And we know that because
we now have introspection techniques where you can actually go inside and look inside the
neural networks to look at the neural circuitry that is being evolved as part of the training
process. And, you know, these things are evolving, you know, general computation functions.
There was a case recently where somebody trained one of these on a chess database and, you know,
just by training lots of chess games, it actually imputed a world model of a chess board, you know,
inside the neural network and, you know, that was able to do original moves.
And so the neural network training process does seem to work.
And then specifically, not only that, but, you know, meta and others recently have
been talking about how so-called overtraining actually works, which is basically continuing
to train the same model against the same data for longer, you know, putting more and more
compute cycles against it.
You know, I've talked to some very smart people in the field, including there,
who basically think that actually that works quite well.
The diminishing returns people were worried about about more training.
And they proved it in a new Lama release, right?
That's the primary technique they use.
Yeah, exactly.
Like one guy in the space basically told me,
basically he's like, yeah,
we don't necessarily need more data at this point
to make these things better.
We maybe just need more compute cycles.
We just train it 100 times more
and it may just get actually a lot better.
So, Juan, the labeling,
it turns out that supervised learning
ends up being a huge boost to these things.
Yeah.
So we've got that.
We've got all of the kind of, you know,
let's say rumors and reports
of various kinds of self-improvement loops,
you know, that kind of underway.
And most of the sort of super advanced practitioners
in the field think that there's now
some form of self-improvement loop that works,
which basically is, you basically get an AI to do what's called chain of thoughts.
You get it to basically go step by step to solve a problem.
You get it to the point where it knows how to do that.
And then you basically retrain AI on the answers.
And so you're kind of basically doing sort of a forklift upgrade across cycles of the reasoning capability.
And so a lot of the experts think that sort of thing's starting to work now.
And then there's still a raging debate about synthetic data,
but there's quite a few people who are actually quite bullish on that.
Yeah.
And then there's even this tradeoff.
There's this kind of dynamic where like LLMs might be okay at writing code,
but they might be really good at validating code.
you know, they might actually be better at validating code than they are at writing it.
That would be a big help.
Yeah, well, but that also means like AOS may be able to self-validate your own code.
Yeah, yeah.
They can validate their own code.
And we have this anthropomorphic bias that's very deceptive with these things because
you think of the model as an it.
And so it's like, how could you have an it that's better at validating code
the writing code, but it's not an it.
What it is is, it's this giant latent space.
It's this giant neural network.
And the theory would be there are totally different parts of the neural network for writing
code and validating code.
And there's no consistency requirement whatsoever that the network would be equally good at both of those
things. And so if it's better at one of those things, right? So the thing that it's good at might be
able to make the thing that it's bad at better and better. Right, right, right, right. Sure, sure.
Right. Sort of a self-improvement thing. And so then on top of that, there's all the other things
coming, right, which is it's everything, there's all these practical things, which is there's an
enormous chip constraint right now. So every AI that anybody uses today is its capabilities
are basically being gated by the availability of chips. But like that will resolve over time.
You know, there's also, if your point out like data labeling, there is a lot of data in these
things now, but there is a lot more data out in the world. And there's, you know, at least in theory,
some of the leading AI companies are actually paying to generate new data. And by the way,
even like the open source data sets are getting much better. And so there's a lot of like data
improvements that are coming. And then, you know, there's just the amount of money pouring into
the space to be able to underwrite all this. And then by the way, there's also just the
systems, right, which is a lot of the current systems, you know, we're built by
by science. And now the really world class engineers are showing up and tuning them up and
getting them work better. And, you know, maybe that's not a, maybe that's not a, which makes
training, by the way, way more efficient as well, not just inference, but also training.
Yeah, exactly. And then even, you know, another improvement area is basically Microsoft
released their Phi small language model yesterday. And apparently it's competitive. It's a very
small model competitive with much larger models. And the big thing they say that they did was they
basically optimized the training set. So they basically de-duplicated the training set.
They took out all the copies and they really optimized on a small amount of training data,
on a small amount of high-quality training data, as opposed to the larger amounts of low-quality
data that most people train on. You add all these up and you've got eight or ten,
different combination of sort of practical and theoretical improvement vectors that are all in
play. And it's hard for me to imagine that some combination of those doesn't lead to like
really dramatic improvement from here. I definitely agree. I think that's for sure going to happen.
Right. Like if you were, so back to Sam's proposition, I think if you were a startup and you
were like, okay, in two years I can get as good as GPT4, you shouldn't do that. Right.
When I hit, that would be a bad mistake. Right. Right. Well, this also goes to, you know, a lot of when
entrepreneurs are afraid of. Well, I'll give you an example. So a lot of entrepreneurs
this thing they're trying to figure out, which is, okay, I really think, I know how to build a
SaaS app that harnesses an LLM to do really good marketing collateral. Let's just make a very similar,
a very similar thing. And so I build a whole system for that. Will it just turn out to be that the
big models in six months will be even better in making marketing collateral just from a simple prompt
such that my apparently sophisticated system is just irrelevant because the big model just does it.
Yeah. Yeah. How are we just talk about that? Like apps. You know, another way you can think about it is
the criticism of a lot of current AI app companies
is their quote-unquote, you know, GPT wrappers.
There's sort of thin layers of wrapper
around the core model, which means the core model
could commoditize them or displace them.
But the counter argument, of course, is it's a little bit
like calling all, you know, old software apps,
you know, database wrappers, you know, wrappers around a database.
It turns out like actually wrappers around a database
is like most modern software, and a lot of that
actually turned out to be really valuable.
And it turns out there's a lot of things to build
around the core engine.
So, yeah, so Ben, how do we think about that
when we run into companies thinking about building apps?
Yeah, you know, it's very tricky question.
because there's also this correctness gap, right?
So, you know, why do we have co-pilots?
Where are the pilots?
Right, where are the I...
There's no AI pilots.
They're only AI co-pilots.
There's a human in the loop on absolutely everything.
And that really kind of comes down to this,
you know, you can't trust the AI to be correct
in drawing a picture or writing a program or, you know,
even like that, writing a course.
brief without making up citations, you know, all these things kind of require a human and
it kind of turns out to be like fairly dangerous to not. And then I think that so what's happening
a lot with the application layer is people saying, well, to make it really useful, I need to turn
this co-pilot into a pilot. And can I do that? And so that's an interesting and hard problem.
And then there's a question of, is that better done at the model level or at some layer
on top that, you know, kind of teases the correct answer out of the model, you know,
by doing things like using code validation or what have you, or is that just something that
the models will be able to do? I think that's one open question. And then, you know, as you get
into kind of domains and, you know, potentially wrappers on things, I think there's a different
dimension than what the models are good at, which is what is the process flow, which is kind
of in database for all to. So on the database kind of analogy, there is like the part of the task
in a law firm that's writing the brief, but there's 50 other tasks and things that have to be
integrated into the way a company works, like the process flow, the orchestration of it. And maybe
there are, you know, on a lot of these things, like if you're doing video production, there's
many tools, or music even, right?
Like, okay, who's going to write the lyrics?
Which AI will write the lyrics and which
AI will figure out the music?
And then, like, how does that all
come together and how do we integrate it
and so forth? And those
things tend to
just require a
real understanding of
the end customer and so forth
in a way. And that's
typically been how, like, applications have
been different than platforms in the past.
It's like there's real knowledge
about how the customer using it wants to function
that doesn't have anything to do
with the kind of intel or is just different
than what the platform is designed to do.
And to get that out of the platform
for a kind of company or a person
turns out to be really, really hard.
And so those things, I think, are likely to work,
you know, especially if the process is very complex.
And it's something that's funny.
As a firm, you know, we're a little more hardcore
technology-oriented. And we've always struggled with those, you know, in terms of, oh, this is
like a some process application for like plumbers to figure out this. And we're like, well,
where's the technology? But, you know, a lot of it is how do you encode, you know, some level of
domain expertise and kind of how things work in the actual world back into the software?
I often think of Intel founders that you can think about this in terms of price, you can
kind of work backwards from pricing a little bit, which is to say sort of business value
and what you can charge for, which is, you know, the natural thing for any technologists to do
is to kind of say, I have this new technological capability and I'm going to sell it to people
and, like, what am I going to charge for it? It's going to be somewhere between, you know, my cost
of providing it and then, you know, whatever markup I think I can justify, you know,
and if I have a monopoly providing it, maybe the markup's infinite. But, you know, it's kind of this,
it's a technology forward, you know, kind of supply forward, you know, pricing model.
there's a completely different pricing model for kind of business value backwards and or sort of,
you know, so-called value pricing, value-based pricing. And that's, you know, to your point,
that's basically a pricing model that says, okay, what's the business value to the customer of the
thing? And if the business value is, you know, a million dollars, then can I charge 10% of that
and get $100,000, right, or whatever. And then, you know, why is it cost $100,000 as compared to
$5,000 is because, well, because to the customer is worth a million dollars. And so they'll pay
10% for it.
Yeah, actually, so a great example of that, like, we've got a company in our portfolio,
Crest AI, that does things like debt collection.
Okay, so if I can collect way more debt with way fewer people with my, you know,
it's a co-pilot type solution, then what's that worth?
Well, it's worth a heck of a lot more than just buying,
an open AI license because an open AI license is not going to easily collect debts or
kind of enable your debt collectors to be massively more efficient or that kind of thing.
So it's bridging that gap between the value.
And I think you had a really important point.
The test for whether your idea is good is how much can you charge for it?
Can you charge the value?
Or are you just charging the amount of work it's going to take the customer to put their
own wrapper on top of open AI. Like that's the, that's the real test to me of like how deep
and how important is what you've done. Yeah. And so to your point on like the kinds of,
you know, the kinds of businesses that technology investors have had a hard time with,
you know, kind of kind of thinking about, you know, maybe accurately is sort of, it's the
company that is, it's a, it's a vendor that has built something where it is a specific solution
to a business problem where it turns out the business problem is very valuable to the
customer. And so therefore, they will pay a percentage of the value provided back to back in the
terms for price for the software. And that actually turns out you can have businesses that are not
very technologically differentiated that are actually extremely lucrative. And then because that
business is so lucrative, they can actually afford to go think very deeply about how technology
integrates into the business, what else they can do. You know, this is like the story of a salesforce.com,
for example, right? And by the way, there's kind of a
a chance, a theory that the models are all getting really good.
There are open source models.
They are like that are awesome.
You know, Lama, Mistral, like, these are great models.
And so the actual layer where the value is going to accrue is going to be like tools,
orchestration, that kind of thing, because you can just plug in whatever the best model
is at the time, whereas the models are going to be competing, you know, in a death battle
with each other and be commoditized down to the, you know, the cheapest one wins and that kind of
thing. So, you know, you could argue that the best thing to do is to kind of connect the power
to the people. Right. Right. So that actually takes us to the next question, and this is a two and one
question. So Michael asks, and I'll say these are diametrically opposed, which is why I paired
them. So Michael asks, why RVC is making huge investments in generative AI startups when it's
clear these startups won't be profitable anytime soon, which is a loaded, loaded question,
but we'll take it. And then Kaiser asks, if AI deflates the cost of building a startup,
how will the structure of tech investment change? And of course, Ben, this goes to exactly what
you just said. So it's basically the questions are diametrically opposed because if you squint
out of your left eye, right, what you see is basically the amount of money being invested
in the foundation model companies kind of going up to the right at a furious pace. You know,
these companies are raising hundreds of millions, billions, billions of dollars. And it's just like,
oh my God, look at these sort of capital, you know, sort of, I don't know, infernos, you know,
that hopefully will result in value at the end of the process. But my God, look at how much money
is being invested in these things. If you squint through your right eye, you know, you think, wow,
that now all of a sudden it's, like, much easier to build software. It's much easier to have
a software company. It's much easier to, like, have a small number of programmers writing
complex software because they've got all these AI co-pilots and all these automated, you know,
software development capabilities that are coming online. And so on the other side, the cost of
building an AI like application startup might, you know, crash and it might just be that like
the, you know, the Salesforce.com might cost, you know, a tenth or a hundredth or a thousandth
amount of money that it took to build, you know, the old database driven Salesforce.com. And so, yeah,
so what do we think of that dichotomy, which is you can actually look, you can actually look
out of either eye and see either cost to the moon as like for startup funding or cost actually
going to zero? Yeah, well, like, so it is interesting. I mean, we actually have companies in both
camps, right? Like I think probably the companies that have gotten to profitability, the fastest,
maybe in the history of the firm have been AI companies. There have been, you know, AI companies
in the portfolio where the revenue grows so fast that it actually kind of runs out ahead of
the cost. And then there are, like, you know, people who are in the foundation model race who
are raising hundreds of millions, you know, even billions of dollars to kind of keep pace and so forth.
They also are kind of generating revenue at a fast rate.
The headcount and all of them is small.
So I would say, you know, where AI money goes, and even, you know, like if you look at OpenAI,
which is the big spender in startup world, which, you know, we are also investors and is, you know,
headcount-wise, they're pretty small against their revenue.
Like, it is not a big company headcount.
Like, if you look at the revenue level on how fast.
they've gotten there. It's pretty small. Now, the total expenses are ginormous, but they're going
into the model creation. So it's an interesting thing. I mean, I'm not entirely sure how to
think about it. But I think, like, if you're not building a foundation model, it will make you
more efficient and probably gets profitability quicker. Right. So the counter, and this is a very
bullish counterargument, but the counterargument to that would be basically that falling costs for, like,
new software companies are a mirage.
And the reason for that is this thing in economics
called the Jevins Paradox,
which I'm going to read from Wikipedia.
So the Jevins Paradox occurs
when technological progress increases the efficiency
with which a resource is used,
reducing the amount of that resource necessary
for any one use.
But the falling cost induces increases in demand,
right, elasticity, enough that the resource use
overall is increased rather than reduced.
Yeah, that's certainly possible.
Right.
And so this is, you see,
You see versions of this, for example, you build in your freeway, and it actually makes traffic jams worse, right?
Because basically what happens is, oh, it's great, now there's more roads, now we can have more people live here, we can have more people that, you know, we can make these companies bigger, and now there's more traffic than ever, and now there's more traffic's even worse.
Or you saw the classic example is during the Industrial Revolution coal consumption, as the price of coal drops, people use so much more coal that actually the overall consumption actually increased.
And people are getting a lot more power, but the result was the use of a lot more coal in the paradox.
And so the paradox here would be, yes, the cost of developing any given piece of software falls,
but the reaction to that is a massive surge of demand for software capabilities.
And so the result of that actually is, although it looks like starting software companies,
the price is going to fall.
Actually, it's going to happen as it's going to rise for the high-quality reason
that you're going to be able to do so much more.
Yeah.
Right?
With software, the products are going to be so much better,
and the roadmap is going to be so amazing of the things you can do,
and the customers are going to be so happy with it that they're going to want more and more and more.
So the result of it
And by the way, another example
of Jevon's Paradox playing out
in another related industries in Hollywood
you know, CGI in theory
should have reduced the price of making movies
in reality has increased it
because audience expectations went up.
Yeah.
And now you go to a Hollywood movie
and it's wall-to-wall CGI.
And so, you know, movies are more expensive
to make than ever.
And so the result of it, you know,
but the result in Hollywood
is at least much more, let's say,
visually elaborate, you know, movies,
whether they're better or not
another question,
but like much more visually elaborate,
compelling kind of visually stunning movies
through CGI.
The version here would be
much better software.
Yeah.
Like radically better software to the end user, which causes end users to want a lot more
software, which causes actually the price of development to rise.
You know, if you just think about like a simple case like travel, like, okay, booking a trip
through Expedia is like complicated.
You're likely to get it wrong.
You're clicking on menus and this and that and the other.
And like, you know, da-da-da-da.
An AI version of that would be like, you know, send me to Paris, put me in a hotel I'll love at the best price.
you know, send me on the best possible kind of airline, an airline ticket.
And then, you know, like, make it, like, really special for me.
And, like, maybe you need a human to go, okay, like, we're going to, you know,
or maybe the AI gets more complicated and says, okay, well, we know the person loves chocolate
and we're going to, like, you know, FedEx in the best chocolate in the world from Switzerland
into this hotel in Paris and this and that and the other.
And so, like, the quality, you can, the quality could get to levels that we can't even imagine today, just because, you know, the software tools aren't, aren't what they're going to be.
So, yeah, that's right.
Yeah, I kind of buy that, actually.
I think I brought in the argument.
You're both.
How about, yeah, or how about I'm going to land in whatever, Boston at 6 o'clock?
I want to have dinner at 7 with a table full of, like, super interesting people.
Yeah, right, right, right, right.
You know.
Yeah.
Right.
Yeah, yeah, yeah.
No travel agent would do that for you today, nor would you want them to.
Yeah, no.
No.
Right, well, and then you think about it, it's got to be integrated into my personal
AI and, like, and this, you know, there's just, like, unlimited kind of ideas that you can do.
And I think this is one of the kind of things that's always been underestimated about humans
is, like, our ability to come up with new things we need.
Like, that has been unlimited.
And there's a very kind of famous case where John Maynard Keynes, who the kind of prominent
economists in the kind of first half of last century, had this thing that he predicted,
which is like nobody, because of automation, nobody would ever work a 40-hour work week.
You know, like because once their needs were met, needs being like shelter and food.
And, you know, I don't even know if transportation was in there.
like that was it it was over and like you would never work past the need for shelter and food like
why would you like there's no reason to but of course needs expanded so then everybody needed a
refrigerator everybody needed not just one car but a car for everybody in the family everybody
needed a television set everybody needed like glorious vacations everybody you know so what are we
going to need next i'm quite sure that i can't imagine it but like somebody's going to imagine it
and it's quickly going to become a need.
Yeah, that's right.
By the way, as Keynes famously said,
his essay, I think, was economic prospects for our grandchildren,
which was basically that.
Yeah.
You just articulated.
So Carl Marx had another version of that.
I just pulled up the quote.
So that society, when, you know,
when the Marxist utopia, socialism is achieved,
society regulates the general production,
thus makes it possible for me to do blah, blah,
to hunt in the morning,
fish in the afternoon,
rear cattle in the evening,
criticize after dinner.
What a glorious life.
What a glorious life.
Like, if I could just list four things
that I do not want to do,
it's hunt, fish, rear cattle, and criticize.
Yeah, yeah.
Right?
And by the way, it says a lot about Marx
that those were his four things.
Well, criticizing being his favorite thing,
I think, is basically communism in a nutshell.
Yeah, exactly.
I don't want to get too political, but yes.
Yes, 100%.
And so, yeah, so it's this, this,
Yeah, what they have, what Keynes and Marks had in common is just this incredibly
constricted, it's incredibly constricted view of what people want to do.
And then, and then correspondingly, you know, the other thing is just like, you know,
people, people, people who want to have a mission.
I mean, probably some people just want to fish and hunt.
Yeah.
But, you know, a lot of people want to have a mission.
They want to have a cause.
They want to have a purpose.
It's actually a good thing in life, it turns out, you know.
It turns, it turns out.
Yeah.
In the startling turn of events.
Okay.
So, yeah, so, yeah, I think that I've long felt, you know, a little bit of the soft
rates the world thing, a decade ago.
I've always thought that basically demand for software is sort of perfectly elastic, possibly to infinity.
And the theory there basically is if you just continuously bring down the cost of software,
you know, which has been happening over time, then basically demand, you know, basically is like basically perfectly correlates upward.
And the reason is because, you know, kind of, as we've been discussing, but it's kind of there's, there's always something else to do in software.
There's always something else to automate. There's always something else to improve.
There's always something to make better. And, you know, in the moment with the constraints that you have today, you may not, you know, think about,
that is, but the minute you don't have those constraints, you'll imagine what it is.
Well, I'll just give you an example. I mean, so I'll give you an example that's playing out
with AI right now, right? So there have been, and we have, you know, we have companies that do
this, you know, there have been companies that have made AI, you know, that have made software
systems for doing security cameras forever, right? And it's like, and for a long time, it was
like a big deal to have software that would do, like, you know, have different security
camera feeds and store them on a DVR and be able to replay them and have an interface that lets
you do that. Well, it's like, you know, AI security cameras all of a sudden can have,
like, actual, semantic knowledge of what's happening in the environment. And so they
can say, you know, hey, that's Ben.
And then they can say, oh, hey, you know, that's Ben, but he's carrying a gun.
Yeah.
Right.
Right.
And by the way, that's Ben and he's carrying a gun, but that's because, like, he hunts on,
you know, on Thursdays and Fridays as compared to that's Mary.
And she never carries a gun and, like, you know, like something is wrong.
And she's really mad, right?
She's got a, yeah, really esteemed expression in her face and we should probably
be worried about it, right?
And so there's, like, an entirely new set of capabilities you can do, just as one example,
for security systems that were never possible pre-AI.
And a security system that actually has a semantic understanding.
of the world is obviously much more sophisticated than the one that doesn't and might actually be
more expensive to make, right?
Right.
Well, and just imagine health care, right?
Like, you could wake up every morning and have a complete diagnostic, you know, like, how am I doing today?
Like, what are all my levels of everything?
And, you know, how should I interpret them, you know, better than, you know, this is one thing
where AI is really good is, you know, medical diagnosis because it's a super high.
dimensional problem, but if you can get access to, you know, your continuous glucose reading,
you know, maybe sequence your blood now and again, this and that and the other, yeah, you've got
an incredible kind of view of things and who doesn't want to be healthier, you know, like now we have
a scale. That's basically what we do. You know, maybe check your heart rate or something,
but like pretty primitive stuff compared to where we could go. Yeah, that's right. Okay, good. All right,
So let's go to the next topic.
So on the topic of data.
So Major Tom asks, as these AI models allow for us to copy existing app functionality at minimal cost,
proprietary data seems to be the most important moat.
How do you think that will affect proprietary data value?
What other moats do you think companies can focus on building in this new environment?
And then Jeff Weisshopped asks, how should companies protect sensitive data, trade secrets, proprietary data, individual privacy in the brave new world of AI?
So let me start with a provocative statement.
Ben, see if you agree with it, which is, you know, you sort of hear a lot, this sort of statement or cliche is like data is the new oil.
And so it's like, okay, you know, data is the key input to training AI, making all this stuff work.
And so, you know, therefore, you know, data is basically the new resource.
It's the limiting resource.
It's the super valuable thing.
And so, you know, whoever has the best data is going to win and you see that directly in how you train AI's.
And then, you know, you also have like a lot of companies, of course, that are now trying to figure out what to do with AI.
And a very common thing you'll hear from companies is, well, we have proprietary data, right?
So I'm a hospital chain or I'm a, you know, whatever, any kind of business, insurance company or whatever.
And I've got all this proprietary data that I can apply, you know, that I'll be able to, you know, build things with my proprietary data with AI that won't just, you know, be something that anybody will be able to have.
Let me argue that basically, let's see, I mean arguing in like almost every case like that, it's not true.
It's basically what the Internet kids would call cope.
It's simply not true.
And the reason it's just not true is because the amount of data available on the Internet and just generally in the environment.
and just generally in the environment is just a million times greater.
And so while it may not, you know,
while it may not be true that I have your specific medical information,
I have so much medical information off the Internet for so many people
and so many different scenarios that it just swamps the value of, quote, your data.
You know, it's just like overwhelming.
And so your proprietary data as, you know, company X will be a little bit useful in the margin,
but it's not actually going to move the needle.
And it's not really going to be a barrier to entry in most cases.
and then let me cite as proof for my belief that this is mostly cope is
there has never been nor is there now any sort of basically any level of rich or sophisticated
marketplace for data market for data there's no there's no large marketplace for data and
in fact in fact what there are is there are very small markets for data so there are these
businesses called data brokers that will sell you you know large numbers of like you know
information about users on the internet or something and they're just small businesses like
they're just not large it just turns out like information on lots of people is just
not very valuable. And so if the data actually had value, you know, it would have a market price
and you would see a transacting and you actually very specifically don't see that, which is sort of
a, yeah, sort of quantitative proof that the data actually is not nearly as valuable as people
think it is. Where I agree, so I agree that the data like just says, here's a bunch of data
and I can sell it without doing anything to the data is like massively over.
I definitely agree with that.
And like maybe I can imagine some exceptions like some, you know, special population genomic databases
or something that are that were very hard to acquire that are useful in some way.
That's, you know, that's not just like living on the Internet or something like that, I could imagine,
where that's super highly structured, very general purpose and not widely available.
But for most data in companies, it's not like that.
and that it tends to not,
it's either widely available or not general purpose.
It's kind of specific.
Having said that, right,
like companies have made great use of data,
for example, a company that you're familiar with,
meta, uses its data to kind of great ends itself,
feeding it into its own AI systems,
optimizing its products in incredible ways.
And I think that, you know, us, Andres and Horowitz,
actually, you know, so we just raised $7.2 billion.
And it's not a huge deal, but we took our data and we put it into an AI system.
And our LPs were able, there's a million questions investors have about everything we've done, our track record, every company we've invested and so forth.
And for any of those questions, they could just ask the AI.
They could be wake up at 3 o'clock in the morning and go, do I really want to trust these guys and go in and ask the AI a question?
And boom, they'd get an answer back instantly.
They'd have to wait for us and so forth.
So we really kind of improved our investor relations product tremendously through use of our data.
And I think that almost every company can improve its competitiveness through use of its own data.
But the idea that it's collected some data that it can go like sell or that is oil or what have you,
that's, yeah, that's probably not true.
I would say.
And, you know, it's kind of interesting
because a lot of the data
that you would think
would be the most valuable
would be like your own code base.
Right, your software that you've written.
So much of that lives in GitHub.
Nobody is actually,
I don't know of any company.
We could work with, you know,
whatever, a thousand software companies.
And do we know any that's like
building their own programming model
on their own code?
Like, or, and would that be a good idea?
Probably not just because there's so
much code out there that the systems have been trained on. So like that's not so much of
advantage. So I think it's a very specific kind of data that would have value. Well, let's make it
actionable then. If I'm running a big company, like if I'm running an insurance company or a
banker or a hospital chain or something like that, or, you know, a consumer package goods
company, Pepsi or something, like what, how should I validate? Like, how should I validate that I actually
have a valuable proprietary data asset that I should really be focusing on using.
versus in the alternate, by the way,
maybe there's other things.
Maybe I should be taking all the effort I would spend
on trying to optimize use of that data,
and maybe I should use it entirely trying to build things
using internet data instead.
Yeah, so I think, I mean, look, if you're, right,
if you're in the insurance business, then, like,
all your actuarial data is both interesting,
and then I don't know that anybody publishes their actual,
or actuarial data.
And so, like, I'm not sure how you would train the model
on stuff off of the internet.
You know, similarly...
Can I challenge that one?
So that would be a good thing.
That would be a good test case.
So I'm an insurance company.
I've got records on 10 million people
and, you know, the actuarial tables
and when they get sick and when they die.
Okay, that's great.
But like, there's lots and lots of actuarial,
general actuarial data on the internet
for large-scale populations, you know,
because governments collect the data
and they process it and they publish reports.
And there's lots of academic studies.
And so, like, is your large data set,
giving you any additional actuarial information
that the much larger data set on the internet
isn't already providing you?
Like are your insurance clients
actually actuarially any different
than just everybody?
I think so because on intake on the
you know, when you get insurance,
they give you like a blood test,
they've got all these things.
I know if you're a smoker and so forth.
And in the, I think in the general data set,
like, yeah, you know who dies,
but you don't know what the fuck they did coming in.
And so what you realize,
are looking for is like, okay, for this profile of person with these kinds of lab results,
how long are they live? And that's where the value is. And I think that, you know, interesting,
like, you know, I was thinking about like a company like Coinbase where, right, they have
incredibly valuable assets in the terms of money. They have to stop people from breaking in.
They've done a massive amount of work on that. They've seen all kinds of break in types. I'm sure they
have tons of data on that.
It's probably, like, weirdly specific to people trying to break into crypto exchanges.
And so, you know, like, I think it could be very useful for them.
I don't think they could sell it to anybody.
But, you know, I think every company's got data that if, you know, fed into an intelligent
system would help their business.
And I think almost nobody has data that they could just go sell.
And then there's this kind of in-between question.
which is what data would you want to let Microsoft or Google or OpenAI
or anybody get their grubby little fingers on?
And that I'm not sure.
That I think is the question that enterprises are wrestling with more than,
it's not so much should we go like sell our data,
but it should we train their own model just so we can maximize the value
or should we feed it into the big model?
And if we feed it into the big model, do all of our competitors now have the thing that we just did?
And, you know, or could we trust the big company to not do that to us?
Which I kind of think the answer on trusting the big company not to F with your data is probably I won't do that.
Well, I mean, yes.
If your competitiveness depends on that, you probably shouldn't do that.
Well, there are at least reports that certain big companies are using all kinds of data that they shouldn't be using to train their models already.
so yep i think like i think those reports are very likely true right or they have open data right
like this is you know we've talked about this before but you know the same companies that are saying
they're not stealing all the data from people you are taking it an unauthorized way refuse to
say open their data like why not tell us where your data came from and in fact they're trying
to shut down all openness no open source no open weights no open data no open nothing
and go to the government and try and get to do that.
You know, if you're not a thief, then why are you doing that?
Right, right, right.
What are you heading?
By the way, there's other twist and turns here.
So, for example, in the insurance example, I kind of deliberately loaded it because you may
know it's actually illegal to use genetic data for insurance purposes, right?
So there's this thing called the Geno Law, Genetic Information Non-Discrimination Act of 2008.
And basically it basically banshears in the U.S.
from actually using genetic data for the purpose of doing, you know, health assessment,
actual actual assessment of it, which, by the way, because now the genomics are getting really
good, like that data probably actually is, you know, among the most accurate data you could have
if you were actually trying to predict, like, when people are going to get sick and die,
and they're literally, they're literally not allowed to use it.
Yeah, it is, I think that this is an interesting, like, weird misapplication of good intentions
in a policy way that's probably going to kill more people than ever get saved by
every kind of health, FDA, et cetera, policy that we have, which is, you know, in a world of
AI, having access to data on all humans, why they get sick, what their genetics were,
et cetera, et cetera, et cetera, is the most, that is, you know, you don't talk about data being
the new oil, like that is the new oil.
the health care oil is, you know, if you could match those up, then we'd never not know why
we're sick. You know, you could make everybody much healthier, all these kinds of things.
But, you know, to kind of stop the insurance company from kind of overcharging people who are
more likely to die, we've kind of locked up all this data. A kind of better idea would be
to just go, okay, for the people who are likely to, like, we subsidize health care, like,
massively for individuals anyway, just, like, differentially, you know, subsidize.
And, you know, and then, like, you solve the problem and you don't lock up all the data.
But, you know, it's typical of politics and policy.
I mean, most of them are like that, I think, yeah.
Well, there's this interesting questions like an insurance, like, basically, one of the
questions people have asked about insurance is, like, if you had perfectly predictive
information on, like, individual outcomes, does the whole concept of insurance actually still
work, right? Because the whole theory of insurance is risk pooling, right? It's precisely the
fact that you don't know what's going to happen in the specific case that means you build these
statistical models and then you risk pool and then you have variable payouts depending on exactly
what happens. But if you literally knew what was going to happen in every case, because, for
example, you had all of this predictive genomic data, then all of a sudden it wouldn't make
sense to risk pool because you just say, well, no, this person is going to cost X, that person is
going to cost Y. There's no. Health insurance already doesn't make sense in that way, right?
Like in France, the idea of insurance is kind of like the, it started with crop insurance where
like, okay, you know, my crop fails. And so we all put money in a pool in case like my crop fails
so that, you know, we can cover it. It's kind of designed for it to risk pool for a catastrophic
unlikely incident.
Like, everybody's got to go to the doctor all the fucking time.
And some people get sicker than others and that kind of thing.
But, like, the way our health insurance works is, like, all medical gets, you know,
paid for through this insurance systems, which is this layer of loss in bureaucracy
and giant companies and all this stuff when, like, if we're going to pay for people's
health care, just pay for people's health care.
Like, what are we doing?
right? And if you want to disincent people from like going for nonsense reasons than just up the co-pay,
like it's like what are we doing? Just. Well, and then from a justice standpoint, from a fairness
standpoint, like what it makes sense for me, you know, would it make sense for me, you know,
if I knew that you were going to be more expensive than me, like, you know, I'm directly, you know,
if everybody knows what future health care costs is per person. Yeah. There has a very good
predictive model for it. You know, societal willingness to all pool in. The, you know,
way that we do today might really diminish. Yeah, yeah. Well, and then, like, you could also,
if you knew, like, there's things that you do genetically, and maybe we give everybody a pass on
that. It's like you can't control your genetics. But then, like, there's things you do behaviorally
that, like, dramatically increases your chance of getting sick. And so maybe, you know, we incentivize
people to stay healthy instead of just, like, paying for them not to die. There's a lot of systemic
fixes we could do to the healthcare system. It couldn't be designed in a more ridiculous way,
I think. Well, it couldn't be designed a more ridiculous way. It's actually more ridiculous than
some other countries, but it's pretty crazy here. Nathan Odie asks, what are the strongest
common themes between the current state of AI and Web 1.0? And so let me start there. Let me give you
a theory, Ben, and see what you think. So I get this question, you know, because of my role and, you know, Ben,
you with me at Netscape. We get this question a lot because of our role early on with the internet.
And there's an, you know, the internet boom was like a major, major event in technology,
and it's still within a lot of, you know, people's memories.
And so, you know, the sort of, you know, people like to reason from analogy.
So it's like, okay, the AI boom must be like the internet boom, starting an AI company,
must be like starting an internet company.
And so, you know, what is this like?
And we actually got a bunch of questions like that, you know, that are kind of analogy
questions like that.
I actually think, you know, and then, Ben, you know, you and I were there for the internet
boom.
So, you know, we live through that and the bust and the boom and the bust.
So I actually think that the analogy doesn't really work.
for the most, it works in certain ways,
but it doesn't really work for the most part.
And the reason is because the internet,
the internet was a network,
whereas AI is a computer.
Yep.
Okay, yeah.
So people understand what we're saying.
It's more like the PC boom.
Or the PC boom, or even I would say the microprocessor,
like my best analogy is to the microprocessor.
Yeah.
Or even to the original computers,
like back to the mainframe era.
And the reason is because, yeah,
look, what the internet did was the internet,
you know, obviously it was a network,
but the network connected together many existing computers,
and then, of course, people built many other new kinds of computers
to connect to the Internet.
But fundamentally, the Internet was a network.
And that's important because most of the sort of industry dynamics,
competitive dynamics, startup dynamics around the Internet,
had to do with basically building,
either building networks or building applications that run on top of networks.
And this, you know, the Internet generation of startups
was very consumed by network effects.
And, you know, all these positive feedback loops that you get
when you connect a lot of people together.
And, you know, things like met, you know, so-called Metcast law, which is sort of the value of a network, you know, expands, you know, kind of the way it expands is you have more people to it.
And then, you know, there were all these fights, you know, all the social networks or whatever fighting to try to get network effects and try to steal each other's users because of the network effects.
And so it's kind of, you know, it's dominated by network effects, which is what you expect from a network business.
AI, like there are some networks effects in AI that we can talk about, but it's more like a microprocessor.
It's more like a chip.
It's more like a computer.
It's a system that basically, right, if data comes in, data gets processed, data comes out, things happen.
That's a computer.
It's an information processing system.
It's a computer.
It's a new kind of computer.
It's a, you know, we like to say the sort of computers up until now have been what are called von Neumann machines, which is to say they're deterministic computers, which is they're like, you know, hyper literal.
And they do exactly the same thing every time.
And if they make a mistake, it's, yes, the programmer's fault.
But they're very limited in their ability to interact with people and understand the world.
We think of AI and large language models as a new kind of computer, a probabilistic computer, a neural network-based computer that, you know, by the way, is not very accurate and doesn't give you the same result every time and, in fact, might actually argue with you and tell you that it doesn't want to answer your question.
Yeah, yeah, which makes it very different in nature than the old computers, and it makes it kind of compulsibility, you know, the ability to build things, big things out of little things more complex.
right but but the capabilities are new and different and valuable and important because it can
understand language and images and you know that all these do all these things that you see when
you use all that means we never solve with deterministic computers we can now go after right
yeah exactly and so i think i think ben i think the analogy and i think the lessons learned are
much more likely to be drawn from the early days of the computer industry or from the early days
of the microprocessor than the early days of the internet does that does that sound right i think so yeah
I definitely think so.
And that doesn't mean there's no boom and bust and all that because that's just the nature of technology.
You know, people get too excited and then they get too depressed.
So there will be some of that, I'm sure.
There will be over-billed-out, you know, potentially of, eventually of chips and power and that kind of thing.
You know, we start with the shortage.
But I agree.
Like, I think networks are fundamentally different in the nature of how they evolved than computers.
And kind of just the adoption curve and all those kinds of things.
things will be different.
Yeah, so then, and this kind of goes to where, how I think the industry is going to
unfold. And so this is kind of my best theory for kind of what happens from here.
It's kind of this, you know, this giant question of like, you know, is the industry going
to be a few God models or, you know, a very large number of models of different sizes and so
forth? So the computer, like famously, you know, the original computers, like the original IBM
mainframes, you know, the big computers, you know, they were very, very large and expensive
and there were only a few of them. And the prevailing view, actually, for a long time was
that's all there would ever be.
And there was this famous statement by Thomas Watson, Sr.,
who was the creator of IBM,
which was the dominant company
for the first, like, you know, 50 years
of the computer industry.
And he said, he said, I didn't believe this is actually true,
but he said, I don't know,
I don't know that the world will ever need
more than five computers.
And I think the reason for that,
it was literally, it was like the government's going to have two,
and then there's like three big insurance companies,
and then that's it.
Who else would need to do all that math?
Exactly.
Yeah, who else would need to,
who else needs to keep track a few
amounts of numbers, who else needs that level of, you know, calculation capability.
It's just not a relevant, you know, it's just not, not a relevant concept.
And by the way, they were, like, big and expensive.
And so who else can afford them, right?
And who else can afford all the headcount required to manage them and maintain them?
I mean, and this is in the days, I mean, these things were big.
These things were so big that you'd have an entire building that got built around a computer, right?
And they'd have, like, they'd famously have all these guys in white lab coats, literally, like,
taking care of the computer because everything had to be kept super clean or the computer
would stop working.
And so, you know, it was this thing where, you know, there's a, it's a,
day we have the idea of an AI god model which is like a big foundation model that you know then we
have the idea of like a god mainframe like there would just be a few a few of these things and by the way
if you watch old science fiction it almost always has this sort of conceit it's like okay there's a big
supercomputer and it either is like doing the right thing or doing the wrong thing and if it's doing the wrong
thing you know that's that's often the plot of the science fiction movies is you have to go in and try to
figure how to fix it or defeat it so it's sort of this this idea of like a single top down thing of course
and that held for a long time like that held for you know the first few decades and then
And, you know, even when computers, computers started to get smaller.
So then you had so-called mini-computers was the next phase.
And so that was a computer that, you know, didn't cost $50 million.
Instead, it costs, you know, $500,000.
But even still, $500,000 is a lot of money.
People aren't putting mini-computers in their homes.
And so it's like mid-sized companies can buy mini-computers, but certainly people can't.
And then, of course, with the PC, they shrunk down to like $2,500.
And then, you know, with the smartphone, they shrunk down to $500.
And then, you know, sitting here today, obviously, you have computers of every shape,
size description all the way down to, you know, computers that cost,
a penny, you know, you've got a computer in your thermostat that, you know, basically
controls the temperature in the room, and it, you know, probably cost a penny, and it's probably
some embedded arm chip with firmware on it. And there's, you know, many billions of those
all around the world. You buy a new car today. It has something, new cars today have something on the
order of 200 computers in them, maybe more at this point. And so you just basically assume with
the chip today, sitting here today, you just kind of assume that everything has a chip in it.
You assume that everything, by the way, draws electricity or has a battery because it needs
to power the chip. And then increasingly you assume that everything's on the internet, because
basically all computers are assumed to be on the internet or they will be.
And so, as a constant, what you have is the computer industry today is this massive pyramid.
And you still have a small number of like these supercomputer clusters or these giant mainframes
that are like the God model, you know, the God mainframes.
And then you've got, you know, a larger number of many computers.
You've got a larger number of PCs.
You've got a much larger number of smartphones.
And then you've got a giant number of embedded systems.
And it turns out like the computer industry is all of those things.
And, you know, what is it?
What size of computer do you want is based on, well, what exactly are you trying to do and who are you and what do you need?
And so if that analogy holds, it basically means actually we are going to have AI models of every conceivable shape, size, description capability, right, based on trained on lots of different kinds of data running at very different kinds of scale, very different privacy, different policies, different security policies.
You know, you're just going to have like enormous variability and variety, and it's going to be an entire ecosystem and not just a couple of companies.
Yeah, let me see what you think of that.
Well, I think that's right.
And I also think that the other thing that's interesting about this era of computing,
if you look at prayers of computing from the mainframe to the smartphone,
a huge source of lock-in was basically the difficulty of using them.
So, you know, nobody ever got fired for buying IBM because, like, you know,
you had people trained on them.
You know, people knew how to use the operating system.
like it was, you know, it was just kind of like a safe choice due to the massive complexity of, like, dealing with a computer.
And then even with the smartphone, like the, you know, why is the Apple computer smartphone so dominant, you know, what makes it so powerful it's well?
Because, like, switching off of it is so expensive and complicated and so forth.
It's an interesting question with AI because AI is the easiest computer to use by far.
it speaks English. It's like talking to a person. And so, like, what is the lock-in there? And so
are you completely free to use the size, price, choice, speed that you need for your particular
task? Or are you locked into the God model? And, you know, I think it's still a bit of an open
question, but it's pretty interesting. And that, that thing could be very different than prior
generations.
Yeah.
Yeah, that makes sense.
And then just to complete the question, what would we say?
So, Ben, what would you say our lessons learned from the Internet era that we live through
that would apply that people should think about?
I think a big one is probably just the boom-bust nature of it that, like, you know, the demand,
the interest in the Internet, the recognition of what it could be was so high that money just
kind of poured in in buckets and you know and then the underlying thing which in internet age was
the telecom infrastructure and fiber and so forth got just unlimited funding and unlimited fiber
was built out and then eventually we had a fiber glut and all the telecom companies went bankrupt
and and that was great fun but you know like we entered in a good place and i think that that's
something like that's probably pretty likely to happen in AI where like you know every company is going to
get funded. We don't need that many AI companies, so a lot of them are going to bust. There's
going to be a huge, you know, huge investor losses. There will be an overbuilt out of chips for sure
at some point. And then, you know, we're going to have too many chips and, you know, some chip
companies will go bankrupt for sure. And then, you know, and I think probably the same thing with
data centers and so forth, like, we'll be behind, behind, behind, and then we'll overbuild at some
point. So that all be very interesting. I think that, and that's kind of the, that's every new
technology. So Carlotta Perez has a great kind of, has done, you know, amazing work on this where
like that is just the nature of a new technology is that you overbuild, you underbuild it, then you
overbuild, and you know, and there's a hype cycle that funds the build out and a lot of money
is lost, but we get the infrastructure and that's awesome because that's when it really gets adopted
didn't change this world. I want to say, you know, with the internet, the other, the other kind
of big kind of thing is the internet went through a couple of phases, right? Like, it went through
a very open phase, which was unbelievably great. It was probably one of the greatest booms to the
economy. It, you know, it certainly created tremendous growth in power in America, both, you know,
kind of economic power and soft cultural power and these kinds of things. And then, you know,
it became closed with the next generation architecture with, you know, kind of discovery on the
internet being owned entirely by Google and, you know, kind of other things, you know, being
owned by other companies. And, you know, AI, I think, could go either way, so it could be very open
or like, you know, with kind of misguided regulation, you know, we could actually force our
way from something that, you know, is open source, open weights, anybody can build it. We'll have a plethora
of this technology will be like use all of American innovation to compete, or will, you know,
we'll cut it all off, we'll force it into the hands of the companies that kind of on the
internet today. And, you know, and we'll put ourselves at a huge disadvantage, I think,
competitively against China in particular, but everybody in the world. So I think that's something
I think that definitely, you know, that we're involved with trying to make sure it doesn't happen,
but it's a real possibility right now.
Yeah.
There's sort of an irony is that networks used to be all proprietary, and then they opened up.
Yeah, yeah, yeah, right.
Landman, Apple Talk, NetBooey, NetBioes.
Yeah, exactly.
And so these are all the early proprietary networks from all individual-specific vendors,
and then the Internet appeared and kind of TCPIP and everything opened up.
The AI is trying to go the other.
I mean, the big company is trying to take AI the other way.
It started out as, like, open, just like, basically.
just like the research.
Everything was open source and AI, yeah.
Right, right.
And now they're trying to,
they're trying to lock it down.
So it's a,
it's a fairly nefarious turn of events.
Yeah, yeah, very nefarious.
You know, like,
and it's remarkable to me.
I mean, it is kind of
the darkest side of capitalism
when a company is so greedy,
they're willing to destroy the country
and maybe the world to like just get a little extra profit.
But, you know, and they do it.
Like the, the really kind of nasty thing
is they claim, oh, it's for safety.
You know, we've created an alien that we can't control.
But we're not going to stop work on it.
We're going to keep building it as fast as we can,
and we're going to buy every freaking GPU on the planet.
But we need the government to come in and stop it from being open.
This is literally the current position of Google and Microsoft right now.
It's crazy.
And we're not going to secure it.
So we're going to make sure that, like, Chinese spies can just, like,
steal our chip plans, take them out of the country,
and we won't even realize for six months.
Yeah, yeah, it has no.
nothing to do with security. It only has to do with monopoly.
Yes. The other, you know, just I've been going back on your point of speculation.
So there's this critique that we hear a lot, right, which is like, okay, you idiots, basically
it's like you idiots, you idiots, you idiots, entrepreneurs, investors. You idiots, it's just like
there's a speculative bubble with every new technology. Like, basically, like, when are you people
going to learn to not do that? Yeah. And there's an old joke. There's an old joke that relates
to this, which is the foremost dangerous words in investing are, this time is different.
The 12 most dangerous words in investing are the foremost dangerous words are in
in investing are this time is different, right?
Like, so like, does history repeat?
Does it not repeat?
The, my sense of it, and you referenced Carlotta Perez's book, which I agree is good,
although I don't think it works as well anymore, we can talk about sometime, but, but, you know,
is a good, at least background piece on this.
You know, it's just like, it's just incontrovertibly true.
Basically, every significant technology advance in history was greeted by some kind
of financial bubble, basically since financial markets had existed.
And this, you know, by the way, this includes like everything from, you know, radio and
television, the railroads, you know, lots and lots of prior.
By the way, there was actually a so-called,
there was an electronics boom bust in the 60s called the,
it was called the Tronics.
Every company had the name Tronix.
And so, you know, there was that.
So, you know, there was like a laser boom bus cycle.
There were all these like boom bus cycles.
And so basically it's like any new technology,
that's what economists call it general purpose technology,
which is to say something that can be used in lots of different ways.
Like it inspires sort of a speculative mania.
And, you know, and look, the critique is like,
okay, why do you need to have the speculative mania?
Why do you need to have the cycle?
because, like, you know, people,
some people invest in the things, they lose a lot of money.
And then there's this bus cycle that, you know,
causes everybody to get depressed, maybe it delays the rollout.
And it's like two things.
Number one is like, well, you just don't know.
Like, if it's a general purpose technology like AI is
and it's potentially useful in many ways, like nobody actually knows up front,
like what the successful use cases are going to be
or what successful companies are going to be.
Like, you actually have to, you have to learn by doing.
You're going to have some misses.
That's venture capital.
Yeah.
Yeah, exactly.
Yeah, exactly.
So, yeah, the true venture capital model,
kind of wires this in, right?
We basically, in core venture capital, the kind that we do, we sort of assume that
half the companies fail, half the projects fail.
And, you know, if any of us, if we or any of our...
Failed completely, like, lose money.
Yeah.
Lose money, exactly, yeah.
And so, like, and of course, if we or any of our competitors, you know, could figure out
how to do the 50% that work without doing the 50% that don't work, we would do that.
But, you know, here we sit 60 years into the field and, like, nobody's figured that out.
So there is, there is that unpredictability to it.
And then the other kind of interesting way to think about,
this is like, okay, what would it mean I have a society in which a new technology did not inspire
speculation? And it would mean having a society that basically is just like inherently like super
pessimistic about both the prospects of the new technology, but also the prospects of entrepreneurship
and, you know, people inventing new things and doing new things. And of course, there are many
societies like that on planet Earth, you know, that just like fundamentally like don't have
the spirit of invention and adventure that, you know, that a place like Silicon Valley does.
And, you know, are they better off or worse off? And, you know, generally speaking, they're
worse off. They're just, you know, less future oriented, less, less, less focused on building
things, less focused on figuring out how to get growth. And so I think there's a, at least my sense,
there's a comes with the territory thing. Like we, we would all prefer to avoid the downside of a
speculative boom-bus cycle, but like it seems to come with the territory every single time.
And I, at least I have not, nobody, I'm aware, no society I'm aware of has ever figured out
how to capture the good without also having the bet. Yeah. And like, why would you? I mean,
it's kind of like, you know, the whole western United States was built off the gold rush. And like
every kind of treatment in like popular culture of the gold rush kind of focuses on the people
who didn't make any money. But there were people who made a lot of money, you know, and found gold.
And, you know, in the internet bubble, which, you know, was completely ridiculed by, you know,
kind of every, every movie, if you go back and watch any movie between like 2001 and 2004,
they're all like how only morons did dot com and this and that and the other.
And there were all these funny documentaries and so forth.
But like, that's when Amazon got started.
You know, that's when eBay got started.
That's when Google got started.
You know, these companies, you know, were started in the bubble in the kind of time
this great speculation, there was gold in those companies.
And if you hit any one of those, like you funded, you know, probably the next set of
companies, you know, which included things like, you know, Facebook and X and, you know,
Snap and all these things.
And so, yeah, I mean, like, that's just the nature of it.
I mean, like, that's what makes it exciting.
And, you know, it's just a, it's an amazing kind of thing that, you know, look,
the transfer of money from people who have excess.
money to people who are trying to do new things
and make the world a better place is the greatest thing
in the world.
And if some of the people with excess money
lose some of that excess money and trying
to make the world a better place, why are you mad about that?
Like that's the thing that I could never have seen.
Like, why would you be mad at young, ambitious people
trying to improve the world, getting funded,
and some of that being misguided?
Like, why is that bad?
Right, right.
As compared to, yeah, as it compared to, especially as compared to everything else in the world and all the people who are not trying to.
So you'd rather like, you know, lots of mansions and boats and jets.
Right.
Right.
Like, what do you talk?
Right.
Right.
Exactly.
Donate money to ruin us.
Yeah, ruinous causes.
Right.
Such as ones that are on the news right now.
Okay.
So, all right.
We were at a minute 20.
We made it all the way through four questions.
We're doing good.
great. So let's call it here. Thank you, everybody, for joining us. And I believe we should do
a part two of this, if not parts three through six, because we have a lot more questions to go. But
thanks everybody for joining us today. All right. Thank you.