No Priors: Artificial Intelligence | Technology | Startups - Big tech earnings and the current AI debates, with Sarah Guo and Elad Gil
Episode Date: March 7, 2024Host-only episode discussing NVIDIA, Meta and Google earnings, Gemini and Mistral model launches, the open-vs-closed source debate, domain specific foundation models, if we’ll see real competition i...n chips, and the state of AI ROI and adoption. Don’t miss our episodes with: Mistral NVIDIA AMD Sign up for new podcasts every week. Email feedback to show@no-priors.com Follow us on Twitter: @NoPriorsPod | @Saranormous | @EladGil Show Notes: (0:00) Introduction (0:27) Model news and product launches (5:01) Google enters the competitive space with Gemini 1.5 (8:23) Biology and robotics using LLMs (10:22) Agent-centric companies (14:22) NVIDIA earnings (17:29) ROI in AI (20:43) Impact from AI (25:45) Building effective AI tools in house (29:09) What would it take to compete with NVIDIA (33:23) The architectural approach to compute (35:42) the roadblocks to chip production in the US (38:30) The virtuous tech cycles in AI
Transcript
Discussion (0)
Today I know Pryors, we're having a special episode of Sarah and me just talking.
Hello, Sarah. How are you?
Hey, a lot. What's going on? I see you a lot.
Not much. Good to see it.
Let's talk about models. What's going on in the model world?
Yeah. I guess there's a lot of hand models that are emerging.
So I was thinking of maybe trying to do that eventually.
It's almost as good of a business as it's investing.
I know, right? Yeah, so it's been a lot that's happened in the model world recently. Obviously, Google launched Gemini, which I think had a few interesting characteristics, both in terms of performance, but also the huge context window, right? It was a million token context window. Companies like Magic, I think in the past have actually put out like a $5 million token context window model and things like that, but it's really exciting to see that. And I think for certain application areas, like biology, longer context windows actually seem to be quite important. And so, for example, if you're doing your protein folding model and you have a short context window,
you're often actually not encapsulating much of the protein, right?
The average protein is, I think, something like 300 amino acids long, at least in the human genome,
but there are things that are dramatically larger than that.
And so you just can't capture it in some of the context windows being used for biological models.
And so I do think this is going to be one of those areas that will end up being more important
that people think, at least in the short run.
But Gemini, at 1.5, seems to have some really interesting performance characteristics.
There's obviously SORA from OpenAI, which was the video model that, you know,
It was just beautiful to watch.
There's other model companies like Pica and others that I think are doing exciting things as well.
And then Mestral or Le Miz launched Le Chatt, which is really the name of the product.
Le Big Model.
Le Big Mac.
I believe they call it Mistral Large.
Yes.
Yes.
Le Large, Mistral Large, they launched that.
And the thing that's really, really impressed me about Mistral is just the velocity of shipping.
It's incredibly impressive.
They went from basically starting the company to almost GPT4 level in less than a year.
Nine months, yep.
It's amazing.
And they have, you know, small performant models.
They have the big Mac or, you know, the large model.
They have chat.
They have multiple languages.
It's just, it's very impressive execution.
So, and then I think the other thing that they just launched or announced was that deal with
Microsoft where, you know, they're now being licensed on to Azure.
And so I think the main models in Azure now are Open AI, Lama, Mistral, and then some of the Microsoft model.
So again, that's striking as well.
So just very impressive progress by that company so far.
I think the design space for what you actually want from models is certainly going to include state-of-the-art capability.
And Mistral is very much going after that, and they've said so.
But I think, like, from the beginning, the company has talked about efficiency and latency.
and the ability to serve different use cases with that
and also, you know, being long-term proponents of retrieval, right?
Like one of the big debates in the research world right now,
I don't know how much of it is a debate,
but people are talking about it.
I'm on one side of this, is that, like, rag and retrieval is dead
with sufficient context.
And I'm curious what you think here.
But I'm more of the belief that it just opens up the set of tradeoffs you can make
between retrieval, more sophisticated retrieval and model reasoning by having a larger context
window versus saying, like, we don't need any ability to work with a specific data set
versus just retrain or stuff something into context.
Yeah, we're going to have both, in my opinion.
The other thing I think that's very under-discussed, and this could lead into agent stuff,
but I'd like to also spend a little bit time on Gemini before we move to agents, is if you look
at a lot of the optimations that are done for areas where you had,
human-related sort of reasoning or other components pre-LLM-based reasoning.
A lot of it was happening at inference time, right?
So when you were doing, when you were trying to build a better poker AI,
a lot of what you did was, you know, certain types of researches or other things
when you hit inference time, right?
You built the model, but at inference it did a lot of extra work.
And I think that's also a little bit under discussed in terms of probably a lot of what's
going to happen in the future, particularly we get into agents and reasoning.
is stuff that's happening at that point of inference,
and then it's used to sort of feedback over
and sort of continuously train or retrain a model over time,
because I think that's the other piece of it is, you know,
from a model perspective,
you spin up a giant data center,
and you spend $100 million over 12 months
overall between all the different works that you do
and everything to launch your next model,
and then you have a file, and then you use that files for the next year
as you train the next model versus saying you're going to do
some sort of continuous upgrading or training.
And so all these things are going to shift over time.
I think it's early in the technology cycle.
And so all these things are going to happen.
You know, one of the companies that has a lot of capabilities
to do interesting things over time, of course, is Google.
So I'm a little bit curious if Gemini has changed your opinion of sort of the AI model race
and what role Google plays in the future, you know, has it not changed your mind much?
I think the question on like whether or not Google has the ability to
do the research work to have a competitive product has been answered, right? Gemini is a very
impressive model. I think the, the capabilities that they have internally that they haven't
released yet around additional like function calling and multimodality are also really, really
impressive. And so the questions around Google are less about do they have, like they have all
of these extraordinary advantages. And you're the ex-googler. Like, I want to hear your opinion. But
they have the distribution, they have the consumer behavior, they have all the data on like
what the search behavior is, they have the data on what queries are valuable and which they
would peel away and turn into like an answer. They know how to build like advertising auction
systems and they have a great research team and enough GPUs and the model capabilities.
Do you think it's progressive enough though?
Do I think the models are progressive enough?
Yeah.
One might actually ask if they're perhaps a little too far in that direction.
Right.
And so I think, like, the question is actually, can they steer Google to, like, focus on being competitive versus the many other demands from their employee base and, like, different missions that are not, you know, brokering the world's information and, like, market cap.
Yeah.
Yeah. It's interesting because the launch of 1.5 has made me more bullish on Google. And I was always actually quite positive on it, right? Like I think I read a blog post a year or a year and a half ago, basically about the model world. And one of the things that I mentioned at the time was I felt like Google was kind of a sleeping giant. Once it awoke, you know, it could really make enormous progress quickly. And just as Mistral is executed from scratch as a startup, which is extremely hard to do, right? You're literally building everything from the ground up. Although obviously there's open source to support you.
and all these other things, but fundamentally,
you're just building an entire company.
It's pretty amazing, right?
Google has really accelerated its efforts,
and it's had a series of launches over the last two, three months
that have been quite impressive in terms of the velocity
from Cold Start to having things that are externally accessible.
And they have all the resources that one would need in order to do extremely well in AI, right?
They have the compute, they have unique proprietary data
as well as all the data from the web,
all the data from YouTube.
They have specialized data that you could potentially opt into,
like, you know, all your emails and your Google Docs.
And, you know, they have this immense corpus of really valuable information.
And then they have amazing talent.
And so really, I think the thing that was lacking until recently was the Will.
And it seems like now, because of the competitive dynamic,
the Will has been reborn, right?
And so it really feels to me like they are going to make really big strides.
going forward, and, you know, it's always possible
with the velocity only increases from here for them.
If I think about the domains in which these general LLMs
are still not as capable, I mean, it's every domain,
but in particular not as capable as we want,
like two of the areas, one you already mentioned,
that I'm excited about include, like,
biology and then robotics, right? So maybe let's talk about that for a second. As a task,
for example, if you ask ChachyPT to design a DNA sequence that can express Prisper Cas9,
it can't do that yet. Right. And if we think about cell design, protein design, protein
optimization, a lot of these are areas where you have researchers showing like really exciting
progress in use of transformers and diffusion models to get to, um, get to,
much better predictions for, for example, drug discovery and target identification.
And so I think, you know, I've seen a number of companies in this area of better understanding of
biology that really feels like a different type of reasoning, a different type of data set.
And as you said, even like specific context window constraints.
And so I think that's an interesting one.
And then on the, I don't know if you wanted to mention the robotic side or if that's something
you've been looking at too.
The robotic stuff seems super interesting.
It's a little bit earlier on than some of the other models and part due to data constraints,
but it seems like there's pretty reasonable ways to generate some of that data now.
So it seems like in general, I wouldn't be surprised if 2024 and 2025 is the year
of proliferation of models, where we're going to start to see an expansion in terms of the
different types that are covered, chemistry and material sciences, et cetera, et cetera.
Robotics will be part of that, biology would be part of that.
maybe physics and math.
I think maybe the last thing
that is happening
from a model perspective
is I think the last few weeks
have seen a lot of different
sort of agent-centric companies
get up and running.
And I think that's been
a really interesting wave.
And some of them, again,
are taking very different approaches
from the traditional,
let's just build a giant LLM.
And they're looking at things
like AlphaGo or some of the game-centric
work that have been done in the past,
you know, how do you build a better poker paper?
How do you build diplomacy?
How do you build Go?
And there you have a very strong notion
of acting sequentially based on changing information.
You have some forms of what's known as self-play.
You know, you have the machine play itself
a billion times ago, and it learns new patterns based on that.
You have really interesting approaches
in heuristics and algorithms at time of inference versus training.
And so I think that that harpus of knowledge is about to hit the world in the context of new products.
And it'll take time for those products to emerge, you know, six months, 12 months a year.
But it does feel like that's another wave is coming where you're taking a fundamentally different approach that involves reinforcement learning,
but is just different in terms of how you think about what you're actually doing in architecting and what you're inferencing and all the rest.
So that's the one other area on the model side that I think is very.
exciting. One thing that I've seen here is that people are getting much smarter about agents as
part of systems versus expecting to simply, like, instruct an agent and have it work with
compounding failure across a bunch of tasks, right, in a general environment across any type of
software, right? And so if it is operating in an environment that supports reinforcement well,
like a game environment, or even a web application environment, but one that is constrained
to particular tasks or agents working in domains that better support a sampling and validation,
like code generation.
Like, I'm really excited about that.
And I feel like I've begun to see the glimpse of some of those things work, whereas a very
real question you could have asked in Q3, Q4 of last year would be like, is any of this stuff
useful, right? Is it anything? And I think now it's like, it is. Yeah. Yeah, people went too broad
to early versus just saying, I'm just going to focus on a handful of targeted use cases or domains.
And I'm going to figure out how do you create feedback loops in those domains so I can actually
train effectively. And so, you know, the very early versions of this, even predating this LLM wave
was, you know, hey, we're going to have a browser plug in and it'll watch everything you do
and then it'll do everything you do, which is a very different problem from saying, hey, we're going
to make our PA better, we're going to make good better, we're going to make customer support
better, we're going to make, you know, X, Y, Z thing better. So I think the targeted approach makes
a lot of sense. Yeah, and I think some of the teams working on this have also, they've actually
experimented with post-training in environments where you can pay for human feedback data, right?
And if you do that, then you actually understand, like, the distribution.
of data you need, the scale of data you might pay for.
And that's very exciting because it turns like the agent problem from one that is
like open-ended untenable to just like how much is it going to cost to make a particular task
work.
And I'm massively oversimplifying here, but that is a very different proposition when scoped
than like as you described the initial set of forays into agents, which is like, you know,
we'll try to do anything.
Yeah, that makes sense.
And I think we'll still get there.
but there is, like, rapid success on this front.
Invidia, everybody's talking about earnings.
What do you make of it?
I think earning money is an excellent idea.
How about you?
I think Jensen understands this better than everybody else.
I think one thing that people have been talking about is whether or not this was a, like, short-term phenomenon, right?
Like if there was only so much demand, and once the supply chain caught up a little bit, there would be less insane growth.
And I think now people are pretty confident, especially hearing Jensen's comment that they expect to continue to be supply constrained through the rest of the year, demand is just like much, much larger than I think most people expect on the Kappex side.
And I think it's like worth understanding the upgrade cycle.
drives that, right? Because there's this huge efficiency incentive to upgrade from A100s to
H-100s to H-200s to B-100s. I was talking to one of my portfolio companies that's buying in the
tens of thousands of GPU size and is skipping to B-100s because they described it as like
free money in terms of training efficiency. It's funny when somebody describes spending hundreds
of millions of dollars is free money, but free money in terms of training efficiency, if you can actually
get access to a cluster of a certain size.
And so if others feel that way,
it is wild how much this expense expands the server market.
Yeah, it's probably a good time to run a hedge fund.
I think in general, one thing that's a little bit under discussed
is a lot of the emphasis on startups and startup rounds
and look, the startup raised $100 million or whatever.
And the reality is a lot of the spend is the big hyperscalers
and then other clouds that are building out right now.
And then I think the other thing,
is that if you were to look at
at least enterprise adoption of AI,
it's still really, really, really early days.
And despite that, if you look at Microsoft Azure revenue
in the last quarter, they mentioned that
revenue grew by 5% from AI-related products,
which if I'm doing the math right,
if it's a $25 billion, a quarter,
Azure sort of revenue,
then that means they're adding something like one,
one and a half billion a quarter
in new spend due to date.
AI, right? So that's $5 or $6 billion annualized. And so, you know, one thing that is a little
bit perhaps not talked about is there's a lot more stuff coming. And over the next two years,
three years, et cetera, as enterprises really adopt us at scale, we should anticipate as well
that, you know, the need for compute will continue to grow. So it's really interesting to see
both this replacement cycle. You're talking about the massive spend by big tech on, um, on
LLMs because they're driving most of the spend on LLMs because they're there are the big
rounds right the big rounds aren't venture capitalists investing billions of dollars is the
big tech companies it's Amazon and Google and Microsoft and sales force in NVIDIA actually
right and then there's the enterprise adoption which is still TBD so yeah there's a lot going on
on this point if you look back a mom you know AI years are like dog years so a year to the
meta earnings beat at the end of January. Did you see this article that David Kahn wrote at
Sequoia, the 200, like AI's $200 billion question? Was this where he basically said based on the
spend, if you think of the ROI you need, then you need to generate hundreds of billions of dollars
in return. Yeah. Or justify all the, yeah, all the spend that you had, yeah. Yeah, very succinct
summary. And I was like, okay, yeah, that is the question. And I feel like the meta earnings beat was
the, like, one-day answer to that question, right? So to your point, they're one of the large
spenders. They said they're going to spend $30 to $37 billion on KAPX in 2024, driven by, like,
AI, driven by servers, right? Mark has this great, like, quote where he's talking about
600K, H-100 equivalent units of compute and saying, like, there's no room for other people. But the response
to all of the investment that has in CAPEX for training and inference at meta over the prior
years has been like a huge earnings beat from better targeting, leading to better conversion,
better recommendations, leading to better engagement, better advertising tools, leading in a better
ROI, as well as like the cost controls that the rest of the industry is doing.
And so they had this one day, I thought it was really nice that the number was exactly this too,
that this one-day ad of $197 billion of market cap.
Biggest single-session ad before Nvidia,
I forget where Nvidia ended up landing after their beat.
But, like, that's the answer, right?
Like, you know, $197 billion of increase in enterprise value on $25, $30 billion of CAPX.
Like, you should keep doing it.
Yeah.
Yeah, it's kind of amazing.
It's kind of a related question because I remember Uri Milner showed me this chart,
which basically he looked at the aggregate increase in startup market cap
and the aggregate increase of what at the time was like feying market cap
and obviously now there's like the magnificent seven or eight or whatever it is.
And so if you looked at the top tech companies at the time,
they added like, I don't know what it was five or ten times the market cap
of all the startup ecosystem combined during the same period of time.
And to some extent you could argue we're going into the same thing,
at least in the short run for AI and we still haven't seen the monster AI companies
and merge from scratch and indebtedly those will exist.
But at least for the next few years,
it seems like where we're going to see that really huge market cap incremental
add, maybe companies like OpenEI and some of the model companies,
but also it seems like increasingly it's just going to be existing companies
adding huge amounts of revenue and earnings and compute and everything else along the way.
So it's back to maybe the right thing to do right now is to start a hedge fund.
I think that also begs a question of how to think about, like, all of the other companies, like, tech and not in terms of amount of impact from AI.
I actually think it would be like a really fun lens to run a hedge from with, because you can take a, you can take a very long-term view of something that feels very secular and just classify companies this way and long short, like take that strategy.
is the only lens. Because, like, I do think that there are a number of services companies
that are squarely in the sites of things that you will be able to significantly automate.
And the only question is which of these management teams is going to have the investment capability,
technical talent, guts conviction to invest the way Mark did through, you know, people were really
get mad about the CAPEX spent for a few years at meta, right? And I think the answer is mostly,
especially some of these services firms, like maybe they partner to get there, but they mostly
will not make the transition, I think. The other thing it isn't really discussed is the impact
it's already having on some businesses. So obviously service now had like a blowout quarter in part
due to AI. So we're starting to see a little bit of that enterprise adoption. One of the folks
from Klarna posted today that they built an AI assistant that's powered by OpenEI, and it's
first four weeks
handled 2.3 million
customer service chats for them.
And so I ended up handling
two-thirds of all their customer
service inquiries.
It was on par with humans
in terms of customer satisfaction.
It was higher accuracy.
So it led to a 25% reduction
and repeat queries.
Customers resolve their errands in
two minutes versus 11 minutes.
It's live 24-7
in over 23 markets
communicating in over 35 languages.
And it performed
the equivalent job of
700 full-time agents.
And so basically, Klarna, in, you know, a few months or a year or however long it took
him to build this, built this customer service chat product, and it replaced 700 people's
worth.
And they say that at this point, they have something like 3,000 full-time agents, and so it cut
the agents needed by about 25%.
Right.
And so it's this really interesting post from Klarna where they announced this.
And then one of the things they announced as part of that is, you know,
longer-term society needs to think about what this means for society because this technology
seems to be so good for certain human-level tasks. And this is back to that point of AI adoption
and the enterprise is just starting. But how many years is it before every enterprise realizes
that they can cut customer support dramatically, at least for certain types of products,
just through adding simple apps? And so I think that's the other thing that is kind of happening
in the background that isn't talked about that much. But, you know,
was already starting to really show its face in pretty interesting ways.
Yeah, well, I do think you're going to get this accelerated adoption that goes use case by use case,
right, where like in any market, you have early adopters that build it in-house or go get
these solutions and are willing to take the risk when you don't actually know, like,
what the impact will be, how well it will work.
But as soon as one payments company does that, and it's a better experience for the customer,
or it has real, like, impact on operating cost,
I think, like, you switch very quickly over to the entire sector
being like, we have to adopt it in order to be competitive on both fronts.
Yeah, this stuff tends to happen slowly and then suddenly all at once,
and I think we're in the slowly phase right now.
And I actually had my team go and take global services and look at that, right?
And so if you look at spend on software in the U.S. right now,
it's about half trillion dollars.
and software spend a year,
if you look at
human-centric services,
just payroll for things
where Gen. AI can probably impact things.
It's $3.5 to $5 trillion.
So if you convert just 10% of that spend
into AI revenue,
you've effectively recreated
the entire U.S. market software industry
and market cap, right?
And so these are huge trends that are coming,
and you can kind of imagine
vertical by vertical,
what are those things going to be?
And then you can,
can ask, is it going to be built as internal tools for companies? Is it going to be a new
company that emerges that serves these things? Or is it going to be an incumbent who figures it out
and adds it? And so this sort of customer support chatbot thing, you know, you would have thought
that there's a company doing this for everyone. And it looks like in this case, they just did it
internally or in-house. But you could also imagine an existing company like a Zendesk or somebody
adopting to this. And the real question is, which of those three scenarios?
is going to happen, at least from a startup perspective.
But from a technology wave perspective, this is massive, right?
And you can build in the feedback loops really easily for this type of product, right?
Because you can have the customer rated or thumbs down at the end of the session, et cetera.
So you have a really good sort of RLHF or some sort of training support as well.
So it's a product that should get better and better and better over time as you use it more.
Yeah, I think one of the things that is an indicator of like where that services spend might,
be that gets externalized is actually like the big tech companies actually have you know they're
tech companies but they have broader businesses than um i think sometimes they're given credit for right
like facebook meta interacts with smbs as advertisers if you look at anybody who has this like large
commerce um type customer base so as you just mentioned clarna or square or meta or shopify like
they've all done this now, and it's working, right? And so I think the fact that these are the
companies that have the technical teams that are capable of doing it in-house is a nice
indicator for, like, well, if it's that effective, everybody else should too. And the question
is, I think not every segment of customer, like retailers with enough of a technical team to
build an e-commerce presence may not build this themselves. Then it's a more likely scenario that
either an incumbent or a new company, be it in Sierra or something else,
ends up owning that customer service segment.
Yeah, 100%.
Yeah, we have a longness internally of like the companies that I think should exist in
this space, right?
Because there's so many obvious ones.
And very few companies exist for most of them, if any, companies.
And so I think it's back to this idea that there are these human capital waves happening
in AI.
And the very first wave we saw was researchers, and they built early,
model companies and they built some of the early applications like Replexity and Harvey and all these
things were actually started by people who were working on models initially. And they were just
closest to the technology so they knew what to do. And then the second wave of human capital was
like infra people because there was the second closest to LLMs. And then the third wave, of course,
is going to end up being application builders. But many of them were not aware that any of this
stuff was important until chat GPT came out 15 months ago. And they're just starting to show up,
right? It takes some nine months stick with their job and a few months to figure out what to do and find a
co-founder and a few months of built prototype. And so we haven't seen anything yet really on the
app wave. You know, all the apps or many of the apps so far, we're started by people who were
very close to the research community and then it's kind of permeated into other areas with
some things growing really fast, right? There's like half a dozen medical scribing apps that all
seem to be growing at a pretty good pace or there's a few other application areas where it seems
like there's a number of people working, but then there's lots and lots of spaces where it seems
like nobody's doing anything, which is kind of weird, honestly. Yeah, there's a joke.
that the foundation model companies are here to replace all the jobs, but they don't understand
what any of the jobs are. And I think there's like a little, a little bit of truth in the sort of
exposure to what happens in, you know, a broad range of companies in terms of functions and
outsource services. And so I think that is the opportunity, right? Like now it's a race for people
who are just great engineers, smart about a domain to go experiment on the fringe of that. And I still
think there's opportunities around. Like you and I have talked about, um, the, uh, domain areas where
you might want specific models or verticalized companies still. And we should, we should talk about
that. But I, uh, I, my team and I just gave a presentation at this AI and production conference
about how if 2023 was the year of infrastructure, like 24 is the year we begin to see applications.
So I think we're pretty aligned there. I do want to ask you like one thing before we move away from
all of the earnings stuff, which is,
The most obvious place somebody is already making money is either, like, cloud providers,
inference providers, or just Nvidia as a chipmaker.
What would it take to compete to have like a second source with invidia?
I think there's a few different approaches, right?
I mean, fundamentally, if you look at what people claim as a defensibility and part of
Nvidia, it's a mix of chip performance, Kuda, and Interconnect.
You know, Nvidia bought Melanix back in 2019.
It was an Israeli company
to basically provide the interconnect side.
I think that was like a $5 billion acquisition,
so it's quite large relative to NBidi's market cap at the time.
And then obviously, Kuda has been developed over many years.
And then obviously, they've iterated really well
on these sort of different generations of chips.
So minimally, you at least need some form of silicon
in this performance,
and then you need to make sure that you're able to use it
effectively and then you're able to scale it,
which is sort of the interconnect side.
And there's the incumbent side of it, right?
AMD is obviously working on this, Intel's trying to, et cetera.
And then there's the startup side of it where we've seen things like Rock emerge
where they have very fast inference for open source models as well as language models,
which is pretty striking.
You have Sarah Brass, which has taken a fundamentally different approach to the chip side as well.
So, you know, there's a few startups that I think have some interesting early hardware,
and then there's some new companies like ads that have talked publicly about how they're really focused
on transformer-based models
and architectures for chips that they're building.
So there will be this potential
wave of second sourcing over time,
but in general,
if you look at many of the most advanced
chip markets historically, at least,
there's tended to be a winner
or I should say a leader,
and then there's tend to be a second place
party. And that was, you know,
during the microprocessor world,
that was Intel and then AMD was number two.
And, you know,
in mobile, it kind of
a little bit, right? You had Qualcomm and Arm doing different things, but both quite successfully,
but I think Qualcomm was always, at least for a period of time with the bigger company,
although Arm is much larger now. I should actually check that in terms of market cap.
Yeah, Qualcomm is $176 billion, and then Arm is $140 billion. So they're pretty close, actually,
no. There used to be a pretty big disparity between the two, and part that's because Arm is being
used now in sort of broader ways. So, you know, you kind of tend to say,
see these market structures and semiconductors where there's a leader and then a second place.
And I think part of that is traditional Moore's Law chip generation related stuff.
I don't know how that will hold up or how that will morph in AI.
I don't know if you have an opinion on that.
Yeah, the, you know, the way Jensen has described advancements in chip performance
tend to be more memory management and new techniques versus just like,
transistors fitting on a particular dye size.
And I think somebody else said,
NVIDIA called it Jensen's law of, like,
ability to get performance from full system.
But the only thing I'd add to your description of competitiveness here
is also like manufacturing,
even for these fabulous chip design companies, is a big deal, right?
Like, so you got to do what you said,
design something better, including interconnect,
design an entire, like build an entire software ecosystem,
could have been around since 2006.
but after that you have to go get capacity at TSM, right?
And then you need to get yield up
and then you need it all to be competitive
in terms of pricing.
I think the desire, like the economic pressure
given $2 trillion of market cap
and more demand than Nvidia can support
is higher than ever,
but I think the moat is actually really, really deep.
And so when I think about like what could be enough
to go disrupt that,
I'm sure you've seen many of these companies,
but I've seen a few different approaches.
It could be a chip and system designed for like specifically very much around latency.
But the other thing that you said, right, like something, for example, optimized two transformers
as an architecture, you're taking a bet around how much stability there is around a particular
architectural approach.
And I think that's felt like a quite good bet for a while now.
But for the first time and a long time, there is some interest in things like statespace models with companies like Cartesia and some alternatives, right?
If you're a really big company with your own use case, right?
If you're meta or you're Google and you either have like the entire ad system recommendations serving, spam, et cetera, or all that like search and your own cloud, then you don't need to make everything work on the software.
ecosystem side, you just need to make one application work. And these companies also acquired teams
in. But that's how you end up with like TPUs and traniums and all that. But I would love to
meet companies in this area and still haven't seen something that's gotten me over the edge,
even in a place that it's so obviously economically fertile. Yeah, I think one thing you pointed out,
which was interesting to expand a little bit on, is TSM. And the whole fabulous semiconductor world
where you're basically, you know, outsourcing the development or the manufacturing of the chip to a handful of players, TSM being the bigger, but there's the biggest, but there's one or two others that are big enough to at least handle some volume.
And, you know, there's been this push to try and repatriate semiconductor manufacturing to the U.S.
And it's run into all sorts of obstacles that are pretty avoidable, environmental reviews that go on endlessly or other things that have prevented.
People actually starting to build these things that take many years to build.
and it's been interesting to watch that in Japan,
they're starting to actually have really interesting development of fabs
specifically for this purpose.
And so I'm increasingly wondering whether Japan emerges
as sort of a second source location
in part to geopolitically hedge Taiwan.
But I think that's something also to kind of watch
in terms of where are you actually seeing fabs go out
and how do you think about that geographic distribution,
but also why is the U.S. in some sense,
getting in its own way for something that has pretty broad-based
to treat it important.
on multiple levels, including national security ones.
If you listen to the TSM CEO about this,
he talks as much about the human capital
and the cultural elements of human capital
required to make a place like TSM work
as the CABX spend, right?
And the access to equipment,
and the need to actually build the fab.
I think that's pretty interesting because, like, you know, we can invest a great deal, but it's very hard to change culture.
And so I do think that there's one version of, like, maybe you have fabs in Japan or Mexico or Southeast Asia or a, like, just a broader global supply chain for chip production.
Or maybe you have robots making chips.
Yeah, I mean, that's all true.
but the flip side of it is Intel has manufactured chips in the U.S. for a long time.
T.I. did historically, right? But Intel still does. So I don't think there's a complete lack of
human capital. Obviously, it's concentrated in part in Taiwan and to a secondary extent in Korea right now.
But I do think there's the capability to do it. And I think, again, there are other things
that are getting in the way, I think even before that. Can you even break around on the plant?
Maybe step one, right? We should start with the basics. And then we can deal with culture
when we actually have a fad.
Yeah.
Well, and I'm, I guess, very willing to believe that these companies and industries didn't exist
in the places they do without, like, great leaders for TMSC or otherwise.
And so, like, maybe it's not a solvable problem.
Like, I'd be curious if you believe in the Intel fad business that they're trying to push
and push to other customers now.
But to me, it's not binary.
It's like, of course, we can, like, make chips in America.
The question is, can we make them without the churn and with the yield and cost to make them competitive?
But maybe it's so important, like, you don't need them to be competitive for some period of time.
Yeah, and also my point is we're already doing that for Intel, right?
Intel's fat businesses in the U.S.
Not the fabulous TMCC style business.
Just making their own chips.
They've been doing it for decades in the U.S.
It's been fine.
It's been high yield, you know.
Yeah, it's been fine, but it's also been behind in terms.
of process technologies, right?
But maybe that's not a human capital issue.
Maybe it's other issues at Intel.
Yeah, it seems like it's a other issue.
Yeah.
I think my general take on the whole market is
the more I learn, the less I know in AI
and it's the opposite of every other field I've ever been in.
Really?
Really?
Yeah, usually the more you learn about something,
yeah, usually the more you learn about something,
the more you can create sort of these straight line hypotheses
or, you know, what you know kind of compounds
and it's static.
And I feel in the AI world, like every week there's like so many new things that your entire world model shifts.
In like a fun way.
Yes. It's fun to be exhausted. But I think, you know, there's just so much going on and the base of innovative.
It really feels like, you know, that early slope into the, you know, the exponent that is a singularity or however you want to phrase it.
But it really feels like this self-reinforcing loop of new stuff.
And honestly, a lot of it was kind of held back in the larger tech companies.
And now it's kind of flourishing externally, and that's creating competitive pressure.
The larger companies and the larger companies are reacting, and that's spawning more startups,
and it's just this really interesting virtuous cycle.
And to some extent, the big tech companies are helping fuel it all by then funding the companies
that are working at very late stages with huge rounds.
And they're funding a lot of the compute in the industry in a way that's, you know,
at least in order of magnitude, maybe two orders of magnitude more than what the venture
community is doing. And so it's this really interesting virtuous cycle of startups come out that
accelerates big tech doing stuff. That causes some people to leave big tech to do some interesting
things externally. They then get funded by big tech and that accelerates both themselves and
big tech. And you have this kind of interesting cycle happening right now. So it's very exciting
days. Yeah, I drew, I drew a slide that has like, as you might hope, like a bunch of reinforcing cycles.
it's very fancy.
And the one I would add to that is like what we started talking about, which is when
something begins to work, if it is actually valuable, like the Klarna thing that you describe,
like at some point, if it's valuable and it moves the needle in the business, you have to do it,
like, as a part of the competitor set.
And so I think like we started with this like narrative driven thing where, you know,
CEOs would say that they're going to do AI because like the markets believed that was
the future and it was very generic.
generic. And you see that show up in the spending numbers or at least the expectations around
spend. I was looking at this survey from one of the investment banks that says like Fortune
1,000 IT budgets go to 5 to 8% this year instead of 3 to 5% generally. And it's all because
of AI. Like that's pretty big, right? That's like two fucking X. And like if that's true,
then that's also part of the reinforcement cycle here. Because if the companies start to work,
then they get to continue building these products, VCs, you know,
or investors like us will keep trying.
So I think it's pretty exciting.
Yeah, it's RLPA, yeah.
Yeah, RLPA, yeah.
Rolls right off the ton.
Reinforcement learning through product adoption feedback.
You're welcome.
Well, I'm just going to plug that in to chat ChbT and have it write the paper.
But I will be sponsoring author if you'll be first author.
Yeah, I'll see if I include you.
Academic violence.
Find us on Twitter at No PryorsPod.
Subscribe to our YouTube channel if you want to see our faces.
Follow the show on Apple Podcasts, Spotify, or wherever you listen.
That way you get a new episode every week.
And sign up for emails or find transcripts for every episode at no dash priors.com.
Thank you.