a16z Podcast - Dwarkesh and Ilya Sutskever on What Comes After Scaling
Episode Date: December 15, 2025AI models feel smarter than their real-world impact. They ace benchmarks, yet still struggle with reliability, strange bugs, and shallow generalization. Why is there such a gap between what they can d...o on paper and in practiceIn this episode from The Dwarkesh Podcast, Dwarkesh talks with Ilya Sutskever, cofounder of SSI and former OpenAI chief scientist, about what is actually blocking progress toward AGI. They explore why RL and pretraining scale so differently, why models outperform on evals but underperform in real use, and why human style generalization remains far ahead.Ilya also discusses value functions, emotions as a built-in reward system, the limits of pretraining, continual learning, superintelligence, and what an AI driven economy could look like. Resources:Transcript: https://www.dwarkesh.com/p/ilya-sutsk...Apple Podcasts: https://podcasts.apple.com/us/podcast...Spotify: https://open.spotify.com/episode/7naO... Stay Updated:If you enjoyed this episode, be sure to like, subscribe, and share with your friends!Find a16z on X: https://x.com/a16zFind a16z on LinkedIn: https://www.linkedin.com/company/a16zListen to the a16z Podcast on Spotify: https://open.spotify.com/show/5bC65RDvs3oxnLyqqvkUYXListen to the a16z Podcast on Apple Podcasts: https://podcasts.apple.com/us/podcast/a16z-podcast/id842818711Follow our host: https://x.com/eriktorenbergPlease note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details please see a16z.com/disclosures](http://a16z.com/disclosures. Stay Updated:Find a16z on XFind a16z on LinkedInListen to the a16z Show on SpotifyListen to the a16z Show on Apple PodcastsFollow our host: https://twitter.com/eriktorenberg Please note that the content here is for informational purposes only; should NOT be taken as legal, business, tax, or investment advice or be used to evaluate any investment or security; and is not directed at any investors or potential investors in any a16z fund. a16z and its affiliates may maintain investments in the companies discussed. For more details please see a16z.com/disclosures. Hosted by Simplecast, an AdsWizz company. See pcm.adswizz.com for information about our collection and use of personal data for advertising.
Transcript
Discussion (0)
Now that compute is big, computer is now very big.
In some sense, we are back to the age of research.
We got to the point where we are in a world
where there are more companies than ideas by quite a bit.
Now there is the Silicon Valley saying that says
that ideas are cheap, execution is everything.
What is the problem of AI and AIGI?
The whole problem is the power.
AI models look incredibly smart on benchmarks, yet their real-world performance often feels far behind.
Why is there such a gap and what does that say about the path to AGI?
From the Dwar Keshe podcast, here's a rare, long-form conversation with Ilya Sutskiever,
co-founder of SSI, exploring what's actually slowing down progress toward AGI.
Dwar Keshe and Ilya dig into the core problems in modern AI systems,
from why RL and pre-training scale so differently, to why generalization, reliability, and sample efficiency
still fall short of human learning.
They also explore continual learning,
value functions,
superintelligence,
and what a future economy
shaped by AI might look like.
You know what's crazy?
I know.
That all of this is real.
Yeah.
Don't you think so?
Meaning what?
Like all this AI stuff
and all this big area.
Yeah, that it's happened.
Like, isn't it straight out of science fiction?
Yeah.
Another thing that's crazy
is like how normal the slow takeoff feels.
the idea that we'd be investing 1% of GDP in AI
like I feel like it would have felt like a bigger deal
you know where right now it just feel like
we get used to things pretty fast turns out yeah
but also it's kind of like it's abstract like
what does it mean
what it means that you see it in the news
that such and such company announced such and such dollar amount
right that's all you see right
it's not really felt in any other way so far
yeah should we actually begin here
I think this is an interesting discussion
Sure.
I think your point about, well, from the average person's point of view,
nothing is that different will continue being true even into the singularity.
No, I don't think so.
Okay, interesting.
So the thing which I was referring to not feeling different is, okay, so such and such company announced some difficult to comprehend dollar amount of investment.
Right.
I don't think anyone knows what to do with that.
Yeah.
But I think that the impact of AI is going to be felt.
AI is going to be diffused through the economy.
There are very strong economic forces for this.
And I think the impact is going to be felt very strongly.
When do you expect that impact?
I think the models seem smarter than their economic impact would imply.
Yeah.
This is one of the very confusing things about the models right now.
how to reconcile the fact
that they are doing so well on e-vals
and you look at the e-vals and you go
those are pretty hard e-vails
they are doing so well
but the economic impact seems to be
dramatically behind and it's almost like
it's very difficult to make sense
of how can the model on the one hand do these amazing things
And on the other hand, like repeat itself twice in some situation
in a kind of an example would be,
let's say you use vibe coding to do something.
And you go to some place and then you get a bug.
And then you tell the model, can you please fix the bug?
And the model says, oh my God, you're so right, I have a bug.
Let me go fix that.
And it reduces a second bug.
And then you tell you, you have this new, this second bug.
And it tells you, oh, my God, how could have done it?
You're so right again.
and brings back the first bug
and you can alternate between those
and it's like how is that possible
it's like
I'm not sure
but it does suggest that
something strange is going on
I have two possible explanations
so here this is the more kind of
a whimsical explanation
is that maybe REL training
makes the models a little bit too single-minded
and narrowly focused
a little bit too
I don't know
unaware, even though it also makes them aware
in some other ways. And because of this, they can't do
basic things. But there is another explanation, which is
back when people were doing pre-training,
the question of what data to train on was answered
because that answer was everything.
When you do pre-training, you need all the data.
so you don't have to think
it's going to be this data or that data
but when people do RRL training
they do need to think
they say okay we want to have
this kind of RL training for this thing
and that kind of RL training for that thing
and from what I hear
all the companies have teams
that just produce new RL environments
and just out it to the training mix
and the question is well what are those
there are so many degrees of freedom
there is such a huge variety
of REL environments you could produce
and one of the one thing you could do
and I think that's something that is done inadvertently
is that people take inspiration from the evals
you say hey I would love our model to do really well
when you release it I want the evals to look great
what would be RL training that could help on this task right
I think that is something that happens
and I think it could explain a lot of what's going on
If you combine this with generalization of the models actually being inadequate,
that has the potential to explain a lot of what we are seeing,
this disconnect between e-val performance and actual real-world performance,
which is something that we don't today exactly even understand what we mean by that.
I like this idea that the real reward hacking is the human researchers who are too focused on the evils.
I think there's two ways to understand or to try to think about what you have just pointed out.
One is, look, if it's the case that simply by becoming superhuman at a coding competition,
a model will not automatically become more tasteful and exercise better judgment about how to improve your code base,
well, then you should expand the suite of environments such that you're not just testing it on having the best performance in a coding competition.
It should also be able to make the best kind of application for X thing or Y thing or Z thing.
And another, maybe this is what you're hinting at, is to say, why should it be the case in the first place that becoming superhuman at coding competitions doesn't make you a more tasteful program or more generally?
Maybe the thing to do is not to keep stacking up the amount of environments and the diversity of environments to figure out an approach would let you learn from one environment and improve your performance on.
on something else.
So I have a human analogy
which might be helpful.
So even the case, let's take
the case of competitive programming since you mentioned
that. And suppose you have two students.
One of them
worked decided they want to be
the best competitive programmer so they will
practice 10,000 hours
for that domain.
They will solve all the problems,
memorize all the proof techniques and be very
very, you know,
be very
skill that quickly and correctly
implementing all the algorithms
and by doing so they became
the best, one of the best.
Student number two thought,
oh, competitive programming is cool.
Maybe they practiced for 100 hours,
much, much less, and they also did really well.
Which one do you think is going to do better
in their career later on?
The second. Right? And I think that's
basically what's going on. The models are much more
like the first student, but even more, because then we
say, okay, so the model
should be good competitive programming, so let's get
every single competitive programming problem ever
and then let's do some data augmentation so we have even more
competitive programming problems and we train on that
and so now you got this great competitive programmer and with this
analogy I think it's more intuitive I think it's more intuitive with this
analogy that yeah okay so if it's so well trained okay it's like
all the different algorithms and all the different proof techniques are like
right at its fingertips and it's more intuitive that with this level of
preparation, it would not necessarily generalize to other things.
But then what is the analogy for what the second student is doing before they do the
100 hours of fine-tuning?
I think it's like they have it.
I think it's the it factor.
Yeah.
Right?
And like I know, like when I was an undergrad, I remember there was a student like this
that studied with me.
So I know it exists.
Yeah.
I think it's interesting to distinguish it from whatever.
pre-training does. So one way to understand
what you just said about
we don't have to choose the data in pre-training
is to say, actually, it's not
dissimilar to the 10,000 hours of practice.
It's just that you get that 10,000 hours of practice
for free because it's
already somewhere in the
pre-training distribution. But it's like
maybe you're suggesting actually
there's actually not that much generalized just from pre-training.
There's just so much data in free training.
But it's like it's not necessarily generalizing better than RL.
Like the main strength of pre-training
is that there is, A, so much
of it.
Yeah.
And B, you don't have to think hard about what data to put into pre-training.
And it's a very kind of natural data, and it does include in it a lot of what people do.
Yeah.
People's thoughts and a lot of the features of, you know, it's like the whole world as projected
by people onto text.
Yeah.
And pre-training tries to capture that using a huge amount of data.
it's very pre-training is very difficult to reason about
because it's so hard to understand the manner
in which the model relies on pre-training data
and whenever the model makes a mistake
could it be because something by chance
is not as supported by the pre-training data
you know and support by pre-training is maybe a loose term
I don't know if I can add anything more useful
on this, but I don't think there is a human analog to pre-training.
Here's analogies that people have proposed for what the human analogy to pre-training is,
and I'm curious to get your thoughts on why they're potentially wrong.
One is to think about the first 18 or 15 or 13 years of a person's life
when they aren't necessarily economically productive,
but they are doing something that is making them understand the world better and so forth.
And the other is to think about evolution as doing some kind of search for 3 billion years,
which then results in a human lifetime instance.
And then I'm curious if you think either of these are actually analogous to pre-training
or how would you think about at least what lifetime human learning is like, if not pre-training?
I think there are some similarities between both of these to pre-training,
and pre-training tries to play the role of both of these.
But I think there are some big differences as well.
the amount of pre-training data is very, very staggering.
Yes.
And somehow a human being, after even 15 years,
with a tiny fraction of that pre-training data,
they know much less.
But whatever they do know, they know much more deeply, somehow.
And the mistakes, like already at that age,
you would not make mistakes that RIAs make.
Yeah.
There is another thing.
You might say, could it be something like evolution?
And the answer is maybe, but in this case, I think evolution might actually have an edge.
Like, there is this, I remember reading about this case where some, you know, that's one thing that neuroscientists do,
or rather one way in which neuroscientists can learn about the brain, is by studying people with brain damage to different parts of the brain.
And some people have the most strange symptoms you could imagine.
It's actually really, really interesting.
and there was one case that comes to mind that's relevant
I read about this person
who had some kind of brain damage
that took out I think a stroke or an accident
that took out his emotional
processing so he stopped feeling any emotion
and as a result of that
he still remained very articulate
and he could solve little puzzles
and on tests he seemed to be just fine
but he felt no emotion
he didn't feel sad he didn't feel
angry he didn't feel animated
and he became somehow
extremely bad at making any decisions at all
it would take him hours to decide
on which socks to wear
and he would make very bad financial decisions
and that's very
what does it say
about
the role of our built in emotions
in making us
like a viable agent essentially
and I guess to connect to your question
about pre-training
it's like maybe
if you're good enough
at like getting everything out of pre-training
you could get that as well
but that's the kind of thing
which seems
well
it may not be possible
to get that
from pre-training
what is
that
clearly not just directly emotion
it seems like some
almost value function like thing
which is telling you which decision to be
like what the end reward for any decision should be
and you think that doesn't sort of implicitly come from
I think it could
I'm just saying it's not one it's not 100% obvious
but what is that
like how do you think about emotions
what is the ML analogy for emotions
it should be some kind of a value function thing
Yeah.
But I don't think there is a great ML analogy
because right now value functions
don't play a very prominent role
in the things people do.
It might be worth defining for the audience
what a value function is
if you want to do that.
I mean, certainly,
I'll be very happy to do that, right?
So when people do reinforcement learning,
the very reinforcement learning is done right now,
how do people train those agents?
So you have a neural net.
and you give it a problem,
and then you tell the model go solve it.
The model takes maybe thousands, hundreds of thousands of actions
or thoughts or something,
and then it produces a solution, a solution is created.
And then the score is used to provide a training signal
for every single action in your trajectory.
So that means that if you are doing something that goes for a long time,
if you're training a task that takes a long time to solve,
you will do no learning at all
until you came up with the proposed solution.
That's how reinforcement learning is done naively.
That's how 01, R1 ostensibly are done.
The value function says something like,
okay, look, maybe I could sometimes, not always,
could tell you if you are doing well or badly.
The notion of a value function is more useful in some domains than others.
So, for example, when you play chess and you lose a piece, you know, I messed up.
You don't need to play the whole game to know that what I just did was bad and therefore whatever preceded it was also bad.
So the value function lets you short circuit the weight until the very end.
Like let's suppose that you started to pursue some kind of, okay, let's suppose that you are doing some kind of a math thing or a programming thing.
and you're trying to explore a particular solution direction
and after let's say after a thousand steps of thinking
you concluded that this direction is unpromising
as soon as you conclude this
you could already get a reward signal
a thousand times steps previously
when you decided to pursue down this path
you say oh next time I shouldn't pursue this path
in a similar situation
long before you actually came up with a proposed solution
This was in the Deep Sikar 1 paper, is that the space of trajectories is so wide that maybe it's hard to learn a mapping from an intermediate trajectory and value.
And also given that, you know, encoding, for example, you'll have the wrong idea, then you'll go back, then you'll change something.
This sounds like such lack of face in deep learning.
Like, I mean, sure, it might be difficult, but nothing deep learning can't do.
Yeah.
So my expectation is that value function should be useful
and I fully expect that they will be used in the future, if not already.
What was I alluding to with the person whose emotional center got damaged
is more that maybe what it suggests is that the value function of humans
is modulated by emotions
in some important way
that's hard-coded by evolution.
And maybe that is important
for people to be effective in the world.
That's the thing I was actually going to
planning on asking you.
There's something really interesting
about emotions of the Valley Function,
which is that it's impressive
that they have this much utility
while still being rather
simple to understand.
So I have two responses.
I do agree
that compared to
the kind of things that we learn
and the things we are talking about,
the kind of as we are talking about,
emotions are relatively simple.
They might even be so simple
that maybe you could map them out
in a human understandable way.
I think it would be cool to do.
In terms of utility, though,
I think there is a thing where,
you know, there is this complexity,
robustness, straight off.
where complex things can be very useful
but simple things
are very useful in a very broad range of situations
and so I think one way to interpret what we are saying
is that we've got these emotions
that essentially evolved mostly
from our mammal ancestors
and then fine-tuned a little bit
while we were hominids just a bit
we do have like a decent amount of social emotions
though, which mammals
may lack.
But they're not very sophisticated.
And because they're not sophisticated
the service so well in this very
different world compared to the one that we've
been living in. Actually, they also
make mistakes. For example, our
emotions, well, I don't know, this hunger
count as an emotion?
It's debatable, but I think, for example,
our intuitive
feeling of hunger
is not
succeeding in
guiding us correctly
in this world with an abundance of
food. Yeah. People have been
talking about scaling data,
scaling parameter, scaling compute.
Is there a more
general way to think about scaling? What are the other
scaling axes?
So
the thing, so
here is a perspective. Here's a perspective
I think might be
true.
So
the way
ML used to work
is that people would just tinker with stuff
and try to
and try to get interesting results.
That's what's been going on in the past.
Then
the scaling insight arrived, right?
Scaling laws.
GPT3. And suddenly
everyone realized we should scale.
And it's just
this is an example of how
language affects thought
scaling is just one word
but it's such a powerful word
because it informs people what to do
they say okay let's try to scale things
and so you say okay so
what are we scaling
and pre-training was a thing to scale
it was a particular scaling
recipe yes
the big breakthrough of pre-training
is the realization that this recipe
is good so you say hey
if you mix some compute
with some data into a neural net of a certain size
you will get results
and you will know that it will be better
if you just scale the recipe up
and this is also great companies love this
because it gives you a very
low risk way
of investing your resources
right
it's much harder to invest your resources in research
compare that
you know if you research you need to have like
go forth researchers and research and
come up with something versus get more data, get more compute, you know it'll get something
from pre-training. And indeed, you know, it looks like I based on various things, some people say on
Twitter, maybe it appears that Gemini have found a way to get more out of pre-training. At some point
though, pre-training will run out of data. The data is very clearly finite. And so then, okay,
what do you do next? Either you do some kind of souped-up pre-training,
different recipe from the one you've done before
or you're doing URL or maybe
something else. But now that
compute is big, computer is now very
big, in some sense we are back
to the age of research. So maybe
here's another way to put it. Up until
2020, from 2012
to 2020, it was the age of research.
Now, from 2020 to 2025
it was the age of scaling or maybe
plus minus, let's add the error bars
to those years. Because people say this is
amazing, you've got to scale more, keep scaling,
the one word, scaling.
But now the scale is so big,
like, is the belief really that, oh, it's so big,
but if you had a honeydex more,
everything would be so different.
Like, it would be different, for sure.
But, like, is the belief that if you just honeydex the scale,
everything would be transformed?
I don't think that's true.
So it's back to the age of research again,
just with big computers.
That's a very interesting way to put it.
but let me ask you the question you just posed then what are we scaling and what is what would it mean to have a recipe because i guess i'm not aware of a very clean relationship that almost looks like a law of physics which existed in pre-training there was a power of law between data or computer parameters and loss what is the kind of relationship we should be seeking and how should we think about what this new recipe might look like
So we've already witnessed a transition from one type of scaling to a different type of scaling,
from pre-training to RL.
Now people are scaling RL.
Now based on what people say on Twitter,
they spend more compute on REL than on pre-training at this point,
because REL can actually consume quite a bit of compute.
You know, you do very, very long rollouts.
Yes.
so it takes a lot of compute to produce those rollouts
and then you get relatively small amount of learning power rollouts
so you really can spend a lot of compute
and I could imagine
like I wouldn't at this at this state
it's more like I wouldn't even call it a scale
scaling I would say hey like what are you doing
and is the thing you are doing
the the most productive thing you could be doing
yeah can you find a most more productive way
of using your compute
we've discussed the value function business earlier
and maybe once people get good at value functions
they will be using their resources more productively
and if you find a whole other way of training models
you could say is this scaling or is it just using your resources
I think it becomes a little bit ambiguous in a sense that when people were in the age of research
back then it was like people say hey let's try this and this and this let's try that
And then that, oh, look, something interesting is happening.
And I think there will be a return to that.
So if we're back in the era of research, stepping back, what is the part of the recipe that we need to think most about?
When you say value function, people are already trying the current recipe, but then having LLM as a judge and so forth.
You can say that's a value function, but it sounds like you have something much more fundamental in mind.
Do we need to go back to should we even rethink pre-training at all and not just add more
steps to the end of that process.
Yeah. So the
discussion about value function,
I think it was interesting. I want
to emphasize that I think the value
function is something like
it's going to make REL more
efficient.
And I think that makes a difference.
But I think that anything you can do
with a value function, you can do
without just more slowly.
The thing which I think is the most fundamental
is that these models somehow just
generalize
dramatically worse
than people
and it's super obvious
that seems like
a very fundamental thing
okay
so this is the crux
generalization
and there's two
sub questions
there's one
which is about sample efficiency
which is
why should it take so much more data
for these models
to learn than humans
there's a second
about even separate
from the amount of data
it takes
there's a question of
why is it so hard
to tease the thing
we want to a model than to a human, which is to say, to a human, we don't necessarily need
a verifiable reward to be able to, you're probably mentoring a bunch of researchers right now,
and you're talking with them, you're showing them your code, and you're showing them how
you think, and from that, they're picking up your way of thinking and how they should do research.
You don't have to set like a verifiable reward for them that's like, okay, this is the next part
of your curriculum, and now this is the next part of your curriculum, and, oh, this training
was unstable and we got to there's not this schleppy bespoke process so perhaps these two issues are
actually related in some way but i'd be curious to explore this this second language was more like
continual learning and this first thing which feels just like um sample efficiency yeah so you know
you could actually wonder one one possible explanation for the human sample efficiency
that needs to be considered is evolution and evolution
evolution has given us a small amount of the most useful information possible.
And for things like vision, hearing, and locomotion,
I think there is a pretty strong case that evolution actually has given us a lot.
So, for example, human dexterity far exceeds, I mean, robots can become dexterous too
if you subject them to like a huge amount of training and simulation.
but to train a robot in the real world
to quickly pick up a new skill
like a person does
seems very out of reach
and here you could say
oh yeah like locomotion
all our ancestors
needed great locomotion
squirrels like
so locomotion may be like
we've got like some unbelievable prior
you could make the same case
for vision you know I believe
Jan Lechan made the point
oh like children
learn to drive after 16 hour
after 10 hours of practice
which is true, but our vision is so good.
At least for me, when I remember myself being five-year-old,
I was very excited about cars back then,
and I'm pretty sure my car recognition was more than addict
but self-driving already as a five-year-old.
You don't get to see that much data as a five-year-old.
You spend most of your time in your parents' house,
so you have very low data diversity.
But you could say maybe that's evolution too.
But then language and math and coding, probably not.
it still seems better than models.
I mean, obviously, models are better
than the average human at language and math encoding,
but are they better at the average human at learning?
Oh, yeah, oh yeah, absolutely.
What I meant to say is that language math and coding
and especially math encoding
suggests that whatever it is that makes people good at learning
is probably not so much a complicated prior,
but something more, some fundamental thing.
Wait, I'm not sure I'm interested.
Why should that be the case?
So consider a skill that people exhibit some kind of great reliability or, you know, if the skill is one that was very useful to our ancestors for many millions of years, hundreds of millions of years, you could say, you could argue that maybe humans are good at it because of evolution, because we have a prior.
An evolutionary prior that's encoded in some very non-obvious way
that somehow makes us so good at it.
But if people exhibit great ability, reliability, robustness, ability to learn
in a domain that really did not exist until recently,
then this is more an indication that people might have just better machine learning period.
But then how should we think about what that is?
Is it a matter of, yeah, what is the ML analogy for what?
There's a couple interesting things about it.
It takes fewer samples.
It's more unsupervised.
You don't have to set a very, like a child learning to drive a car.
A teenager learning to drive a car is like not exactly getting some pre-built
verifiable reward.
It comes from their interaction.
with the machine and with the environment.
And yet, it takes much of your samples.
It seems more unsupervised.
It seems more robust.
Much more robust.
The robustness of people is really staggering.
Yeah.
So is it like, okay, and do you have a unified way of thinking about
why are all these things happening at once?
What is the ML analogy that would, that could be,
could realize something like this?
So, so this is where, you know, one of the things that you've been asking about is,
How can, you know, the teenage driver kind of self-correct
and learn from their experience without an external teacher?
And the answer is, well, they have their value function, right?
They have a general sense, which is also, by the way, extremely robust in people.
Like, whatever it is, the human value function,
whatever the human value function is with a few exceptions around addiction,
it's actually very, very robust.
and so for something like a teenager that's learning to drive
they start to drive
and they already have a sense of how they're driving
immediately
how badly they're unconfident
and then they see okay
and then of course the learning speed of any teenager
is so fast after 10 hours you're good to go
yeah it seems like humans have some solution
but I'm curious about like well how are they doing it
and like why is it so hard to like
how do we need to reconceptualize the way we're training models
to make something like this possible
You know, that is a great question to ask, and it's a question I have a lot of opinions about.
But unfortunately, we live in a world where not all machine learning ideas are discussed freely, and this is one of them.
So there's probably a way to do it.
I think it can be done.
The fact that people are like that, I think it's a proof that it can be done.
there may be another blocker though which is there is a possibility that the human neurons
actually do more compute than we think and if that is true and if that plays an important
role then things might be more difficult but regardless I do think it points to the
existence of some machine learning principle that I have opinions on but unfortunately
circumstances make it hard to discuss in detail.
Nobody listens to this podcast, Alia.
Yeah.
I am curious, if you say we are back in an era of research,
you were there from 2012 to 2020.
And do you have, yeah, what is now the vibe going to be
if we go back to the era of research?
For example, even after Alex Net,
the amount of compute that was used to run experiments
kept increasing and the size of frontier systems kept increasing. And do you think now that
this era of research will still require tremendous amounts of compute? Do you think it will require
going back into the archives and reading old papers? What is, maybe what was the vibe of like
you were at Google and Open AI in Stanford at these places when there was like a more of a
vibe of research
what kind of
things should we
be expecting
in the community
so
one consequence
of the age
of scaling
is that
there was
this
scaling
sucked out
all the air
in the room
yeah
and so
because scaling
sucked out
all the air
in the room
everyone
started to do
the same
thing
we got to
the point
where
we are
in a world where there are
more companies than ideas
by quite a bit. Actually
on that, you know, there is this
Silicon Valley saying that says
that ideas are cheap
execution is everything.
And people say that a lot.
And there is truth to that.
But then I saw someone
say on Twitter something
like, if ideas
are so cheap, how come no one's having
any ideas? And I think
it's true too. I think
like if you think about
a research progress
in terms of bottlenecks
there are several bottlenecks
if you go back to the
and one of them is ideas
and one of them is your ability
to bring them to life
which might be compute
but also engineering
so if you go back to the 90s
let's say you had people
who had pretty good ideas
and if they had much larger computers
maybe they could demonstrate that their ideas
were viable but they could not
so they could only have very, very small demonstration
and did not convince anyone.
So the bottleneck was compute.
Then in the age of scaling,
computers increased a lot.
And of course,
there is a question of how much computer is needed,
but compute is large.
So compute is large enough
such that
it's like not obvious
that you need that much more compute
to prove some idea.
like I'll give you an analogy
Alexnet was built on two GPUs
that was the total amount of compute use for it
the transformer
was built on 8 to 64
GPUs no single transformer paper
experiment used more than 64
GPUs of 2017
which would be like what two GPUs of today
so the Resnet
right
many like even the
you could argue that the
like 01 reasoning
was not the most compute
heavy thing in the world
so they're definitely
for research
you need
definitely some amount of compute
but it's far from obvious
that you need the absolutely largest
amount of compute ever for research
you might argue
and I think it is true
that if you want to build the absolutely best system
if you want to build the absolutely
best system
so much more compute, and especially if everyone is within the same paradigm,
then compute becomes one of the big differentiators.
Yeah, I guess while it was possible to develop these ideas,
I'm asking you for the history because you were actually there.
I'm not sure what actually happened, but it sounds like it was possible to develop these ideas
using minimal amounts of compute, but it wasn't, the transformer didn't immediately become famous.
It became the thing everybody started doing and then started experimenting on top of
and building on top of
because it was validated
at higher and higher levels of compute.
Correct.
And if you at SSI have 50 different ideas,
how will you know which one is the next transformer
and which one is, you know, brittle
without having the kinds of compute
that other frontier labs have?
So I can comment on that,
which is the short comment is that,
you know, you mentioned SSI,
specifically for us
the amount of compute
that SSI has for research
is really not that small
and I want to explain why
like a simple math can explain
why the amount of compute that we have
is actually a lot more comparable
for research than one might think
now explain
so
SSI has raised
three billion dollars
which is like
not small by
it's like a lot by any
absolute sense but you could say but look at the other
companies raising much more
but a lot of
their compute goes for inference
like these big numbers
these big loans it's earmarked for inference
that's number one
number two
you need if you want to have a product
on which you do inference you need to have a big
staff of engineers
of salespeople
a lot of the research
needs to be dedicated for producing all kinds of product-related features.
So then when you look at what's actually left for research, the difference becomes a lot smaller.
Now, the other thing is that if you are doing something different, do you really need the absolute
maximal scale to prove it? I don't think that's true at all. I think that in our case,
we have sufficient compute
to prove to convince ourselves
and anyone else
that what we're doing is correct.
There's been public estimates
that companies like Open AI spend
on the order of
five, six billion dollars a year
just so far on experiments.
This is separate from the amount of money
they're sending on inference and so forth.
So it seems like they're spending more a year
running research experiments
than you guys have in total funding.
I think it's a question of what you do with it.
It's a question of what you do with it.
Like they have like is the more, I think in their case and the case of others,
I think there is a lot more demand on the training compute.
There's a lot more different work streams.
There are different modalities.
There is just more stuff.
And so it becomes fragmented.
How will SSI make money?
You know, my answer to this question is something like,
we just focus right now, we just focus on the.
research and then the answer to that question will reveal itself. I think there will be lots
of possible answers. Is exercise plans still to straight shot superintelligence? Maybe. I think that
there is merit to it. I think there's a lot of merit because I think that it's very nice to not
be affected by the day-to-day market competition. But I think there are two reasons that may cause us to
change the plan. One is pragmatic if timelines turned out to be long, which they might. And second,
I think there is a lot of value in the best and most powerful AI being out there impacting
the world. I think this is a meaningfully valuable thing. But then, so why is your default plan
to straight-trade superintelligence? Because it sounds like, you know, open AI, anthropic, all these
other companies, their explicit thinking is, look, we have weaker and weaker intelligences that
the public can get used to and prepare for. And why is it potentially better to build a super
intelligence directly? So I'll make the case four and against. Yeah. The case four is that you are,
so one of the challenges that people face when they're in the market is that they have to
participate in the rat race. And the rat race is quite difficult in that it exposed.
you to difficult trade-offs which you need to make.
And it is nice to say, we'll insulate ourselves from all this
and just focus on the research and come out only when we are ready and not before.
But the counterpoint is valid too.
And those are opposing forces.
The counterpoint is, hey, it is useful for the world to see powerful AI.
It is useful for the world to see powerful AI.
because that's the only way you can communicate it.
Well, I guess not even just that you can communicate the idea.
Communicate the AI.
Not the idea.
Communicate the AI.
What do you mean communicate the AI?
Okay, so let's suppose you read an essay about AI.
And the essay says AI is going to be this and AI is going to be that and it's going to be this.
And you read it and you say, okay, this is an interesting essay.
Right.
Now suppose you see an AI doing this and AI doing that, it is incomparable.
like basically I think that there is a big benefit from AI being in the public
and that would be a reason for us to not be quite straight shot yeah well I guess it's not
even that which I do think that is an important part of it the other big thing is I can't
think of another discipline in human engineering and research where the end artifact was made
safer mostly through just thinking about how to make it safe as opposed to why are airplane crashes
per mile so much lower today than there were decades ago? Why is it so much harder to find a bug in Linux
than it would have been decades ago? And I think it's mostly because these systems were deployed
to the world. You noticed failures. Those failures were corrected and the systems became more
robust. And I'm not sure why AGI and superhuman intelligence would be any different, especially
given, and I hope we can talk, we're going to get to this, it seems like the harms of
superintelligence are not just about like having some malevolent papercliper out there, but
it's just like, this is a really powerful thing, and we don't even know how to conceptualize
how people will interact with it, what people will do with it, and having gradual access
to it seems like a better way to maybe spread out the impact of it and to help people prepare
for it.
Well, I think on this point, even in...
In the straight-shot scenario, you would still do a gradual release of it.
Is how I would imagine it.
Gradualism would be an inherent component of any plan.
It's just a question of what is the first thing that you get out of the door.
That's number one.
Number two, I also think, you know, I believe you have advocated for continual learning
more than other people.
And I actually think that this is an important and correct.
thing and here is why
so one of the
things so I'll give you another
example of how
thinking, how language affects thinking
and in this case
this will be two words
two words that have shaped
everyone's thinking I maintain
first word
AGI
second word pre-training
let me explain
so the word
the term AGI
why does
this term exist? It's a very particular term. Why does it exist? There's a reason. The reason that
the term AGI exists is, in my opinion, not so much because it's like a very important
essential descriptor of some end state of intelligence, but because it is a reaction to a different
term that existed. The term is narrow AI. If you go back to ancient,
history of
gameplay and AI, of
checkers AI, chess AI, computer games AI
everyone would say look at this
narrow intelligence. Sure the
chess AI can beat Kasparov but it can't do
anything else. It is so narrow
artificial narrow intelligence.
So in response as a
reaction to this, some people
said, well this is not
good. It is so narrow.
What we need is general
AI.
General AI, an AI that can just do all the
things.
The second, and that term, just got a lot of traction.
The second thing that got a lot of traction is pre-training.
Specifically, the recipe of pre-training, I think the current, the way people do RL now is
maybe undoing the conceptual imprint of pre-training, but pre-training had the property.
You do more pre-training, and the model gets better at end.
everything more or less uniformly.
General AI.
Pre-training gives AGI.
But the thing that happened with AGI and pre-training
is that in some sense they overshot the target.
Because by the kind, if you think about the term AGI,
you will realize, and especially in the context of pre-training,
you will realize that a human being is not an AGI.
because a human being, yes, there is definitely a foundation of skills,
a human being, a human being lacks a huge amount of knowledge.
Instead, we rely on continual learning.
We rely on continual learning.
And so then when you think about, okay, so let's suppose that we achieve success
and we produce some kind of safe super intelligence,
the question is, but how do you define it?
where on the curve of continual learning
is it going to be. I will produce
like a super intelligent
15 year old that's very eager to go
and you say, okay, I'm going to, they don't know very much
at all. The great student, very eager.
You go and be a programmer.
You go and be a doctor.
Go and learn. So you could imagine that
the deployment itself will involve some kind of a
learning trial and error period.
It's a process
as opposed to you drop the
finished thing. Okay. I
I see.
So you're suggesting that the thing you're pointing out with superintelligence
is not some finished mind which knows how to do every single job in the economy.
Because the way, say, the original, I think, Open AI Charter or whatever defines AGI
is like it can do every single job that every single thing a human can do.
You're proposing instead a mind which can learn to do every single job.
Yes.
And that is superintelligence.
And then, but once you have the learning algorithm, it gets deployed into the world
the same way a human laborer might join an organization.
And it seems like one of these two things might happen.
Maybe neither of these happens.
One, this super efficient learning algorithm becomes superhuman, becomes as good as you
and potentially even better at the task of ML research.
and as a result, the algorithm itself
becomes more and more superhuman.
The other is, even if that doesn't happen,
if you have a single model,
I mean, this is explicitly your vision,
if you have a single model
where instances of a model
which are deployed through the economy,
doing different jobs, learning how to do those jobs,
continually learning on the job,
picking up all the skills that any human could pick up,
but actually picking them all up at the same time
and then amalgamating the learnings,
you basically have a model
which functionally becomes super intelligent
even without any sort of recursive self-improve
in software, right?
Because you now have one model
that can do every single job in the economy
and humans can't merge our minds
in the same way.
And so do you expect some sort of
intelligence explosion from broad deployment?
I think that it is likely
that we will have rapid economic growth.
I think the broad deployment,
like there are two arguments
you could make, which are conflicting.
One is that, look, if indeed you get, once indeed you get to a point where you have an
AI that can learn to do things quickly, and you have many of them, then they will, then
there will be a strong force to deploy them in the economy unless there will be some kind
of a regulation that stops it, which by the way they might be.
But I think the idea of very rapid economic growth for some time,
I think it's very possible from broad deployment.
The other question is how rapid it's going to be.
So I think this is hard to know because on the one hand,
you have this very efficient worker.
On the other hand, the world is just really big and there's a lot of stuff.
And that stuff moves at a different speed.
But then on the other hand, now the AI could, you know,
so I think very rapid.
economic growth is possible. And we will see
like all kinds of things like
different countries with different rules
and the ones which have the friendlier rules
the economic growth will be faster.
Hard to predict. It seems to me that this is
a very precarious situation
to be in where
look in the limit, we know that this
should be possible because if you have
something that is as good as a human
at learning, but which can merge
its brains, merge
there are different instances in a way that humans
can't merge. Already, this
seems like a thing that should physically be possible. Humans are possible. Digital computers are
possible. You just need both of those combined to produce this thing. And it also seems like this
kind of thing is extremely powerful. And economic growth is one way to put it. I mean,
Dyson Spears is a lot of economic growth. But another way to put it is just like you will have
potentially a very short period of time. Because a human on the job can, you know, you're hiring
people to SSI in six months. They're like net productive probably, right? A human, like,
It learns really fast.
And so this thing is becoming smarter and smarter very fast.
How do you think about making that go well?
And why is SSI position to do that well?
Or what is SSI's plan there, basically, is what I'm trying to ask.
Yeah.
So one of the ways in which my thinking has been changing is that I now place more importance on AI being deployed.
incrementally and in advance.
One very difficult thing about AI
is that we are talking about systems
that don't yet exist.
And it's hard to imagine them.
I think that one of the things that's happening
is that in practice,
it's very hard to feel the AGI.
It's very hard to feel the AGI.
We can talk about it.
But it's like, it's like talking about like the long future, like imagine like having a conversation about like, how is it like to be old when you're like old and frail and you can have a conversation, you can try to imagine it, but it's just hard and you come back to reality where that's not the case.
And I think that a lot of the issues around AGI and its future power stem from the fact.
that it's very difficult to imagine.
Future AI is going to be different.
It's going to be powerful.
Indeed, the whole problem, what is the problem of AI and AIGI?
The whole problem is the power.
The whole problem is the power.
When the power is really big, what's going to happen?
And one of the ways in which I've changed my mind over the past year
And so that change of mind may back, may, I'll say, I'll hedge a little bit,
may back propagate into the plans of our company, is that,
so if it's hard to imagine, what do you do?
You've got to be showing the thing.
You've got to be showing the thing.
And I maintain that, I think most people who work on AI also can't imagine it.
because it's too different from what people see on a day-to-day basis.
I do maintain, here's something which I predict will happen.
That's a prediction.
I maintain that as AI becomes more powerful
than people will change their behaviors.
And we will see all kinds of unprecedented things
which are not happening right now.
and I'll give some examples
I do like
I think for better or worse
the frontier companies
will play a very important role
in what happens
as will the government
and the kind of things that I think
we'll see
which you see the beginnings of
companies that are
fierce competitors starting
to collaborate
on AI safety
you may have seen
open AI and anthropic
doing a first small
step but that did not exist that's actually something which I predicted in one of my
talks about three years ago that such a thing will happen I also maintain it as
AI continues to become more powerful more visibly powerful there will also be a
desire from governments and the public to do something and I think that this is a
very important force of showing the AI that's number one number two okay so
So then the AI is being built.
What needs to be done?
So one thing that I maintain that will happen
is that right now, people who are working on AI,
I maintain that the AI doesn't feel powerful
because of its mistakes.
I do think that at some point the AI will start to feel powerful, actually.
And I think when that happens,
we will see a big change in the way all AI companies
approach safety.
They'll become much more.
paranoid. I say this as a prediction that we will see happen. We'll see if I'm right. But I think
this is something that will happen because they will see the AI becoming more powerful. Everything
that's happening right now I maintain is because people look at today's AI and it's hard to
imagine the future AI. And there is a third thing which needs to happen. And I think this is
this, and I'm talking about it in broader terms, not just from the perspective.
type of SSI
because you ask me about our company
but the question is okay so then what should
what should the companies aspire to build
what should they aspire to build
and there has been one big idea
that actually everyone has been locked in
locked into which is the
self-improving AI
and why did this happen
because there is fewer
ideas than companies
but I maintain that there is something
that's better to build
and I think that everyone will actually want that
it's like the AI
that's robustly aligned
to care about sentient life specifically
I think in particular
there's a case to be made
that it will be easier
to build an AI that cares about sentient life
than an AI that cares about human life alone
because the AI itself will be sentient
and if you think about things like mirror neurons
and human empathy for animals
which is you know you might argue
it's not big enough, but it exists.
I think it's an emergent property from the fact that we model others
with the same circuit that we used to model ourselves
because that's the most efficient thing to do.
So even if you got an AI to hear about sentient beings,
and it's not actually clear to me that that's what you should try to do
if you solve the alignment,
it would still be the case that most sentient beings will be AIs.
There will be trillions, eventually quadrillions of AIs.
humans will be a very small fraction of sentient beings.
So it's not clear to me if the goal is some kind of human control over this future civilization,
that this is the best criterion.
It's true.
I think that it's possible.
It's not the best criterion.
I'll say two things.
I think that thing number one,
I think that if there
So I think that care for sentient life
I think there is merit to it
I think it should be considered
I think that it will be helpful
if there was some kind of
a short list of ideas
that then
the companies when they are in the situation
could use that's number two
number three I think it would be really
materially helpful if
the power
of the most powerful superintelligence
was somehow capped
because it would address
a lot of these concerns.
The question of how to do it,
I'm not sure,
but I think that would be materially helpful
when you're talking about
really, really powerful systems.
Before we continue the 11th discussion,
I want to double click on that.
How much room is there at the top?
How do you think about superintelligence?
Do you think, I mean,
using this learning efficiency idea
maybe it's just extremely fast at learning new skills or new knowledge.
And does it just have a bigger pool of strategies?
Is there a single cohesive it in the center that's more powerful or bigger?
And if so, do you imagine that this will be sort of godlike
in comparison to the rest of human civilization,
or does it just feel like another agent or another cluster of agents?
So this is an area where different people have different intuitions.
I think it will be very powerful for sure.
I think that what I think is most likely to happen is that there will be multiple such AIs being created roughly at the same time.
I think that if the cluster is big enough, like if the cluster is literally continent-sized, that thing could be really powerful indeed.
If you literally have a continent-sized cluster, those AIs can be very powerful.
And all I can tell you is that if you're talking about extremely powerful AI
is like truly dramatically powerful, then yeah, it would be nice if they could be restrained
in some ways or if there was some kind of an agreement or something.
Because I think that if you are saying, hey, like if you really, like what is the concern
of superintelligence?
What is one way to explain the concern?
If you imagine a system that is sufficiently powerful,
like really sufficiently powerful,
and you could say, okay, you need to do something sensible,
like care for sentient life, let's say,
in a very single-minded way.
We might not like the results.
That's really what it is.
And so maybe, by the way, the answer is that you do not build an REL agent
in the usual sense.
And actually, I'll point several things out.
I think human beings are a semi-a-rel agent,
you know, we pursue a reward
and then the emotions or whatever
make us tire out of the reward,
we pursue a different reward.
The market is like
kind, it's like a very short-sighted
kind of agent. Evolution is the same.
Evolution is very intelligent in some ways,
but very dumb in other ways.
The government has been designed
to be a never-ending fight
between three parts,
which has an effect.
So I think things like this.
Another thing that makes this
discussion difficult is that we are talking
about systems that don't exist
that we don't know how to build
right that's the other thing
and that's actually my belief I think what people are doing
right now will go some distance
and then peter out
it will continue to improve but it will also
not be it so the it
we don't know how to build
and I think that a lot
hinges on
understanding
reliable generalization
now say another thing
which is like, you know, one of the things that you could say
that cause alignment to be difficult is that human value,
that it's, it's, um, your ability to learn human values is fragile.
Then your ability to optimize them is fragile.
You actually learn to optimize them.
And then can't you say, are these not all instances of unreliable generalization?
Why is it that human beings appear to generalize so much better?
What generalization was much better?
What would happen in this case?
what would be the effect.
But those we can't,
we can't,
like those questions are
right now still answerable.
How does one think about
what AI going well
looks like?
Because I think you've scoped out
how AI might evolve
will have these sort of
continual learning agents.
AI will be very powerful.
Maybe there will be many different AIs.
How do you think about
lots of continent
compute size intelligences
going around?
How dangerous is that?
How do you make
that less dangerous and how do we do that in a way that protects a equilibrium where there
might be misaligned AIs out there and bad actors out there. So one reason why I liked the AI
that cares for sentient life, you know, and we can debate on whether it's good or bad. But
if the first N of these dramatic systems actually do care for, you know,
love humanity or something
you know careful sentient life obviously this also needs to be achieved
this needs to be achieved
so if this is achieved by the first end of those systems
then I can see it go well
at least for quite some time
and then there is the question of what happens in the long run
what happens in the long run how do you achieve
a long run equilibrium
and I think that there
there is an answer as well
and I don't like this answer
but
it needs to be considered
in the long run
you might say okay so if you have a world
where powerfully eyes exist
in the short term you could say
okay you have universal high
income
you have universal high income
and we all doing well
but we know that
what do the Buddhists say
change is the only constant
and so things change
and there is some kind of
government political structure thing
and it changes because
these things have a shelf life
you know some new government
thing comes up and it functions and then after
some time it stops functioning
that's something that you see
happening all the time and so
I think that for the long run equilibrium
one approach
you could say okay so maybe every person
will have an AI that will do their bidding
and that's good
and if that could be maintained indefinitely
Definitely. That's true. But the downside with that is, okay, so then the AI goes and like earns, you know, earns money for the person and, you know, advocates for their needs in like the political sphere. And maybe then writes a little report saying, okay, here's what I've done. Here's the situation. And the person says, great, keep it up. But the person is no longer a participant. And then you can say that's a precarious place to be in. But so I'm going to preface by saying,
I don't like this solution
but it is a solution
and the solution is
if people become part AI
with some kind of neuralink plus plus
because what will happen as a result
is that now the AI understands something
and we understand it too
because now the understanding is transmitted wholesale
so now if the AI is in some situation
now it's like
you are involved in that situation yourself fully
And I think this is the answer to the equilibrium.
I wonder if the fact that emotions, which were developed millions,
or in many cases billions of years ago, in a totally different environment,
are still guiding our actions so strongly,
is an example of alignment success.
To maybe spell out what I mean,
the brainstem has these...
I don't know if it's more accurate to call it a value function or reward function,
but the brainstem has a directive of it saying mate with somebody who's more successful.
The cortex is the part that understands, what does success mean in the modern context?
But the brainstem is able to align the cortex and say,
however you recognize success to be, and I'm not smarter enough to understand what that is,
you're still going to pursue this directive.
I think there is, so I think there's a more general point.
I think it's actually really mysterious how
the brain
encodes high level desires
sorry how evolution encodes high level
desires like it's pretty easy
to understand how
evolution would endow us with
the desire for food that
smells good because smell is a chemical
and so just
pursue that chemical it's very easy to imagine such
evolution doing such a thing
but
evolution also has endowed
us with all these social desires
like we really
We care about being seen positively by society.
We care about being in a good standing.
Like all these social intuitions that we have,
I feel strongly that they're baked in.
And I don't know how evolution did it.
Because it's a high-level concept that's represented in the brain.
Like what people think, like, let's say you are like,
you care about some social thing.
It's not like a low-level signal.
smell. It's not something that, for which there is a sensor. Like the brain needs to do a lot of
processing to piece together lots of bits of information to understand what's going on socially
and somehow evolution said that's what you should care about. How did it do it? And he did it quickly
too. Yeah. Because I think all these sophisticated social things that we care about, I think they
evolved pretty recently. So evolution had an easy time hardcoding this high level desire and
I maintain or at least I'll say
I'm unaware of good hypothesis
for how it's done. I had some
ideas I was kicking around but none of them
are satisfying. Yeah and
what's especially impressive is it was a desire that you
learned in your lifetime. It kind of makes sense because
your brain is intelligent. It makes sense why we're able to learn
intelligent desires but your point is that the desire
is maybe this is not your point but one way to
understand it is, the desire is built into the genome, and the genome is not intelligent,
right? But it's able to, you're somehow able to describe this feature that requires, like,
it's not even clear how you define that feature, and you can get it into the, you can build it
into the genes. Yeah, essentially. Or maybe I'll put it differently. If you think about the tools
that are available to the genome, it says, okay, here's a recipe for building a brain,
and you could say, here is a recipe for connecting the dopamine neurons to, like, the smell sensor.
and if the smell is a certain kind of
good smell you want to eat that
I could imagine the genome doing that
I'm claiming that it is harder to imagine
it's harder to imagine the genome
saying you should care about
some complicated computation
that your entire brain that like a big chunk of your brain does
that's all I'm claiming I can tell you
like a speculation I was wondering how it could be done
and let me offer a speculation
and I'll explain why the speculation is probably false
So the speculation is
Okay
So the brain
It's like
The brain has those regions
You know the brain regions
We have our cortex right
Yeah
And has all those brain regions
And the cortex is uniform
But the brain regions
And the neurons in the cortex
They kind of speak to their neighbors mostly
And that explains why you get brain regions
Because if you want to do some kind of speech processing
All the neurons that do speech
need to talk to each other
and because neurons can only speak
to their nearby neighbors
for the most part
it has to be a region
all the regions
are mostly located
in the same place
from person to person
so maybe evolution
hard-coded
literally a location on the brain
so it says
oh like when like
the GPS of the brain
GPS coordinates such and such
when that fires
that's what you should care about
like maybe that's what evolution did
because that would be
within the toolkit of evolution
yeah although there are examples
where, for example, people who are born blind
have that area of their cortex
adopted by another sense.
And I have no idea, but I'd be surprised
if the desires or the reward functions
which require visual signal no longer worked,
you know, people who have their different areas
of their cortex co-opt-out.
For example, if you no longer have vision,
can you still feel the sense
that I want people around me to like me and so forth,
which usually there's also visual cues for.
So I actually fully agree with that.
I think there's an even stronger counter argument of this theory,
which is, like, if you think about people,
so there are people who get half of their brains removed in childhood.
And they still have all their brain regions,
but they all somehow move to just one hemisphere,
which suggests that the brain regions, the location is not fixed.
And so that theory is not true.
It would have been cool if it was true.
but it's not.
And so I think that's a mystery,
but it's an interesting mystery.
Like the fact is,
somehow evolution was able to endow us
to care about social stuff very, very reliably.
And even people who have like all kinds of strange
mental conditions and deficiencies
and emotional problems tend to care about this also.
What is SSI planning on doing differently?
So presumably your plan is to be one of the frontier companies
when this time arrives.
And then what is,
presumably you started SSI
because you're like
I think I have a way of approaching
how to do this safely
in a way that the other companies don't
what is that difference
so the way I would describe it as
there are some ideas
that I think are promising
and I want to investigate them
and see if they are indeed promising or not
it's really that simple
it's an attempt
I think that if the ideas are now to be correct
these ideas that we discussed
around
standing generalization.
If these ideas turn out to be correct,
then I think we will have something worthy.
Will it turn out to be correct?
We are doing research.
We are squarely age of research company.
We are making progress.
We've actually made quite good progress over the past year,
but we need to keep making more progress, more research.
And that's how I see it.
I see it as an attempt to be...
an attempt to be a voice and a participant.
People have asked your co-founder and previous CEO left to go to META recently,
and people have asked, well, if there was a lot of breakthroughs being made,
that seems like a thing that should have been unlikely.
I wonder how you respond.
Yeah, so for this, I will simply remind a few facts that may have been forgotten.
and I think these facts which provide the context,
I think they explained the situation.
So the context was that we were fundraising
at a $32 billion valuation
and then Meta came in
and offered to acquire us
and I said no,
but my former co-founder
like in some sense said yes.
And as a result,
he also was able to enjoy
from a lot of near-term liquidity, and he was the only person from SSI to join meta.
It sounds like SSI's plan is to be a company that is at the frontier when you get to this
very important period in human history where you have superhuman intelligence and you have
these ideas about how to make superhuman intelligence go well, but other companies will be
trying their own ideas. What distinguishes SSI's approach to making superintelligence go well?
the main thing that distinguishes SSI is its technical approach.
So we have a different technical approach that I think is worthy
and we are pursuing it.
I maintain that in the end there will be a convergence of strategies.
So I think there will be a convergence of strategies where at some point
as AI becomes more powerful,
it's going to become more or less clearer to everyone,
the strategy should be.
And it should be something like, yeah,
you need to find some way to talk to each other
and you want your first
actual, like real super intelligent AI
to be aligned and somehow be
you know, care for sentient life,
careful people, democratic,
one of those, some combination of thereof.
And I think
this is the condition
that everyone should strive for
and that's what the society is striving for
and I think that this time
if not already all the other companies
will realize that they're striving towards the same thing
and we'll see I think that the world will truly change
as the air becomes more powerful
and I think a lot of these forecasts will
like I think things will be really different
and people will be acting really differently
what speaking of forecast
what are your forecast to this system
term you're describing which can learn as well as a human and subsequently as a result becomes superhuman
I think like five to 20 five to 20 years so I just want to unroll your how you might see the world coming
it's like we have a couple more years where these other companies are continuing the current approach and it stalls out
and stalls out here meaning they earn no more than low hundreds of billions in revenue or how do you think about what's
Staling out means.
Yeah.
I think it could stole out
and I think stolen out
will look like
it will all look very similar
among all the different companies,
something like this.
I'm not sure because I think
I think even with
I think even
I think even we stolen out
I think these companies could make
a stupendous
stupendous revenue.
Maybe not profits because
they will be
they will need to work hard
to differentiate each other
from themselves, but revenue definitely.
But there's something in your model implies that
when the correct solution does emerge,
there will be convergence between all the companies.
And I'm curious why you think that's the case.
Well, I was talking more about convergence
on their largest strategies.
I think eventual convergence on the technical approach
is probably going to happen as well.
But I was alluding to convergence
to the largest strategies.
What exactly is the thing that should be done?
I just want to better understand
how you see the future enrolling.
So currently we have these different companies
and you expect their approach
to continue generating revenue
but not get to this human-like learner.
So now we have these different forks of companies.
We have you, we have thinking machines,
there's a bunch of other labs.
Yes.
And maybe one of them figures out the correct approach.
But then the release of their product
makes it clear to other people how to do this thing.
I think it won't be clear how to do its thing,
but it can be clear that something different is possible.
Right.
And that is information.
and I think people will then be trying to figure out how that's how that works
I do think though that one of the things that's that I think you know not addressed here
not discussed is that with each increase in the AI's capabilities I think there will be
some kind of changes but I don't know exactly which ones in how things are being done
and so like I think it's going to be
important, yet I can't spell out
what that is exactly.
And how are the,
by default, you would expect the
model company that has, the model company that has that
model to be getting all these gains because they
have the model that is learning how to do all,
has the skills and knowledge that it's
building up in the world.
What is the reason to think that the benefits of that
would be widely distributed and not just
end up at whatever model company
gets this continuous learning loop going
first? Like, I think
that empirically what happened, so
here is what I think is going to happen
number one
I think empirically
when
let's look at
let's look at how things have gone so far
with the AIs of the past
so one company produced an advance
and the other company scrambled
and produced some similar things
after some amount of time
and they started to compete in the market
and push the prices down.
And so I think from the market perspective,
I think something similar will happen there as well.
Even if someone, it's okay,
we are talking about the good world, by the way,
where what's the good world?
What's the good world?
Where we have these powerful human-like learners
that are also like, and by the way,
maybe there's another thing we haven't discussed
on the spec of the super intelligent AI
that I think is worth considering
is that you make it narrow
can be useful and narrow at the same time
so you can have lots of narrow super intelligent AIs
but suppose you have many of them
and you have some company
that's producing a lot of profits from it
and then you have another company that comes in
and starts to compete
and the competition is going to work
is through specialization.
I think what's going to happen is that the way competition,
like competition loves specialization.
And you see it in the market,
you see it in evolution as well.
So you're going to have lots of different niches.
And you're going to have lots of different companies
who are occupying different niches in this kind of world.
But you might say, yeah, like one AI company
is really quite a bit better at some area of really complicated
economic activity
and a different company
is better at another area
and the third company
is really good at litigation
and that's the way
is contradicted by what
human like learning
implies is that like
it can learn
it can but
but you have accumulated learning
you have a big investment
you spent a lot of compute
to become really really
really good
really phenomenal at this same
and someone else spent
a huge amount of computer
and a huge amount of experience
to get really really good
at some other thing
right you apply a lot
of human learning to get there
but now like you are you are
at this high point
where someone else would say look like
I don't want to start learning what you've learned
to go through this.
That would require many different companies
to begin at the human
like continual learning agent at the same time
so that they can start their different research
in different branches.
But if one company
gets that agent first
or gets that learner first,
it does then seem like
well, you know,
like we just don't
just thinking about every single job in the economy,
you just have instance learning each one seems tractable for a company.
Yeah, that's a valid argument.
My strong intuition is that it's not how it's going to go.
My strong intuition is that, yeah, like the argument says it will go this way,
but my strong intuition is that it will not go this way.
That this is the, you know, in theory,
there is no difference between theory in practice and practice there is.
And I think that's going to be one of those.
A lot of people's models of recursive self-improvement
literally explicitly state
we will have a million Ilias in a server
that are coming in with different ideas
and this will lead to a superintelligence emerging very fast.
Do you have some intuition about how parallelizable
the thing you are doing is?
What are the gains from making copies of Ilya?
I don't know.
I think there will definitely be diminishing returns
because you want people who think differently
rather than the same.
I think that if they were little copies of me,
I'm not sure how much more incremental value you'd get.
I think that,
but people who think differently,
that's what you want.
Why is it that it's been,
if you look at different models,
even released by totally different companies,
trained on potentially non-overlapping datasets,
it's actually crazy how similar LLMs are to each other?
Maybe the datasets are not as non-overlapping.
it seems.
But there's some sense
that's like even if an individual human
might be less productive than the future AI,
maybe there's something to the fact that human teams
have more diversity than teams of AI's might have.
But how do we elicit meaningful diversity
among AI?
So I think just raising the temperature
just results in gibberish.
I think you want something more like
different scientists of different prejudices
or different ideas.
How do you get that kind of diversity
among AI agents?
So the reason there has been no diversity
I believe is because of pre-training.
All the pre-trained models are the same, pretty much,
because they're pre-trained on the same data.
Now, RRL and post-training is where some differentiation starts to emerge
because different people come up with different RL training.
Yeah.
And then I've heard you hint in the past about self-play as a way to either get data
or match agents to other agents of equivalent intelligence
to kick-off learning,
how should we think about
why there's no public
proposals of this kind of thinking
working with LLMs?
I would say there are two things to say.
I would say that the reason why I thought
self-played was interesting
is because it offered a way
to create models
using compute only without data.
Right? And if you think that data
is the ultimate bottleneck,
then using compute only is very interesting.
so that's what makes it interesting now the the thing is that self-play at least the way it was done in the past when you have agents which are somehow compete with each other it's only good for developing a certain set of skills it is too narrow it's only good for like negotiation conflict certain social skills strategizing that kind of stuff and so if you care about this
those skills, then self-play
will be useful. Now, actually,
I think that self-play did
find a home, but just in a different
form, in a different form.
So things like debate,
prove a verifier,
you have some kind of
an LLM as a judge, which is
also incentivized to find mistakes in your work.
You could say this is not exactly self-play,
but this is, you know, a related
adversarial setup that people are doing, I believe.
And really self-play is an
example of, is a special case of more general, like, competition between agents, right?
The response, the natural response to competition is to try to be different. And so if you
were to put multiple agents and you tell them, you know, you all need to work on some problem,
and you're an agent and you're inspecting what everyone else is working, you're going to say,
well, if they're already taking this approach, it's not clear I should pursue it. I should pursue
something differentiated. And so I think that something like this could also create an incentive
for a diversity of approaches. Yeah. Final question. What is research taste? You're obviously
the person in the world who is considered to have the best taste in doing research in AI.
You were the co-author on many of the biggest,
the biggest things that have happened in the history of deep learning
from Alex Net to GP3 to so on.
What is it that, how do you characterize how you come up with these ideas?
I can answer.
So I can comment on this for myself.
I think different people do it differently.
But one thing that guides me personally is an analysis.
aesthetic of how AI should be by thinking about how people are.
But thinking correctly, like, it's very easy to think about how people are incorrectly.
But what does it mean to think about people correctly?
So I'll give you some examples.
The idea of the artificial neuron is directly inspired by the brain, and it's a great idea.
Why? Because you say, sure, the brain has all these different organs.
It has the folds, but the falls probably don't matter.
Why do we think that the neurons matter?
Because there is many of them.
It kind of feels right, so you want the neuron.
You want some kind of local learning rule
that will change the connections.
You want some local learning rule
that will change the connections between the neurons.
Right?
It feels plausible that the brain does it.
The idea of the distributed representation.
The idea that the brain,
you know, the brain responds to experience
on neural that should learn from experience,
not response.
The brain learns from experience.
the neural network experience
and you kind of ask yourself
is something fundamental or not fundamental
how things should be
and I think that's been guiding me a fair bit
kind of thinking from multiple angles
and looking for almost beauty
beauty simplicity
ugliness there is no room for ugliness
it's just beauty simplicity
elegance correct inspiration from the brain
and all of those things need to be present at the same time
and the more they are present
the more confident you can be in a top-down belief
and then the top-down belief
is the thing that sustains you
when the experiments contradict you
because if you just trust the data all the time
well sometimes you can be doing a correct thing
but there's a bug
but you don't know that there is a bug
how can you tell that there is a bug
how do you know if you should keep debugging
or you conclude it's the wrong direction
well is the top-down
well how should you can say the things have to be this way
something like this has to work
therefore we got to keep going
that's the top down.
And it's based on this like multifaceted beauty
and inspiration by the brain.
All right.
We'll leave it there.
Thank you so much.
Thank you so much.
Oh, all right.
Appreciate it.
That was great.
Yeah.
I enjoyed it.
Yes, me too.
Thanks for listening to this episode of the A60Z podcast.
If you like this episode, be sure to like, comment, subscribe,
leave us a rating or review, and share it with your friends and family.
Or more episodes, go to YouTube, Apple Podcasts, and Spotify.
Follow us on X at A16Z and subscribe to our Substack at A16Z.com.
Thanks again for listening, and I'll see you in the next episode.
This information is for educational purposes only and is not a recommendation to buy, hold, or sell any investment or financial product.
This podcast has been produced by a third party and may include pay promotional advertisements, other company references, and individuals unaffiliated with A16Z.
Such advertisements, companies, and individuals are not endorsed by AH Capital Management LLC, A16Z, or any of its affiliates.
Information is from sources deemed reliable on the date of publication, but A16Z does not guarantee its accuracy.
