a16z Podcast - a16z Podcast: Revenge of the Algorithms (Over Data)... Go! No?
Episode Date: October 23, 2017with Frank Chen, Steven Sinofsky, and Sonal Chokshi There are many reasons why we’re in an “A.I. spring” after multiple “A.I. winters” — but how then do we tease apart what’s real vs. wh...at’s hype when it comes to the (legitimate!) excitement about artificial intelligence and machine learning? Especially when it comes to the latest results of computers beating games, which not only captures our imaginations but has always played a critical role in advancing machine intelligence (whether it’s AI winning Texas Hold’em poker or beating the world human champ in the ancient Chinese game of Go). But on learning that Google DeepMind’s AlphaGo can master the game of Go without human knowledge — or more precisely: “based solely on reinforcement learning, without human data, guidance, or domain knowledge beyond game rules” — some people leap too far towards claims of artificial generalized intelligence. So where can we then generalize the findings of such work — unsupervised learning, self-play, etc. — to other specific domains? What does it mean for entrepreneurs building companies (and what investors look for)? And what does it mean for how we, as humans, learn… or rather, how computers can also learn from how we learn? Deal and research operating team head Frank Chen and a16z board partner Steven Sinofsky ponder all this and more, in conversation with Sonal Chokshi, in this episode of the a16z Podcast. We ended last time with the triumph of data over algorithms and begin this time with the triumph of algorithms over data … is this the end of big data?
Transcript
Discussion (0)
Hi, everyone. Welcome to the A6 and Z podcast. I'm Sonal, and I'm here today bringing back the band together again, the AlphaGo band, I guess. I don't know how else to describe us, but we have the head of our deal in research team operation, Frank Chen. We have Steven Sinovsky, A6 and Z board partner. Just to give people some quick context, you don't have to have heard our previous podcast on AlphaGo. But when AlphaGo, the algorithm produced by Deep Mind, beat the world champ in Korea in playing Go, which is an ancient Chinese game, we had a
podcast where we discussed a lot of the themes and some of the broad things around that.
And what we'd like to talk about today is their latest paper, but not just only specifically,
but more broadly, what this means for where we actually are, what's hype, what's real in
AI or artificial intelligence. So welcome, guys. Welcome. So, like, one thing that, like,
you read this paper, and the paper's published in nature, it's pretty dense. It has 17 authors
on it. And so it's quite the force. But the thing that sort of jumps out is the paper, the
blogs that everybody is the first thing you read is one step closer to creating general purpose
AI and immediately like my AI winter fears of hype antenna pop up because like everything is one
step closer but like you could take very very very tiny little steps or you could overhype
them and we've heard these promises throughout history and especially around board games so we
solved checkers and people were like oh one step away from general intelligence then we solved
chess. Same thing. We're one step away from general intelligence. Then we did the, oh, I can find
bacteria that causes infection. Therefore, we must be one step away. And I think the fallacy is that
because we're doing these things that are considered sort of high cognition, like there's smart people
play chess, smart people figure it out. Then solving one thing that a smart person does must lead us
to the next thing that a smart person does, which will lead us to the next. And that's always been
the fallacy, which is it actually hasn't quite generalized. Well, in fact,
In fact, they don't even really compound all that much.
They're all fairly discreet.
The one thing that's different now is that they are all building on these new artificial
intelligence or machine learning techniques and taking a whole bunch of data, training
a model on it, and developing solutions that beat all their algorithmic ones.
And that was a big thing about the first alpha-go was like, this is data over algorithms.
And all of a sudden, here we are-
algorithms over data or or a model over data but you know frank you have an interesting way of looking
at this because you know you play go and you understand that you play go i didn't know that enough
more than me i only play chess i don't even know chinese chess and go all right yeah but but like
it's it's not like a generalized problem you know it comes with like a whole bunch of constraints
and things that make it solvable by algorithm yeah so what are some traits
about Go that are not like the real
world. All the rules are completely
well known. The state of
play is completely well known, right? In the
real world, mostly we live in a fog of war
situation where we know some things and we don't know other
things. It would cost us something to go figure out
and Go, and in particular
the representations they chose, the entire board state
is known to both players, right? So there's a lot
of things that aren't like the real world
at all when we do these board games. Now,
having said that, why are people
getting breathless yet again around
one step closer to AGI? I said,
Authors don't claim generalized intelligence. They're talking about being able to do apply some of these techniques in other domains.
Yeah. So you give credit where credits do. Let's sort of review what they actually achieved. And they are like very impressive achievements. So what they achieved was they now have a go player that beat all of their other go players. AlphaGo zero. Right. Well, they have they named them by the people they beat. So they had AlphaGo fan, Alpha Go Lee. This one's AlphaGo Zero. So AlphaGo Zero, their latest one, has beat every iteration, which has beat all of the best human players.
So they have the best of go-playing algorithms in the world.
And the difference in how they got to this one is there was no training data from human
games.
So all the other ones had been bootstrapped by, let me go watch what the human players do and
see if I can mimic that.
And that very broadly speaking is the approach to machine learning called supervised learning.
And not a trivial amount.
I mean, they had 100,000 games that they started from to train it on like 48 of
of Google's TPUs for days on end.
That's exactly right.
So a fleet of machines, a ton of data.
This one started from nothing, which is it didn't even really know the rules.
It had a loss function, which is reinforcement learning's way of sort of improving the algorithm
is over time.
So let me just give a quick intuitive explanation of reinforcement learning.
So it's exactly like the game that used to play when you were a kid called hotter or colder.
Somebody would go hide something.
And then you would try to get to discover it.
Are you getting hotter?
I'm getting hotter.
Hotter, cold, cold again, right.
And then the hotter or colder is the loss function.
That is the thing that tells you if you're getting closer or not.
It's basically trial and error.
That's exactly right.
And so that's the fundamental approach in AlphaGo Zero, which is they have this loss function
that describes, are you more likely to win the game, having made this move or not?
Right.
So that's all they had.
And then they didn't have human input.
They didn't have human games.
They basically said, I have this locks function that tells you whether you're more likely
to win this game or not.
And then it played itself.
One other important thing I think it has, it also knows.
it has the codified rules of the game.
That is the one human input it actually got.
Right.
And it makes it a very, very constrained problem
because there's a whole bunch of the decision trees
like hot or cold.
You could climb up the sofa.
You could leave the house.
You could go all over the place for days on end.
Whereas this just tells you know
it can only be on the floor.
These are the rules.
It's not going to be under the sofa.
It's a 17 by several.
Very little of real life is actually like you get the rules told you that way.
It's black and white pieces.
There are rules.
Yeah, you switch turns, like all of that stuff matters a lot.
It does.
And one more thing to add, by the way, too, because you mentioned the 48 TPUs, it's really significant that they got it down to four in this case.
Yeah.
That's a huge, like, on the power side, energy, like, you know, it's a simple architecture.
It's a simple architecture.
And thinking from the point of view of a software developer, now you're back to one machine.
You're not like, oh, my God, I need to go rent this massive cloud with massive storage and massive interconnects.
and like I need to figure out how to provision the cluster and manage the cluster, you're
back down to one machine.
This stuff was pretty impressive.
It's, you know, they did four TPUs.
It was three days of playing.
Three days total.
Like not, like it's very achievable, you know, like on your Amazon credits.
That's exactly right.
And if you think about sort of the approach that most startups take to artificial intelligence
to TAY, they basically take the supervised learning approach.
And step one, raise money so that you can go get a data.
set that's annotated, train your neural network, make recommendations, right? And you could be
one or five or ten or fifty million dollars in getting that data set depending on how complicated
the data set is. Right. So the reason people are so excited about this is, look, this had no
data aside from rules of the game. It basically played itself. And by day three, it was better
than everything that had been trained before. Nearly an order of magnitude. Yeah. It was
doubt the order's a magnitude of all of this.
It was 408 TPUs versus four.
It was three days versus 40.
It was 30 million trained games versus 4.9.
So order of magnitude improvement on all of those dimensions.
And so like let's give credit where credit is due.
This is a very impressive technical achievement.
And then the question that we sort of entered the session with, okay, does this make us
more likely to be able to create an artificial general intelligence where the learning
algorithms generalize across domains?
In other words, can I take the breakthroughs here and make a better pick-and-pack robot for Amazon
or make a better health care predictor to discover whether you have cancer?
Or Salesforce forecasting, code generation, or a whole bunch of stuff.
And the interesting thing is that also the techniques here, this is just proof of something
that people have been talking about for a long time.
Oh, yeah.
Reinforcement learning's been around for ages.
And so part of what's interesting to me is that reinforcement learning has been around for a long time.
Obviously, the modeling has been around for a long time.
Self play.
All of these things.
And this, like so many times these steps are like somebody new to the domain looking and pulling together a bunch of unrelated things.
And just coming up with a very elegant, incredibly elegant solution.
And it's super impressive.
But it's not clear it generalizes.
And I think that was one of the things that jumped out for me in reading about it, you know, when you hear, oh, and then the next step on this train of generalized television is drug discovery.
protein folding and quantum chemistry and material and it's like all of a sudden I'm trying to
figure out protein folding we like what are the constraints on protein folding well we know
there are amino acids and we know that they have to be in three dimensions but actually nobody
else know that there's no codified rules like nobody has the rules of protein folding so there's
there's no rules there's no perfect understanding of the search space there's like what's the loss
function like why would you how would you even write a loss function I do want to push back a little bit though
because far be it from any of us in here to hype this up.
But there's something unique happening here,
which at least I perceive this in the paper,
which is that I was struck by the analogy to evolution.
Like, this is how human beings have evolved.
This is evolution that we learn by trial and error on a massive million scale.
So I don't want to completely dismiss the idea
that we can get to some kind of generalized intelligence.
I mean, of course, I understand and agree,
but what are the limits and what are the possibilities that can actually take us there
and where are we constrained?
Just to break that down a bit more.
Yeah. So I love the evolution analogy, right? Because in the paper, they talk about how, you know, when it started out, it was making sort of naive moves. It was sort of greedy and trying to capture all the tokens. And then it got to very sophisticated patterns that humans have discovered over thousands of years playing this game, teaching this game, codifying it in books. And it figured that out. And not only did it figure out, it figured out things that humans haven't quite codified. Right. So if you play more games, right?
sort of, it developed an intelligence that humans haven't yet because it's going through
its own reinforcement thing and thousands of generations of games playing each other, sort
of arrive at places that maybe humans would have gotten to if we played another
thousand years, but like, you know, it figured out in three days. So I think there is something
incredibly profound going on here where you're basically, you're accelerating natural selection
cycles in computers. Now, where I think the analogy breaks down is, oh, and therefore we can
apply this approach to every other problem and it's just going to be a straightforward application
of these set of ideas to those problem sets and then we're going to have evolution at that
scale. In other words, eating different problems right in exactly the same way. And I think that's
where Steve and I have a little skepticism. Yeah, well, and I think also like the key with all of
these, if you just take an abstract view of it, you know, you have this awesomely elegant solution
to an intractable problem, which is just for on any measure super cool. But that doesn't make
generalizable to any other space. There are many, many super cool things that surface and you have to be
careful as like an engineer or a founder or a person applying this to go, okay, well, what are the
elements of this solution that one would need to have as a precursor to applying it? And we saw the
same thing with all of the work on supervised learning. Like first you need a data set that is
clean and then it has to be like labeled really well and then you have to have a neural network
model and you have to do all the weights and all this model and so there were all these things
where you couldn't just say hey I'd like to make the most money in our Q4 sales let's machine
learn our way there and you're like you can't just show up on Thursday and yeah and do that every
year salespeople they develop a model for how they want to sell how the customers what the
prospects are they know the rules like they're all these rules like there are this many salespeople
They can only call so many people per day.
They know which customers to call.
They know what the quotas are going to look like.
They know what the product.
And you start to think, well, maybe there is something here because it's a space where, like,
actually the history might not be as well as applicable to supervise learning as you would like.
And so if there was a way to look at this through the lens of, like, what would be the optimal system to navigate
or what would be the sales matter?
And I think that there are more things like that than fewer.
I don't know how many, you know, properties of nature.
are amenable to this? Because for the most part, we don't understand them. Right. It's limited
by what the human knows, right? We can't codify those rules. What are the domains that you see
some transferability of these types of things? Well, I love the idea of sort of sales forecasting,
right? Because essentially what you're doing, the intuition is that if I could play act my salespeople
doing a million different things and a million different situations against a million different set
of prospects, and I could sort of simulate that, then maybe best practices would just come out of
that, right? Because essentially that's what I'm doing in real life.
I raise a ton of money. I hire a sales leader. That sales leader hires a bunch of people,
gives them a playbook, and has them call a bunch of prospects. I can only run so many experiments,
right? Because every call is an experiment. So the idea would be if I could simulate what happens
inside these calls and simulate 10, 100, 1,000, a million times more, then I'd get much better
best practices emerging from that. So super intriguing idea.
Off the top of my head, you start wanting to think about like, okay, what about, you know, cybersecurity?
And you think you're looking at your code and you know in any given moment, like the type of code that can be in a specific place.
And you know the rules and the syntax and it's well understood.
And you know patterns that are all so bad.
And so today what you do in the people that have tried to apply machine learning to code, they just have a lot of examples of like missing equal signs or missing semicolons or operators or.
wrong. But if you think about it, like, the syntax of the language is completely fixed. And so
you're back at very much a rule-based kind of, you know, rule-based is a constraint. These are
the only ways you can put the symbols together. What you don't have is right or wrong win or lose
with code. And you don't have a loss function that's that objective. There are interesting things to
me that like there's just complexity of line that you, that is a very common measure. Like, boy,
something with three sets of parentheses in it,
it's likely to have a bug in it
just because it has three sets of parentheses in it.
You know, something that's missing,
you know, some enclosure kind of rules
like using brackets, not using brackets.
These kind of things you can actually flag,
like you can think of Lint just flags them
in C and C++ as just sort of
high risk behaviors.
So there are things that you can put in
to like sort of, and so then you start to think,
wow, it would be really interesting
to have a tool that is able to look at code
and sort of just,
very different than previously, which was just literally looking at the syntax, but doing millions
of examples of generating code and finding bad examples, would it be able to do a better job
at finding bad examples than the next piece of code that flows into it?
I mean, one of the points you guys made last time in our last podcast is that at the end of the
day, these things aren't working in isolation. It's not like there's one magic approach.
You know, there's always a combination of techniques that come together to actually build real
products. How would this sort of fit into that? Because one of the thoughts that I had is that
clearly this kind of approach, even if you don't have clearly defined rules, will always
be more beneficial in places where we don't have any data, like any big data, just like
humans, like kids learning from n equals two, like their parents or n equals one, if there's
a single parent.
Yeah, I think if we think about what humans do, they have many different types of intelligences.
Yeah.
And we have many different types of strategies for solving problems.
So my guess is that the artificial intelligences that we create will be similar.
lots of different strategies and one of the interesting research items is which strategy should I employ to solve this problem? Because this is something the brain does effortlessly. The strategies that employ is to understand conversation are different than planning a trip than, you know, making sure you don't fall down versus like long range planning. Like how do I choose the best career? Right. So all of these are very different problems and somehow your brain kind of picks a good strategy for each one. Pedro Domingo talks about this as sort of, you know, in his book,
master algorithm, which is we know that there's all of these different techniques, but
kind of what we're missing is sort of the synthesizer, the thing that will know, what
strategy should I pick to solve this problem? And it's something that the brain seems to just
do. What's interesting about that observation is that's where we are with machine learning today
to begin with. Which is? Which is just like there were all of these different networks that you
can model with so many different layers and how many parameters you want to use. And there's
like this art to it right now. And I find that particular interesting because anytime there's
an art to something, there's an opportunity to start a company around that art and to build
out a product that surpasses the best practices in whatever field you're going after, whether it's
analyzing traffic or figuring out how to drive a car. And so that is the opportunity because I don't
think there's not like some path where in the next two years there's the meta algorithm that
knows which what way to pick like that's so so the the best thing to do now is to become well versed
and all of them and my gut just tells me that anytime you're at a point like this the the most
interesting solutions are actually going to end up being a hybrid you of like the thing that
used to work that everybody said doesn't work well enough plus the new thing that everybody
says is going to replace the old thing and that's basically been the entire history of AI
computing in general well computing for sure but AI specifically which is
Like, everybody always says the new thing, that's it.
We finally got it.
Huge step.
The old stuff is all done.
And the best example for me of that is how everybody said machine learning was going to replace all of natural language processing.
But if you dig into any of the work that's been going on, even the most state-of-the-art translation, which goes any language pair to any language pair, well, the input and the output all rely on the old school, like from the 1970s natural language stuff, just to do some very basic.
bookkeeping, it's very basic stuff.
You talk about like building a spell checker or whatever.
Parts of speech finder.
Right. And all the image, all the image stuff, like, wow, you know, if you want to figure
out features, it turns out when you start doing features of image recognition, you're using
a whole bunch of old school edge detection and contrast and finding objects and all of the
stuff.
And so you can't just like show up and say, now we're going to understand, we're going to be
the unsupervised learning company because the question everyone's going to ask is, well, how
are you going to make whatever you're doing practical?
What I do remember about hearing.
the stories of the early days of NLP and observing parts of this firsthand as well is how the
entire field and community, a lot of them had very strong opinions about, you know, there's this
whole phase of like expert domain knowledge building and really that's the only way to actually
make NLP work at scale. There was all these things they had to do because it was before the days
of big data. They couldn't even conceive of the Google scale big data. And then they went to this
world where, oh my God, we don't even have to have these kinds of constraints and way of doing
things because we have all this data. And now it's sexy to me that you can flip that model again
and almost say you don't even need big data with the results like this kind of paper because you don't
have to have anything. Yeah. This is the revenge of the algorithms moment, which is it's always been
about the data. If we had more data, we'd have more accurate models. And what this experiment
showed is, look, all I needed was the rules and a really good loss function, some very clever
programming and I get better performance than I had when I had the data set, right? So that's
the tantalizing, oh, look, you could do it without data. And like, look, as an investor, that'd be
awesome. If I could fund AI experiments without this step of collect data, annotate data, train
model, we could run a ton more experiments. Exactly. Because it's an age of abundance.
You bought up Pedro Domingo, and it was funny because one of the comments he made on the heels of
this announcement, which I thought was quite interesting, is like, well, you know, AlphaGo Zero learned after
like five million games. Humans took only so many thousands of years. And I'm sort of like,
that's the whole point of computers is to learn it in three days. You know, time is money. I don't
care if it took like, you know, five million games. It took three days. Time is more important than
this amount. You know, again, it's an amazing accomplishment. It's just, I always try to look at these
in the context of how these innovations tend to happen. You know, if you just look, the same thing
happened with search, like search itself. When people were inventing all the techniques of search,
they were working on these tiny machines with all of these physics constraints about how much they could compute.
And like, it was a whole thesis just to be able to go get like the archives of the New York Times to search through it.
And the same thing happened with spell checkers.
Like, wow, we'd love to have 50,000 words in the spell checker, but we don't have that much memory.
And then there was a whole, well, now we have a lot of memory.
And so now we're spending all this money to try to compile all the words.
And then someone said, why don't we just use all of the internet as the spelling dictionary?
Then there's no spelling dictionary or the IME if it's an Asian language.
And so this curve just keeps repeating itself.
And I think that neither is going to win out because data is always going to be valuable
because it's an input that you can't escape.
Whereas the flip side is not everything has data.
So if there's this opportunity, like, you know, I look at like a great example is like
how many of your lift drivers also have ways turned on just because they know the maps
are entirely accurate, but they don't know about accidents or emergency.
closures or weather or whatever.
And so having that data combined with the data plus algorithms, that hybrid is going to continue
to win.
Right.
Because only if you really want to be practical and actually solve the problem.
But I have to say one thing about this, which is so fascinating that you use the example
of search and you describe this curve that we're on.
And part of startups is sort of, you know, getting at the right place and point in time
along that curve and where you are in the moment.
Because sometimes I think a lot of academics or even people are really in love with their own
ideas, sometimes lose a big picture.
Like they've had this build up this expertise.
one area and they don't realize, like, practically speaking, the world has changed around you.
Because the point that I think is fascinating as well in the innovation story using the
search example is that Google was like the 15th search company to come around before it hit
success.
And that is kind of relevant to think about.
Right.
And it's super important, too, from the company building perspective, which is they had this
algorithm, which we all talked about back then, page rank, you know, appropriately named
but not.
That's fair, Larry.
But the, or web page ranking as well.
But the interesting thing was very rarely is like an.
algorithm like this secret sauce for a company. Because you could look at what goes on from the
outside and pretty much reverse engineer and algorithm. So again, back to, but if you have like
a data source that you can actually. And a monetization model. And it turns out one of the things
that Google built out was like they were crawling the web faster than Alta Vista. They made this bet
because they were machine learning people before machine learning was cool. They made the bet that having
the data was going to be. And it turns out that was what the barrier to entering the search market,
even for Microsoft and Bing
was like sucking in the entire internet
fast enough.
Step one.
Crawl the internet
while it's growing
at exponential rates.
And so,
you know,
back to,
but I actually want to bring up
one more thing
that I just think is kind of
really interesting
from a practical point of view.
I love it.
You're like the practical
person in this podcast.
I feel particularly practical today
because, well,
I can't play Go.
So I got to add something.
I'm just kidding.
The thing that I find
the most fascinating
about all of these solutions
in the space
is the engineering
of a product
that you can make a commitment to customers that works,
and then when it doesn't work, you can figure out why.
And so one of the things that's so interesting about all of this is debugging.
Interesting.
And how does that really fit in?
And sure, with Go, you lost the game.
And, of course, while they're building all of this,
they're figuring out, whoa, what did we do wrong to make this move repeatedly?
They're doing all of that debugging over the past end months.
But in, you know, if you just all of a sudden apply this to the enterprise space
or to adjusting news feeds or a zillion other things that you can think of.
Like, figuring out where it goes wrong,
like, that's actually really critical to a business.
Like, you can't put a product out there,
pick our sales forecasting example.
And then it's wrong.
Like, you can't just go, ooh, machines make mistakes just like people.
Your VP of sales could have messed up too.
Because nobody pays money to, like, a computer for it to be wrong.
Yeah.
And so how do we think about that?
Yeah, this whole idea is sort of transparency,
behind these models? In other words, do we know why they're behaving the way they're behaving?
Demystifying the black box.
You know, the superactive area of research right now, right? Which is how do I make the deep
learning models more transparent so that I can debug them? I can verify them. I can make
sure there's no systematic bias in them, right? Because until that, you couldn't do important
things like, hey, can this person have a loan or not? Because the government will say, you cannot
make that decision unless we understand why it is that you're making that decision. But you've made
the argument, Frank, that when it comes to
say self-driving cars, we kind of
it's no better or different than what the human
mind does. We don't know. We can't interrogate
the black box of the human mind
that's driving that car. So the
counterpoint of that is that
well, you're right, Stephen, that you know
you're paying for this computer to be smarter, but the
reality also is the stuff and not that smarter
than humans anyway, so who cares? Right, but
the problem is first, that's everybody
else, not me. Yeah. Like I
like I'm the best driver on the road.
It's all the other people.
But this is, I mean, you know, I like to always go back to this, the wonderful, wonderful
research on computers and society that the Stanford professors, Nasson Reeves did,
because one of the things that they really realized really back in the early 80s.
Oh, Clifford Nass.
Oh, Iron Rees.
Which was that there's something about a mechanical device that produces answers that makes the human brain
ascribe way more authority to it than there necessarily should be.
In fact, they even apply this to the world of voice.
recognition systems. They applied it to voice recognition. They applied it to chat bots before they
were chatbots. I mean, it's so the interesting thing is that I don't know how it's going to
take a major change in society for people collectively to just go, ah, people make mistakes
and that's okay, but machines can't. Like we have, especially in the United States, a very,
very low tolerance for devices making mistakes. Except when machines are doing a kind of alien
intelligence that humans cannot do.
Because again, the most fascinating thing about our last talk about AlphaGo, this current
talk is that at the end of the day, even though it opened, the AlphaGo Zero opened and closed
with similar moves to what humans would do, it converged very quickly.
There was a whole set of things in the middle that it did that were just things that humans
would never have done.
It's actually then augmenting us in a very different way because it's adding a completely
foreign intelligence.
So it's not even comparable to our own.
Yeah.
It's a different type of intelligence.
in the same way that animals have a different
type of intelligence. And then you make all kinds
of category errors when you say
giraffes aren't intelligent or bats
aren't intelligent. They're just intelligent in a
different way. Exactly. And we can't use the
human yardstick to compare them. It's the reason
we are more forgiving when a dog
pees all over your couch and your five-year-old
kid does the same freaking thing. Right.
That's definitely the case. The challenge that
we have in technology
is just the perception.
Like you, I mean, you can, you just see it
in the discussion of news feeds.
and algorithms and how people are like,
these should just be right.
I mean, like, yes.
And people should just have debugged this before it all happened.
And actually, it's not even all that crazy, sophisticated what's going on.
And what's weird is, of course, at the same time, people are critical of what's on the front
page of newspapers or what's in the first five minutes of TV newscast, which is literally
the same decision made by a human being who's just deciding what should we show first on CNN versus
Fox News.
Some human just made up with some black box.
their brain, which is augmented by their title of in charge of production.
Yeah.
And yet infinitely forgiving of those choices.
And so I actually think that there's a lot here.
And I don't have an answer, but I think it's no different than any other software,
which is if you're going to make something and offer it to people in a commerce situation.
It better damn work and it better not be wrong.
It better be clear like how it does it to you.
I mean, like people like Excel was really great at doing math.
quickly and then one day I find myself at the Naval War College having to explain
myself to a bunch of generals like how do we know it's right I literally just had
to sit there going I mean it's just right it just works and and it's one of
those those are the moments that contribute to me feeling like a lot of of empathy
for what's going on in the marketplace now about this and why I'm so so alarm
not alarm that's totally the wrong word so focused on this you know know the
outcome because like like until you've just sat there with a bunch of people
who control nuclear weapons telling you
we're using a spreadsheet to calculate it, the weight of being right doesn't really hit you.
Right.
Because before that, we were just like, oh, my God, it doesn't seem to make mistakes.
And we were like super ecstatic.
And it may, look, the charts are cool.
Yay, ship it.
And then one day they're like, is it going to work or not work?
And you're like, you can't, we couldn't prove it.
Like, that was the essence of it.
Like, I couldn't go to people at Boeing or people at the Navy or wherever in Wall Street and prove that Excel worked.
And then, of course, 30 years later, like, every time there's a mistake in Excel,
like it's a mistake in the human that typed it.
Like Ken Rogoff did a bunch of economic predictions
that forecast the recession in 2008,
and then all of a sudden you find out his model was wrong,
but the recession happened anyway.
Did he make the right prediction or not the right prediction?
And like, how does that work?
And I think that's what's going on here, too, with these things.
I have a philosophical question then for you guys.
That wasn't philosophical enough.
Well, one more philosophy.
We're being very philosophical here.
You know, you said something about how people make judgment calls
for what news to show on television, et cetera,
and we have these expectations of all algorithms.
And one of the topics we've discussed on this podcast,
Frank, you and I discuss it with Faye Faye,
is this idea of bias in algorithms
and how algorithms by definition can be biased.
Is one possibility of this type of work
because they kept using the phrase tabular rasa,
which of course I find so fascinating
because in human development,
there's an analogous world of this
where there was this theory
that the human brain was also a tabula rasa or blank slate,
and then they quickly learn, like, no,
we have millions of people,
years of evolution and DNA that's actually
we've actually coming in inheriting things.
Is there now a possibility that
algorithms can write themselves
at a true tabula rasa-like
way given this type of work?
Is that just way out there?
That one's above my pay grade.
These guys are throwing up their hands for our listeners
by the way.
I will give an example of sort of the things
that are hardwired in your DNA.
So they've done a bunch of experiments that
if you watch somebody's hand
getting pricked, your muscles in your
own hand will involuntarily contract. Now, the big caveat is if that person has a different skin
color than you, your hand doesn't contract. Really? I didn't know that. In other words, there's
something going on in your lizard brain that says, that's an other's hand. Therefore, it's not
relevant to me. Therefore, whatever reflexes caused your hands to contract in sympathy when somebody
jammed a needle into that hand, that's an example of this sort of pre-wiring. You know, the question
is, can algorithms be pre-wired that way?
Or unprewired, even?
Or unprewired.
Because that, to me, is where the opportunity lies.
I mean, you could definitely write a set of rules that said, you know, treat other people
in a different way, right?
That would be top down.
The peril that most people talk about today is not so much that the algorithms are biased,
but you've fed it not enough data so that your prediction is biased.
So the classic example of this is in the early days of vision recognition, some of the image
classification algorithms were categorizing people with dark skins as gorillas.
That happened because they didn't feed it enough data of dark skin people.
So when people talk about bias in algorithms, they're mostly talking about this phenomenon,
which is the human researcher or the human programmer selected an incomplete data set,
and therefore you got biased results, as opposed to somehow the architecture of the neural
network is biased inherently.
And I think that that's a very important point that is being where.
well studied, it's well articulated, but particularly in the supervised learning case,
the data that's input, at least today, almost every data set that you have is going to be
and have some inherent bias. Because you weren't aware of these factors when it was being
collected in the first place, because I think it's fair to say the awareness of all of these
issues is at an all-time high relative to that. But then again, you look at all the medical
studies and you're like, well, there haven't been very many women in most of these studies,
or there have been only women but studying a drug developed by.
Is it true in biology and genetics research too?
There's a lot of limitations.
Right.
And so, but I do think that anybody today embarking on using supervised learning, whether for all or part of the solution, that that data set is implicitly challenged.
And in particular, it's the labels.
Because even the which labels you pick or which labels you omit is going to create some bias in the model that you're unaware of.
That's exactly right.
Your cultural background, your history, the way you grew up will lead you to label an image a certain way.
That may be different than somebody who, right?
So what is the ground truth?
The best example of this are just looking at images and like sentiment analysis is such a big thing.
But images, like facial expressions have been studied all around the world for decades, multiple decades, about like what is happy, what is sad, what is questionable.
The same with speech and intonation.
Like sometimes, you know, ending on a high note sounds like you're asking a question unless you're in another culture where that's making an exclamation.
Or it can be vocal fry and people are complaining that women shouldn't speak that way.
The vocal fry thing is a perfect example of that going on right now.
And actually, the high note was the sort of the 80s version of that same thing.
And so it goes in vogue too, right.
Be super careful about the labels.
And to Frank's point, that's a great place for transparency.
Because if you can go to a future or a potential customer and talk about these, this is what our model is based on when we're forecasting your sales or telling you to optimize your assembly line.
That could really matter to that process.
So high level takeaways.
Hybrid works always.
This is a thing.
This is like a refrain.
I feel like I should get you a T-shirt that says hybrid works, God damn it.
Well, be careful because I'm really against a lot of hybrids.
Oh, yeah.
Like hybrid cloud computing.
Right, right.
Hybrid in terms of solutions between old and new for coding.
The old stuff that works definitely does work.
Old and new, academic and industrial, like exactly, all the things that make it practical versus pure, so to speak, purest.
I think another big takeaway is this revenge of the algorithms moment.
Yes.
There is so much momentum right now that says, basically, we're one labeled data set away from glory, right?
And this result basically shows you, wow, there's a lot of mileage that you can get out of reinforcement learning where there's no data, no labels.
One takeaway for me is just the element of surprise, because I've been thinking,
a lot, you know, just how, I mean, I used to work in the world of how humans learn, but what's
fascinating to me is that the system played itself at a level calibrated to itself at
every level. And in human learning, that doesn't happen. You're taught by your parents. You learn
from, you know, adults. You learn from people have more experience. There's all these different
things that happen. And the thing that's so fascinating me is that to me, it is very evolution-like.
It's like a big bang moment. And while I wouldn't hype and say that means we're going to end up here,
I do think that's very amazing, especially when you think about the surprises like, in the
In the paper, one of the things they talk about is that the system learned something that's
actually very easy for humans to learn way later in the game.
It's actually, I think, like, I forgot, Chichot or some specific move to, like, the latter.
The fact that the system took forever to learn something that's very first for humans, I just,
I'm endlessly fascinated by this relationship between humans and computers and what we can
learn from computers and vice versa, and also what it means for the field when you can actually
add how people learn to the field of artificial intelligence machine learning.
Yeah, with reinforcement learning, one of the challenges has been, sometimes you get into this training epic where there's no improvement, right?
Because you're just playing these games over and over again.
And like, what is it that causes the next game to be smarter?
And this one was like in three days, it got incredibly smart.
So that's off to these guys.
And let's also add the other takeaway, which is that you can do amazing things with simple architectures because we're at a point in a moment right now where you can have four TPUs instead of 48.
For me, one of the takeaways is if you can frame your problem with a set of constraints and a set of rules and a fixed set of operations, that's a very powerful concept that changes, because we've been so data-centric, people have stopped trying to think of their solutions algorithmically, and it's entirely possible that there is an algorithm.
Which is the same thing that I think we would have said a year ago, which is everybody was so focused on machine learning, the traditional algorithmic approaches might have been overlooked.
And here we are again with proof, but still, now we have to go back and say, well, you know, there's some basic machine learning that can work.
There's some basic algorithmic stuff.
But that key is really the set of constraints.
Yeah, for people who are interested in, I'd highly recommend Andre Caparthe's YC talk on this.
You can search for it.
It's on Google slides.
And the whole talk is basically about where will artificial general intelligence come from?
And he basically compares and contrast this sort of rule-based world of go to something messy like.
how would you build a pick and pack robot for Amazon?
And so if you're interested in this topic of what's the difference between role-based board games
versus the messy real world, it's a great presentation.
That's great for the generalized intelligence side.
So any parting messages for entrepreneurs building companies that have very specialized things?
Because one thing we have argued in this podcast, including you, Stephen, when we did a podcast
on building product and with machine learning, is that sometimes the best places to play
are very specific domain focus, whether or not they have a lot.
clearly explicit set of rules or not. So any thoughts on that? Yeah. So one takeaway is one of the
ways we evaluate startups from an investor's point of view is have you pick the best techniques
to solve the business problem that you're trying to solve. And I think this paper basically
opens up a frontier of exploration that you might not have thought about before. Because I think
if you started an AI-centric company today, you would definitely be on the get data, get labels,
train model, make prediction path. And this opens up another area you can try and figure out
am I solving a business problem where this supervised learning approach is going to lead me to
glory, this reinforcement learning approach is going to lead me to glory as opposed to supervised
learning or unsupervised learning?
I think is the most interesting thing for me in looking at different companies and just
and talking to the founders is just that there's this world going on of advances and new things
all the time.
And you could get on the treadmill of always trying to be the newest thing.
And we knew that the next generation of companies, you know, two years ago,
we're just going to be, you know, whatever you were doing before plus AI.
But the thing was it wasn't clear that that was always the best solution.
And then you replaced AI with, no, it has to be data and machine learning and labeling.
And so what Frank was saying is just super important to internalize,
which is that part of being a founder and building a new product is knowing that the reason
why you're choosing the technologies you're choosing and not just because you think that's what
investors are looking for.
Our jobs and that side are to actually fair.
out like who actually has a handle on the problem they're solving and a line that they can
draw from that problem to their chosen solution path.
And that is often, too often overlooked.
That's exactly right.
My favorite joke that I tell is when I asked, when I ask entrepreneurs what machine
learning algorithms are using in their product when they claim to be in machine learning
based startup, the answer I am not looking for is, I don't know, I'm sure we're using the good ones.
Well, on that note, you guys, thank you for joining the A6 and Z podcast.
Thank you.