Acquired - Google: The AI Company
Episode Date: October 6, 2025Google faces the greatest innovator's dilemma in history. They invented the Transformer — the breakthrough technology powering every modern AI system from ChatGPT to Claude (and, of course,... Gemini). They employed nearly all the top AI talent: Ilya Sutskever, Geoff Hinton, Demis Hassabis, Dario Amodei — more or less everyone who leads modern AI worked at Google circa 2014. They built the best dedicated AI infrastructure (TPUs!) and deployed AI at massive scale years before anyone else. And yet... the launch of ChatGPT in November 2022 caught them completely flat-footed. How on earth did the greatest business in history wind up playing catch-up to a nonprofit-turned-startup?Today we tell the complete story of Google's 20+ year AI journey: from their first tiny language model in 2001 through the creation Google Brain, the birth of the transformer, the talent exodus to OpenAI (sparked by Elon Musk's fury over Google’s DeepMind acquisition), and their current all-hands-on-deck response with Gemini. And oh yeah — a little business called Waymo that went from crazy moonshot idea to doing more rides than Lyft in San Francisco, potentially building another Google-sized business within Google. This is the story of how the world's greatest business faces its greatest test: can they disrupt themselves without losing their $140B annual profit-generating machine in Search?Sponsors:Many thanks to our fantastic Fall ‘25 Season partners:J.P. Morgan PaymentsSentryWorkOSShopifyAcquired’s 10th Anniversary Celebration!When: October 20th, 4:00 PM PTWho: All of you!Where: https://us02web.zoom.us/j/84061500817?pwd=opmlJrbtOAen4YOTGmPlNbrOMLI8oo.1Links:Sign up for email updates and vote on future episodes!Geoff Hinton’s 2007 Tech Talk at GoogleOur recent ACQ2 episode with Tobi LutkeWorldly Partners’ Multi-Decade Alphabet StudyIn the PlexSupremecyGenius MakersAll episode sourcesCarve Outs:We’re hosting the Super Bowl Innovation Summit!F1: The MovieTravelpro suitcasesGlue Guys PodcastSea of StarsStepchange PodcastMore Acquired:Get email updates and vote on future episodes!Join the SlackSubscribe to ACQ2Check out the latest swag in the ACQ Merch Store!Note: Acquired hosts and guests may hold assets discussed in this episode. This podcast is not investment advice, and is intended for informational and entertainment purposes only. You should do your own research and make your own independent decisions when considering any financial transactions.
Transcript
Discussion (0)
I went and looked at a studio, well, a little office that I was going to turn into a studio nearby, but it was not good at all.
It had dropped ceiling, so I could hear the guy in the office next to me.
You would be able to hear him talking on episodes.
Third co-host.
Third co-host.
Is it Howard?
No, it was like a lawyer.
It seemed to be like talking through some horrible problem that I didn't want to listen to, but I could hear every word.
Does he want millions of people listening to his conversations?
Right.
All right.
All right.
let's do a podcast let's do a podcast
who got the truth
is it you is it you is it you
who got the truth now
is it you is it you
sit me down
say it straight
another story on the way
who got the truth
welcome to the fall
2025 season of Acquired
the podcast about great companies
and the stories and playbooks behind them
I'm Ben Gilbert
I'm David Rosenthal
and we are your hosts.
Here's a dilemma.
Imagine you have a profitable business.
You make giant margins on every single unit you sell,
and the market you compete in is also giant,
one of the largest in the world, you might say.
But then on top of that, lucky for you,
you also are a monopoly in that giant market
with 90% share and a lot of lock-in.
And when you say monopoly,
monopoly as defined by the U.S. government.
That is correct.
But then, imagine this, in your research lab, your brilliant scientists come up with an invention.
This particular invention, when combined with a whole bunch of your old inventions by all your other brilliant scientists, turns out to create the product that is much better for most purposes than your current product.
So you launched the new product based on this new invention, right?
Right.
I mean, especially because out of pure benevolence, your scientists had published research papers about how,
awesome the new invention is, and lots of the inventions before also. So now there's new startup
competitors quickly commercializing that invention. So of course, David, you change your whole
product to be based on a new thing, right? Uh, this sounds like a movie. Yes, but here is the
problem. You haven't figured out how to make this new incredible product anywhere near
as profitable as your old giant cash printing business. So maybe you shouldn't launch that new
product. David, this sounds like quite the dilemma to me. Of course, listeners, this is Google
today and in perhaps the most classic textbook case of the innovators dilemma ever. The entire AI
revolution that we are in right now is predicated by the invention of the transformer out of the
Google brain team in 2017. So think OpenAI and ChatGBTDBT, Anthropic, NVIDIA, hitting all-time
highs, all the craziness right now depends on that one reason.
research paper published by Google in 2017. And consider this. Not only did Google have the
densest concentration of AI talent in the world 10 years ago that led to this breakthrough,
but today they have just about the best collection of assets that you could possibly ask for.
They've got a top-tier AI model with Gemini. They don't rely on some public cloud to host their
model. They have their own in Google Cloud that now does $50 billion in revenue. That is real
scale. They're a chip company with their tensor processing units or TPUs, which is the only real
scale deployment of AI chips in the world besides Nvidia GPUs. Maybe AMD, maybe, but these are
definitely the top two. Somebody put it to me in research that if you don't have a foundational
frontier model or you don't have an AI chip, you might just be a commodity in the AI market. And
Google is the only company that has both.
Google still has a crazy bench of talent,
and despite ChatGPT becoming kind of the Kleenex of the era,
Google does still own the text box,
the single one that is the front door to the Internet
for the vast majority of people
anytime anyone has intent to do anything online.
But the question remains,
what should Google do strategically?
Should they risk it all and lean into their birthright
to win in artificial intelligence?
Or will protecting their gobs of profit
from search hamstring them as the AI wave passes them by. But perhaps first, we must answer
the question, how did Google get here, David Rosenthal? So listeners, today, we tell the story
of Google, the AI company. Woo. You like that, David? Was that good? I love it. Did you hire
like a Hollywood script writing consultant without telling me? I wrote that 100% myself with no
AI. Thank you very much. No AI. Well, listeners, if you want to know every time an episode
drops, vote on future episode topics, or get access to corrections from past episodes, check out
our email list. That's Acquired.fm slash email. Come talk about this episode with the entire
acquired community in Slack after you listen. That's Acquired.fm slash Slack.
Speaking of the Acquired community, we have an anniversary celebration coming up.
We do. Ten years of the show. We're going to do an open Zoom call with everyone to celebrate,
Kind of like how we used to do our LP calls back in the day with LPs.
And we are going to do that on October 20th, 2025 at 4 p.m. Pacific Time.
Check out the show notes for more details.
If you want more required, check out our interview show, ACQ2.
Our last interview was super fun.
We sat down with Toby Lutke, the founder and CEO of Shopify,
about how AI has changed his life and where he thinks it will go from here.
So search ACQ2 in any podcast player.
and before we dive in, we want to briefly thank our presenting partner, J.P. Morgan Payments.
Yes, just like how we say every company has a story. Every company's story is powered by payments,
and J.P. Morgan Payments is a part of so many of their journeys from Seed to IPO and beyond.
So with that, this show is not investment advice. David and I may have investments in the companies we discuss,
and this show is for informational and entertainment purposes only. David, Google, the AI company.
So, Ben, as you were alluding to in that fantastic intro, really,
Really up in the game again.
If we rewind 10 years ago from today, before the Transformer paper comes out, all of the following people, as we've talked about before, were Google employees.
Ilya Sitz-Skiever, founding chief scientist of Open AI, who along with Jeff Hinton and Alex Koshowski
had done the seminal AI work on AlexNet and just published that a few years before all three of them
were Google employees, as was Dario Amadeh, the founder of Anthropic, Andre Carpathy, chief scientist at Tesla
until recently, Andrew Ng, Sebastian Thrun, Noam Shazir, all the deep-mind folks, Demis
Sassabas, Shane Legg, Mustafa Suleiman,
Mustafa now, in addition to it in the past,
having been a founder of DeepMind,
runs AI of Microsoft.
Basically, every single person
of note in AI worked
at Google, with the one exception of
Jan Lacoon who worked at Facebook.
Yeah, it's pretty difficult to trace
a big AI lab now
back and not find Google in its origin story.
Yeah, I mean, the analogy here is
it's almost as if at the dawn
of the computer era itself, a single company, like, say, IBM, had hired every single person
who knows how to code. So it would be like, you know, if anybody else wants to write a computer
program, oh, sorry, you can't do that. Anybody who knows how a program works at IBM. This is how
it was with AI and Google in the mid-2010s. But learning how to program a computer wasn't so hard
that people out there couldn't learn how to do it. Learning how to be an AI researcher,
significantly more difficult. Right. It was the stuff of very specific PhD programs with a,
very limited set of advisors, and a lot of infighting in the field of where the direction of
the field was going, what was legitimate versus what was crazy, heretical, religious stuff.
Yep. So then, yes, the question is, how do we get to this point? Well, it goes back to the start
of the company. I mean, Larry Page always thought of Google as an artificial intelligence company.
And in fact, Larry Page's dad was a computer science professor and had done his PhD at the University
of Michigan in machine learning and artificial intelligence.
which was not a popular field in computer science back then.
Yeah, in fact, a lot of people thought specializing in AI was a waste of time
because so many of the big theories from 30 years prior to that
had been kind of disproven at that point,
or at least people thought they were disproven.
And so it was frankly contrarian for Larry's dad
to spend his life and career and research work in AI.
And that rubbed off on Larry.
I mean, if you squint, page rank, the page
rank algorithm that Google was founded upon is a statistical method. You could classify it as part of
AI within computer science. And Larry, of course, was always dreaming much, much bigger.
I mean, there's the quote that we've said before on this show in the year 2000, two years
after Google's founding, when Larry says artificial intelligence would be the ultimate version of
Google. If we had the ultimate search engine, it would understand everything on the web. It would
understand exactly what you wanted, and it would give you the right thing. That's obviously
artificial intelligence. We're nowhere near doing that now. However, we can get incrementally closer,
and that is basically what we work on here. It's always been an AI company. Yep, and that was in 2000.
Well, one day, in either late 2000 or early 2001, the timelines are a bit hazy here, a Google engineer
named Georges Herrick is talking over lunch with Ben Gohm's famous Google engineer, who I think would go on
to lead search, and a relatively
new engineering hire named Nome Shazir. Now, Georges was one of Google's first 10 employees,
incredible engineer, and just like Larry Page's dad, he had a PhD in machine learning from
the University of Michigan. And even when George went there, it was still a relatively rare
contrarian subfield within computer science. So the three of them are having lunch, and
George says offhandedly to the group that he has a theory from his time as a PhD student.
that compressing data is actually technically equivalent to understanding it.
And the thought process is,
if you can take a given piece of information and make it smaller,
store it away, and then later reinstantiate it in its original form,
the only way that you could possibly do that
is if whatever force is acting on the data actually understands what it means
because you're losing information, going down to something smaller,
and then recreating the original thing.
It's like you're a kid in school.
You learn something in school.
You read a long textbook.
You store the information in your memory.
Then you take a test to see if you really understood the material.
And if you can recreate the concepts, then you really understand it.
Which kind of foreshadows big LLMs today are like compressing the entire world's knowledge
into some number of terabytes that's just like the smash-down little vector set.
Little, at least compared to all the information in the world.
but it's kind of that idea, right?
You can store all the world's information in an AI model
in something that is like kind of incomprehensible
and hard to understand,
but then if you uncompress it,
you can kind of bring knowledge back to its original form.
Yep, and these models demonstrate understanding, right?
Do they? That's the question.
And that's the question.
They certainly mimic understanding.
So this conversation is happening.
This is 25 years ago.
And Noam, the new hire,
the young buck, he sort of stops in his tracks and he's like, wow, if that's true, that's
really profound. Is this in one of Google's microchitchens? This is in one of Google's microkitchens. They're
having lunch. Where did you find this, by the way, a 25-year-old? This is in the Plex. This is like
a small little passage in Stephen Levy's great book that's been a source for all of our Google
episodes in the Plex. There's a small little throwaway passage in here about this, because
this book came out before chat GPT and AI and all that. So, Nome kind of latches on to
George, and keeps vibing over this idea. And over the next couple months, the two of them
decide, in the most Googly fashion possible, that they are just going to stop working on everything
else. And they're going to go work on this idea, on language models and compressing data,
and can they generate machine understanding with data? And if they can do that, that that would be good
for Google. I think this coincides with that period in 2001 when Larry Page fired all the
managers in the engineering organization.
And so everybody was just doing whatever they wanted to do.
Funny.
So there's this great quote from Georges in the book.
A large number of people thought it was a really bad thing for Noem and I to spend our talents on.
But Sanjay Gamowat, Sanjay, of course, being Jeff Dean's famous prolific coding partner, thought it was cool.
So, Georges would posit the following argument to any doubters that they came
cross. Sanjay thinks it's a good idea. And no one in the world is as smart as Sanjay. So why should
Nome and I accept your view that it's a bad idea? It's like if you beat the best team in football,
are you the new best team in football no matter what? Yeah. So all of this ends up taking
Noman, Georges, deep down the rabbit hole of probabilistic models for natural language.
Meaning, for any given sequence of words that appears on the internet, what is the probability
for another specific sequence of words to follow.
This should sound pretty familiar
for anybody who knows about LLM's work today.
Oh, kind of like a next word predictor.
Yeah, or next token predictor if you generalized it.
Yep.
So the first thing that they do with this work
is they create the did you mean spelling correction in Google search.
Oh, that came out of this?
That came out of this.
Nome created this.
So this is huge for Google
because obviously it's a bad user experience.
when you mistype a query and then need to type another one. But it's a tax to Google's
infrastructure because every time these mistyped queries are going, well, Google's infrastructure
goes and serves the results to that query that are useless and immediately overwritten with
the new one. Right. And it's a really tightly scoped problem where you can see like,
oh, wow, 80% of the time that someone types in god groomer, oh, they actually mean dog groomer
and they retype it. And if it's really high confidence, then you actually just correct it
without even asking them and then ask them if they want to opt out instead of opting in.
It's a great feature, and it's sort of a great first use case for this in a very narrowly scoped domain.
Totally. So they get this way and they keep working on it, Norman George,
and they end up creating a fairly large, I'm using large in quotes here, you know, for the time,
language model that they call affectionately Phil, the probabilistic hierarchical, inferential learner.
These AI researchers love creating their backronyms.
They love their word puns.
Yeah.
Yep.
So fast forward to 2003 and Susan Mujiski and Jeff Dean are getting ready to launch AdSense.
They need a way to understand the content of these third-party webpages, the publishers,
in order to run the Google Ad corpus against them.
Well, Phil is the tool that they use to do it.
Huh. I had no idea that language models were involved in this.
Yeah. So Jeff Dean borrows Phil and famously uses it to code up his implementation of AdSense in a week because he's Jeff Dean.
And boom, ad sense. This is billions of dollars of new revenue to Google overnight.
Because it's the same corpus of ads that are ad words that are search ads that they're now serving on third-party pages.
they just massively expanded the inventory for the ads that they already have in the system.
Thanks to Phil.
Thanks to Phil.
All right, this is a moment where we got to stop and just give some Jeff Dean facts.
Jeff Dean is going to be the through line of this episode of, wait, how did Google pull that off?
How did Jeff Dean just go home and over the weekend rewrite some entire giant distributed system
and figure out all of Google's problems?
Back when Chuck Norris facts were big, Jeff Dean Fax became a thing internally at Google.
I just want to give you some of my favorites.
The speed of light in a vacuum used to be about 35 miles per hour.
Then Jeff Dean spent a weekend optimizing physics.
So good.
Jeff Dean's pin is the last four digits of pie.
Only Googlers would come up with these.
Yes.
To Jeff Dean, NP means no problemo.
Oh, yeah.
I've seen that one before.
I think that one's my favorite.
Yes.
Oh, man.
So, so good.
also a wonderful human being who we spoke to in research and was very, very helpful. Thank you, Jeff. Yes. So language models
definitely work, definitely going to drive a lot of value for Google. And they also fit pretty beautifully
into Google's mission to organize the world's information and make it universally accessible and
useful if you can understand the world's information and compress it and then recreate it. Yeah,
that fits the mission, I think. I think that checks the box. Absolutely. So, Phil,
get so big that apparently by the mid-2000s, Phil is using 15% of Google's entire data center
infrastructure. And I assume a lot of that is ad-sense ad-serving, but also did you mean and all
the other stuff that they start using it for within Google. So early natural language systems
computationally expensive. Yes. So, okay, now mid-2000s, fast forward to 2007, which is a very,
very big year for the purposes of our story.
Google had just recently launched the Google Translate product.
This is the era of all the great, great products coming out of Google that we've talked about.
Maps and Gmail and docs and all the wonderful things that Chrome and Android are going to come later.
They had like a 10-year run where they basically launched everything you know of at Google except for search.
Truly, in a 10-year run.
And then there were about 10 years after that from 2013 on where they basically didn't launch any new products that you've heard.
about until we get to Gemini, which is this fascinating thing. But this 03 to 2013 era was just
so rich with hit after hit after hit. Magical. And so one of those products was Google Translate,
you know, not the same level of user base or perhaps impact on the world as Gmail or Maps or
whatnot, but still a magical, magical product. And the chief architect for Google Transly
was another incredible machine learning Ph.D. named Franz Ack. So Franz had a background in natural language processing and machine learning, and that was his PhD, he was German. He got his PhD in Germany. At the time, DARPA, the Defense Advanced Research Projects Agency, Division of the government, had one of their famous challenges going for machine translation. So Google and Franz, of course, enters this. And Franz,
and Franz builds an even larger language model that blows away the competition in this year's
version of the DARPA Challenge, this is either 2006 or 2007, gets a astronomically high
blue score for the time.
It's called the bilingual evaluation understudy is the sort of algorithmic benchmark for judging
the quality of translations, at the time higher than anything else possible.
So Jeff Dean hears about this in the work.
The Franz and the Translate team have done, and it's like, this is great. This is amazing.
When are you guys going to ship this in production?
Oh, I heard this story.
So, Jeff and Noem talk about this on the Dorcasch podcast.
Yes.
That episode is so, so good.
And Franz is like, no, no, no, no, Jeff, you don't understand.
This is research.
This isn't for the product.
We can't ship this model that we built.
This is a N-gram language model.
Grams are like a number of words in a cluster.
and we've trained it on a corpus of two trillion words from the Google search index.
This thing is so large, it takes it 12 hours to translate a sentence.
So the way the DARPA challenge worked in this case was you got a set of sentences on Monday,
and then you had to submit your machine translation of those set of sentences by Friday.
Plenty of time for the servers to run.
Yeah, they were like, okay, so we have whatever number of hours it is from Monday to Friday.
Let's use as much compute as we can to translate these couple sentences.
Hey, learn the rules of the game and use them to your advantage.
Exactly.
So, Jeff Dean being the engineering equivalent of Chuck Norris, he's like, hmm, let me see your code.
So Jeff goes and parachutes in and works with the translate team for a few months,
and he re-architects the algorithm to run on the words in the sentences in parallel instead of sequentially.
because when you're translating a set of sentences or a set of words in a sentence,
you don't necessarily need to do it in order.
You can break up the problem into different pieces, work on it independently.
You can parallelize it.
And you won't get a perfect translation, but imagine you just translate every single word.
You can at least go translate those all at the same time in parallel,
reassemble the sentence, and like mostly understand what the initial meaning was.
Yeah.
And as Jeff knows very well, because he and Sanjay basically built it with Erz Holze.
Google's infrastructure is extremely parallelizable.
Distributed, you can break up workloads into little chunks, send them all over the various data centers that Google has,
reassemble the projects, return that to the user.
They are the single best company in the world at parallelizing workloads across CPUs across multiple data centers.
CPUs, we're still talking CPUs here.
Yep.
And Jeff's work with the team gets that average sentence translation time down from 12 hours.
to 100 milliseconds.
And so then they ship it in Google Translate, and it's amazing.
This sounds like a Jeff Dean fact.
Well, you know, it used to take 12 hours, and then Jeff Dean took a few months with it.
No, it's 100 milliseconds.
Right, right, right, right, right.
So this is the first large, I'm using large in quotes here, language model used in production
in a product at Google.
They see how well this works.
Like, hmm, maybe we could use this for other things.
predicting search queries as you type.
That might be interesting.
And of course, the crown jewel of Google's business
that also might be interesting application for this,
the ad quality score for AdWords is literally
the predicted click-through rate on a given set of ad copy.
You can see how an LLM that is really good at ingesting information,
understanding it and predicting things based on that
might be really useful for calculating ad quality for Google.
Yep, which is the direct translation to Google's bottom line.
Indeed. Okay. So obviously all of that is great on the language model front.
I said 2007 was a big year. Also, in 2007,
begins the sort of momentous intersection of several computers.
computer science professors on the Google campus. So in April of 2007, Larry Page hires Sebastian Thrun
from Stanford to come to Google and work first part-time and then full-time on machine learning
applications. Sebastian was the head of sale at Stanford, the Stanford Artificial Intelligence
Laboratory, legendary AI laboratory that was big in the sort of first wave.
wave of AI back in the 60s, 70s, when Larry's dad was active in the field, then actually
shut down for a while and then had been restarted and re-energized here in the early 2000s.
And Sebastian was the leader, the head of sale.
Funny story about Sebastian, the way that he actually comes to Google, Sebastian was kind
enough to speak with us to prep for this episode.
I didn't realize it was basically an aqua hire.
He and some, I think it was grad students, were in the process of starting a company, had
term sheets from benchmark and Sequoia. Yes. And Larry came over and said, what if we just
acquire your company before it's even started in the form of signing bonuses? Yes, probably a very
good decision on their part. So Sale, this group within the CS department at Stanford, not only had
some of the most incredible, most accomplished professors and PhD AI researchers in the world.
they also had this stream of Stanford undergrads
that would come through and work there as researchers
while they were working on their CS degrees
or symbolic system degrees
or whatever it was that they were doing
as Stanford undergrads.
One of those people was Chris Cox,
who's the chief product officer at Meta.
Yeah, that was kind of how he got his start
in all of this, in AI,
and Facebook and Meta are going to come back into the story
here in a little bit.
Wow.
You really can't make this up.
Another undergrad who passed through sale while Sebastian was there
was a young freshman and sophomore
who would later drop out of Stanford to start a company
that went through Y Combinator's very first batch in summer 2005.
I'm on the edge of my seat. Who is this?
Any guesses?
Dropbox, Reddit. I'm trying to think who else was in the first batch.
Oh, no, no. But way more on the nose for this episode.
The company was a failed local mobile social network.
Oh, Sam Altman, looped.
Sam Altman.
That's amazing.
He was at sale at the same time.
He was at sale.
Yep.
It's an undergrad researcher.
Wow.
Wild, right?
We told you that it's a very small set of people that are all doing all of this.
Man, I miss those days.
Sam presenting at the WWC with Steve Jobs on stage with the double pop collar.
Right?
Different time in tech.
The double-popped collar.
That was amazing.
That was a vibe.
That was a moment.
Oh, man.
All right.
So, April 2007, Sebastian comes over from sale into Google, Sebastian Thrun.
One of the first things he does over the next set of months is a project called Ground Truth for Google Maps.
Which is essentially Google Maps.
It is essentially Google Maps.
So before Ground Truth, Google Maps existed as a product, but they had to get all the mapping data
from a company called Tele Atlas.
I think there were two.
They were sort of a duopoly.
Navtech was the other one.
Yeah, Navtech and Tele Atlas.
But it was this kind of crappy source of truth map data
that everyone used and you really couldn't do any better
than anyone else because you all just used the same data.
Yep.
It was not that good and it cost a lot of money.
Teleatlas and Navtech were multi-billion dollar companies.
I think maybe one or both them were public at some point,
then got acquired, but a lot of money, a lot of revenue.
Yeah.
And Sebastian's first thing was Street View, right? So he already had the experience of orchestrating this fleet of all these cars to drive around and take pictures.
Yes. So then, coming into Google, Ground Truth is this sort of moonshot-type project to recreate all the Tele Atlas data.
Mostly from their own photographs of streets from Street View. And they incorporated some other data. There was like census data they used. I think it was 40-something data sources to bring it all together. But Ground Truth was this very ambitious.
effort to create new maps from whole cloth.
Yep. And just like all of the AI and AI-enabled projects within Google that we're talking
about here works very, very well, very quickly. Huge win.
Well, especially when you hire a thousand people in India to help you sift through all
the discrepancies and the data and actually hand-draw all the maps.
Yes, we are not yet in an era of a whole lot of AI automation.
So on the back of this win with Ground Truth, Sebastian starts lobbying to Larry and Sergey.
Hey, we should do this a lot. We should bring in AI professors, academics. I know all these people into Google part-time. They don't have to be full-time employees. Let them keep their posts in academia, but come here and work with us on projects for our products. They'll love it. They get to see their work used by millions and millions of people. We'll pay them. They'll make a lot of money. They'll get Google stock. And they get to stay professors at their academic institutions.
win, win, win, win. So as you would expect, Larry and Sergey, you're like, yeah, yeah, yeah, that's a good idea. Let's do that. More of that. So in December of 2007, Sebastian brings in a relatively little-known machine learning professor from the University of Toronto named Jeff Hinton to the Google campus to come and give a tech talk. Not yet hiring him, but come give a tech talk to, you know, all the folks at Google and talk about,
Some of the new work, Jeff, that you and your Ph.D. and postdoc students there at the University of Toronto are doing on blazing new paths with neural networks.
And Jeff Hinton, for anybody who doesn't know the name, now very much known as the godfather of neural networks and really the godfather of kind of the whole direction that AI went in.
Modern AI.
He was kind of a fringe academic at this point in history.
I mean, neural networks were not a respected sub-tree of AI.
No, totally not.
And part of the reason is there had been a lot of hype 30, 40 years before, around neural networks that just didn't pan out.
So it was effectively everyone thought disproven and certainly backwater.
Yep.
Then, do you remember from our NVIDIA episodes my favorite piece of trivia about Jeff Hinton?
Oh, yes, that his grandfather, great-grandfather was George Bull.
Yep. He is the great-great-grandson of George and Mary Boole, who invented Boolean algebra and Boolean logic.
Which is hilarious, now that I know more about this, because that's the basic building block of symbolic logic, of defined deterministic computer science logic.
And the hilarious thing about neural nets is it's not. It's not symbolic AI. It's not, I feed you the specific instructions and you follow
a big if-then tree. It is non-deterministic. It is the opposite of that field.
Which actually just underscores again how sort of heretical this branch of machine learning and
computer science was. Right. So Ben, as you were saying earlier, neural networks not a new
idea and had all of this great promise in theory, but in practice just took too much computation
to do multiple layers. You could really only have a single or maybe small,
single-digit number of layers in a computer neural network up into this time.
But Jeff and his former postdoc, a guy named Jan Lacoon,
started vandalizing within the community,
hey, if we can find a way to have multi-layered, deep-layered neural network,
something we call deep learning, we could actually realize the promise here.
It's not that the idea is bad, it's that the implementation,
which would take a ton of compute
to actually do all the math,
to do all the multiplication required
to propagate through layer after layer
after layer of neural networks
to sort of detect and understand
and store patterns.
If we could actually do that,
a big multilayered neural network
would be very valuable
and possibly could work.
Yes.
Here we are now in 2007, mid-2000s.
Moore's Law has increased enough
that you could actually start to try to test
some of these theories.
Yep. So, Jeff comes and he gives this talk at Google. It's on YouTube. You can go watch it.
We'll link to it in the show notes. This is incredible. This is an artifact of history sitting
there on YouTube. And people at Google, Sebastian, Jeff D, and all the other folks who are
talking about, they get very, very, very excited. Because they've already been doing stuff like this
with Translate and the language models that they're working with. That's not using deep neural networks,
that Jeff's working on.
So here's this whole new architectural approach
that if they could get it to work
would enable these models
that they're building to work way better,
recognize more sophisticated patterns,
understand the data better,
very, very promising.
Again, kind of all in theory at this point.
Yeah.
So Sebastian Thrun brings Jeff Hinton
into the Google Fold after this tech talk,
I think first as a consultant
over the next couple of years
and then this is a basic.
Later, Jeff Hitton technically becomes an intern at Google.
Like, that's how they get around the part-time, full-time policies here.
Yep.
He was a summer intern in somewhere around 2011, 2012.
And mind you, at this point, he's like 60 years old.
Yes.
So in the next couple of years after 2007 here,
Sebastian's concept of bringing these computer science, machine learning academics,
into Google as contractors or part-time or interns,
basically letting them keep their academic posts and work on big projects for Google's products
internally goes so well that by late 2009, Sebastian and Larry and Sergey decide, hey, we should
just start a whole new division within Google. And it becomes Google X, the Moonshot Factory.
The first project within Google X, Sebastian leads himself.
Oh, David, don't say it. Don't say it.
I won't say the name of it.
We will come back to it later.
But for our purposes, for now, the second project would be critically important, not only for our story, but...
To the whole world.
Everything in AI, changing the entire world.
And that second project is called Google Brain.
But before we tell the Google Brain story, now is a great time to thank our friends at J.P. Morgan Payments.
Yes.
So today, we are going to talk about one of the core components of J.P. Morgan Payments
their Treasury solutions. Now, Treasury is something that most listeners probably do not spend a lot of time thinking about, but it's fundamental to every company.
Yep. Treasury used to be just a back office function, but now great companies are using it as a strategic lever.
With JP Morgan Payments Treasury Solutions, you can view and manage all your cash positions in real time and all of your financial activities across 120 currencies in 200 countries.
And the other thing that they acknowledge really in their whole strategy is that every,
business has its own quirk, so it's not a cookie-cutter approach. They work with you to figure
out what matters most for you in your business and then help you gain clarity, control, and
confidence. So whether you need advanced automation or just want to cut down on manual
processes and approvals, their real-time treasury solutions are designed to keep things running
smoothly, whether your treasury is in the millions or billions or perhaps like the company
we're talking about this episode in the hundreds of billions of dollars. And they have some great
strategic offerings like pay-by-bank, which lets customers pay you directly from their bank account.
It's simple, secure, tokenized, and you get faster access to funds and enhance data to
optimize revenue and reduce fees. This lets you send and receive real-time payments instantly
just with a single API connection to J.P. Morgan. And because J.P. Morgan's platform is global,
that one integration lets you access 45 countries and counting and lets you scale basically infinitely
as you expand. As we've said before, J.P. Morgan payments moves 10.
trillion dollars a day. So scale is not an issue for your business. Not at all. If you're wondering how to
actually manage all that global cash, J.P. Morgan again has you covered with their liquidity and
account solutions that make sure you have the right amount of cash and the right currencies in the
right places for what you need. So whether you're expanding into new markets or just want more
control over your funds, J.P. Morgan Payments is the partner you want to optimize liquidity,
streamline operations, and transform your treasury. To learn more about how J.P.
Morgan can help you and your company. Just go to JPMorgan.com slash acquired and tell them that Ben
and David send you. All right, David. So Google Brain. So when Sebastian left Stanford full
time and joined Google full time, of course, somebody else had to take over sale. And the person
who did is another computer science professor, brilliant guy named Andrew Ng. This is like all the
hits. All the hits. This is all the AI hits on this episode. So what does Sebastian do?
He recruits Andrew to come part-time, start spending a day a week on the Google campus.
And this coincides right with the start of X, and Sebastian formalizing this division.
So one day in 2010-2011 timeframe, Andrew's spending his day a week on the Google campus,
and he bumps into who else, Jeff Dean.
And Jeff Dean is telling Andrew about what he and Franz have done with language models
and what Jeff Hinton is doing in deep learning.
Of course, Andrew knows all this,
and Andrew's talking about what he in Sale are doing at Stanford,
and they decide, you know, the time might finally be right
to try and take a real big swing on this within Google
and build a massive, really large, deep learning model
in the vein of what Jeff Hinton has been talking about
on highly paralyzable Google infrastructure.
And when you say the time might be right,
Google had tried twice before,
and neither project really worked.
They tried this thing called Brains on Borg.
Borg is sort of an internal system
that they used to run all of their infrastructure.
They tried the cortex project,
and neither of these really worked.
So there's a little bit of scar tissue
in the sort of research group at Google
of our large-scale neural network
is actually going to work for us
on Google infrastructure.
So the two of them,
Andrew Ng and Jeff Dean,
pull in Greg Carrotto,
who is a neuroscience PhD
and amazing researcher,
was already working at Google. And in 2011, the three of them launch the second official
project within X, probably enough, called Google Brain. And the three of them get to work
building a really, really big, deep neural network model. And if they're going to do this,
they need a system to run it on. You know, Google is all about taking this sort of frontier
research and then doing the architectural and engineering system to make it actually run.
Yes. So Jeff Dean is working on this system on the infrastructure and he decides to name
the infrastructure distbelief, which of course is a pun, both on the distributed nature of the
system and also on, of course, the word disbelief because...
Noah thought I was going to work. Most people in the field thought this was not going to work.
and most people in Google thought this was not going to work.
And here's a little bit on why.
And it's a little technical, but follow me for a second.
All the research from that period of time pointed to the idea that you needed to be synchronous.
So all the compute needed to be sort of really dense, happening on a single machine with really high parallelism,
kind of like what GPUs do, that you really would want it all sort of happening in one place.
So it's really easy to kind of go look up and see, hey, what are the computed values for everything else in the system before I take my next?
move. What Jeff Dean wrote with disbelief was the opposite. It was distributed across a whole bunch
of CPU cores and potentially all over a data center or maybe even in different data centers.
So in theory, this is really bad because it means you would need to be constantly waiting around
on any given machine for the other machines to sync their updated parameters before you
could proceed. But instead, the system actually worked asynchronously without bothering to go and
get the latest parameters from other cores. So you were sort of updating parameters on
stale data. You would think that wouldn't work. The crazy thing is it did. Yes. Okay. So you've
got disbelief. What do they do with it now? They want to do some research. So they try out,
can we do cool neural network stuff? And what they do in a paper that they submitted in 2011,
right at the end of the year, is, and I'll give you the name of the paper first.
building high-level features using large-scale, unsupervised learning.
But everyone just calls it the cat paper.
The cat paper.
You talk to anyone at Google, you talk to anyone at AI.
They're like, oh, yeah, the cat paper.
What they did was they trained a large nine-layer neural network
to recognize cats from unlabeled frames of YouTube videos
using 16,000 CPU cores on a thousand different machines.
And listeners, just to like underscore how seminal this is, we actually talked with Sundar in prep for the episode, and he cited seeing the cat paper come across his desk as one of the key moments that sticks in his brain in Google's story.
Yeah.
A little later on, they would do a TGIF where they would present the results of the cat paper.
You talk to people at Google.
They're like that.
TGIF.
Oh, my God.
That's when it all changed.
Yeah.
It proved that large neural networks could actually learn meaningful patterns.
without supervision and without label data, and not only that, it could run on a distributed
system that Google built to actually make it work on their infrastructure. And that is a huge
unlock of the whole thing. Google's got this big infrastructure asset. Can we take this
theoretical computer science idea that the researchers have come up with and use disbelief to
actually run it on our system? Yep. That is the amazing technical achievement here.
that is almost secondary to the business impact of the cat paper.
I think it's not that much of a leap to say that the cat paper led to probably hundreds of billions of dollars of revenue generated by Google and Facebook and by dance over the next decade.
Definitely. Pattern recognizers in data.
So YouTube had a big problem at this time.
time, which was that people would upload these videos, and there's tons of videos being
uploaded to YouTube. But people are really bad at describing what is in the videos that they're
uploaded. And YouTube is trying to become more of a destination site, trying to get people to
watch more videos, trying to build a feed, increase dwell time, et cetera, et cetera. And the
problem is the recommender is trying to figure out what to feed. And it's only just working off
titles and descriptions that people were writing about their own videos.
Right. And whether you're searching for a video or they're trying to figure out what video
to recommend next, they need to know what the video is about. Yep. So the cap paper proves
that you can use this technology, a deep neural network running on distbelief, to go inside
of the videos in the YouTube library and understand what they were about and use that data
to then figure out what videos to serve to people.
If you can answer the question cat or not a cat,
you can answer a whole lot more questions, too.
Here's a quote from Jeff Dean about this.
We built a system that enabled us to train pretty large neural nets
through both model and data parallelism.
We had a system for unsupervised learning
on 10 million randomly selected YouTube frames,
as you were saying, Ben.
It would build up unsupervised representations
based on trying to reconstruct the frame
from the high-level representations.
We got that working and training
on 2,000 computers using 16,000 cores.
After a little while, that model was actually able to build a representation
at the highest neural net level where one neuron would get excited by images of cats.
It had never been told what a cat was,
but it had seen enough examples of them in the training data
of head-on facial views of cats that that neuron would then turn on for cats
and not much else.
It's so crazy.
I mean, this is the craziest thing about unlabel data, unsupervised learning,
that a system can learn what a cat is
without ever being explicitly told what a cat is
and that there's a cat neuron.
Yeah. And so then there's a iPhone neuron
and a San Francisco Giants neuron
and all the things that YouTube recommends.
Not to mention porn filtering,
explicit content filtering.
Not to mention copyright identification
and enabling revenue share with copyright holders.
Yeah, this leads to everything in YouTube
basically puts YouTube on the path
to today becoming the single biggest property on the internet and the single biggest media
company in the planet, this kicks off a 10-year period from 2012 when this happens until
chat GPT on November 30th, 2022, when AI is already shaping the human existence for all of us
and driving hundreds of billions of dollars of revenue. It's just in the YouTube feed and then
Facebook borrows it and they hire Jan Lacoon and they start Facebook AI research.
And then they bring it into Instagram and then TikTok and bite dance take it. And then it goes back to Facebook and YouTube with reels and shorts. This is the primary way that humans on the planet spend their leisure time for the next 10 years. This is my favorite David Rosenthalism. Everyone talks about 2022 onward as the AI era. And I love this point from you that actually for anyone that could make good use of a recommender system and a classifier system, basically in a company with a social feed, the AI era started in 2012.
Yes, the AI era started in 2012, and part of it was the cap paper.
The other part of it was what Jensen at Envidia always calls the Big Bang moment for AI, which was Alex Nett.
Yes.
So we talked about Jeff Hinton.
Back at the University of Toronto, he's got two grad students who he's working with in this era.
Alex Kershefsky and Ilya Sutskiber.
Of course.
Future co-founder and chief scientist of OpenA.I.
And the three of them are working with Jeff's deep neural network ideas and algorithms to create an entry for the famous ImageNet competition in computer science.
This is Fei-Fei's thing from Stanford.
It is a annual machine vision algorithm competition.
And what it was was Faye Faye had assembled a database of 14 million images that were hand-labeled.
Famously, she used a mechanical Turk on Amazon, I think, to get them all hand-labeled.
Yes, and I think that's right.
And so then the competition was, what team can write the algorithm that without looking at the labels, so just seeing the images, could correctly identify.
the largest percentage, the best algorithms that would win the competitions year over year,
we're still getting more than a quarter of the images wrong. So like 75% success rate, great,
way worse than a human. Can't use it for much in a production setting when quarter of the time you're
wrong. So then the 2012 competition. Along comes Alexnet. Its error rate was 15%. Still high,
but a 10% leap from the previous best
being a 25% error rate
all the way down to 15 in one year.
A leap like that had never happened before.
It's 40% better than the next best.
Yes.
On a relative basis.
Yes.
And why is it so much better, David?
What did they figure out
that would create a $4 trillion dollar company in the future?
So what Jeff and Alex and Ilya did
is they knew,
like we've been talking about all episode,
that deep neural networks had all this potential
and Moore's a lot advanced enough
that you could use CPUs
to create a few layers.
They had the aha moment
of what if
we re-architected
this stuff, not to run
on CPUs
but to run on a whole
different class
of computer chips
that were by their very nature
highly, highly, highly parallelized.
video game graphics cards, made by the leading company in the space at the time, Invidia.
Not obvious at the time, and especially not obvious that this highly advanced cutting-edge academic computer science research...
That was being done on supercomputers, usually.
That was being done on supercomputers with incredible CPUs would use these toy video game cards.
That retail for $1,000.
Yeah, it lasts at that point in time, a couple hundred bucks.
So the team in Toronto, they go out to like the local Best Buy or something.
They buy two Nvidia, G-Force GTX-580s, which were Nvidia's top of the line gaming cards at the time.
The Toronto team rewrites their neural network algorithms in CUDA, Nvidia's programming language.
they train it on these two off-the-shelf GTX-580s
and this is how they achieve their deep neural network
and do 40% better than any other entry in the ImageNet competition.
So when Jetson says that this was the big bang moment of artificial intelligence,
A, he's right, this shows everybody that, holy crap,
if you can do this with two off-the-shelf GTX-580s,
imagine what you could do with more of them or with specials.
chips. And B, this event is what sets Nvidia on the path from a somewhat struggling PC
gaming accessory maker to the leader of the AI wave and the most valuable company in the world
today. And this is how AI research tends to work, is there's some breakthrough that gets you
this big step change function. And then there's actually a multi-year process of optimizing from
there where you get these kind of diminishing returns curves on breakthroughs where the first
half of the advancement happens all at once. And then the second half takes many years after that to figure out. But it's rare and amazing and it must be so cool when you have an idea, you do it, and then you realize, oh my God, I just found the next giant leap in the field.
It's like I unlock the next level to use the video game analogy. Yes. I leveled up. So after Alex Net, the whole computer science world is a buzz.
people are starting to stop doubting neural networks at this point.
Yes.
So, after Alex Knapp, the three of them from Toronto,
Jeff Hinton, Alex Kershevsky, and Ilya Sitzkiever,
do the natural thing.
They start a company called DNN Research,
deep neural network research.
This company does not have any products.
This company has AI researchers.
Who just won a big competition.
And predictably, as you might imagine,
it gets acquired by Google almost immediately.
Oh, are you intentionally shortening this?
That's what I thought the story was.
Oh, it is not immediately.
Oh, okay.
There's a whole crazy thing that happens where the first bid is actually from Baidu.
Oh, ho.
Oh, I did not know that.
So Baidu offers $12 million.
Jeff Hinton doesn't really know how to value the company and doesn't know if that's fair.
And so he does what any academic would do to best determine the market value of the company.
He says, thank you so much.
I'm going to run an auction now, and I'm going to run it in a highly structured manner
where every time anybody wants to bid, the clock resets, and there's another hour
where anybody else can submit another bid.
No way.
So.
I didn't know this.
This is crazy.
He gets in touch with everyone that he knows from the research community, who is now
working at a big company, who he thinks, hey, this would be a good place for us to do
our research. That includes Bidu. That includes Google. That includes Microsoft. And there's
one other. Facebook, of course. It's a two-year-old startup. Wait, so it does not include Facebook?
It does not include Facebook. Think about the year. This is 2012. So Facebook's not really in the
AI game yet. They're still trying to build their own AI lab. Yeah, yeah, because Jan Lecun and Fairwood
start in 2013. Is it Instagram?
Nope. It is the most important part of the end of this episode.
Wait, well, it can't be Tesla, because Tesla is older than that.
Nope. Well, Open AI wouldn't get founded for years. Wow. Okay, you really got me here.
What company slightly predated Open AI doing effectively the same mission?
Oh, ha ha ha ha. Of course, of course. Hiding in plain sight.
Deep Mind. Wow.
DeepMind, baby. They are the fourth bidder in a four-way auction for DNN research. Now, of course,
right after the bidding starts, DeepMind has to drop out. They're a startup. They don't actually
have the cash to be able to buy. Yeah. Didn't even cross my mind because my first question was like,
where the hell would they get the money? Because they had no money. But Jeff Hinton already knows and respects Demis,
even though he's just doing this at the time startup called DeepMind. That's amazing.
Wait, how is DeepMind in the auction, but Facebook is not? Isn't that wild?
That's wild.
So the timing of this is concurrent with the, it was then called Nips.
Now it's called Nureps conference.
So Jeff Hinton actually runs the auction from his hotel room at the Hara's Casino in Lake Tahoe.
Oh my God.
Amazing.
So the bids all come in and we got to thank Cade Metz, the author of Genius Makers,
great book on the whole history of AI that we're actually going to reference a lot in this episode.
The bidding goes up and up and up.
at some point Microsoft drops out, they come back in, told you DeepMind drops out. So it's
bydo and Google really going at the end. And finally at some point, the researchers look at each
other and they say, where do we actually want to land? We want to land at Google. And so they
stop the bidding at $44 million and just say, Google, this is more than enough money. We're going
with you. Wow. I knew it was about $40 million. I did not know that old story. It's almost like
Google itself and the Dutch auction IPO process. Right.
How fitting.
That's kind of a perfect DNA, yes.
Wow.
And the three of them, we're supposed to split at 33 each, and Alex and Ilya go to Jeff
and say, I really think you should have a bigger percent.
I think you should have 40 percent, and we should each have 30, and that's how it ends up
breaking down.
Ah, wow.
What a team.
Well, that leads to the three of them joining Google Brain directly and turbocharging everything
going on there.
Spoiler alert, a couple years later.
Astro Teller who would take over running Google X after Sebastian Thrun left.
He would get quoted in the New York Times in a profile of Google X that the gains to Google's
core businesses and search and ads and YouTube from Google Brain have way more than funded
all of the other bets that they have made within Google X and throughout the company over the years.
It's one of these things that if you make something a few percent better that happens to do
tens of billions of dollars or hundreds of billions of dollars in revenue, you find quite a bit
of loose change in those couch cushions. Yes, quite a bit of loose change. But that's not where the
AI history ends within Google. There is another very important piece of the Google AI story
that is an acquisition from outside of Google, the AI equivalent of Google's acquisition of
YouTube. That's what we talked about in a minute ago. Deep Mind. But before we tell the deep mind,
story. Now is a great time to thank a new partner of ours, Sentry. Yes. Listeners, that is
S-E-N-T-R-Y, like someone's standing guard. Yes. Sentry helps developers debug everything from errors to
latency and performance issues, pretty much any software problem, and fix them before users get
mad. As their homepage puts it, they are considered, quote-unquote, not bad by over 4 million
software developers. And today we're talking about the way that Century works with another
company in the acquired universe, Anthropic. Anthropic used to have some older monitoring
systems in place. But as they scaled and became more complex, they adopted Century to find
and fix issues faster. So when you're building AI models, like we're talking about all
episode here, small issues can ripple out into big ones fast. Let's say you're running a huge
compute job like training a model. If one node fails, it can have massive downturn
extreme impacts, costing huge amounts of time and money. Sentry helped Anthropic detect bad
hardware early so they could reject it before causing a cascading problem and taking debugging
down to hours instead of days for them. And one other fun update from Century, they now have an
AI debugging agent called Sear. Sear uses all the context that Century has about your app usage
to run root cause analysis as issues are detected. It uses errors, span data, logs, and tracing,
and your code to understand the root cause, fix it, and get you back to shipping.
It even creates pull requests to merge code fixes in.
And on top of that, they also recently launched Agent and MCP server monitoring.
AI tooling tends to offer limited visibility into what's going on under the hood, shall we say?
Century's new tools make it easy to understand exactly what's going on.
This is everything from actual AI tool calls to performance across different models
and interactions between AI and the downstream.
services. We're pumped to be working with Century. We're big fans of the company and of all the
great folks we're working with there. They have an incredible customer lists, including not only
Anthropic, but Cursor, Versal, Linear, and More. And actually, if you're in San Francisco or
the Bay Area, Century is hosting a small invite-only event with David and I in San Francisco for
product builders on October 23rd. You can register your interest at century.io slash acquired. That's
Century, S-E-N-T-R-Y-O-S-A-O-A-O-A-O-A-O-C-A-O-A-W-A-W-A-R-A-R-A-S-A-R-A-R-W-R-A-R-R-E-N-A-R-I-R-R-E-N-A-R-R-E-R-E-R-E-R-A-R-R-U-S-A-R-E-R-S-E-R-U-S-E-R-A-R-R-S-A-R-S-E-R-E-S-E-R-S-S-E-E-R-S-I-S-R-R-S-E-S-I-R-1. So I remember when this deal
happened, just like I remember when the Instagram deal happened. Because the number was big at
the time. It was big, but I remember it for a different reason. It was like when Facebook bought
Instagram, like, oh my God, this is, wow, what a tectonic shift in the landscape of tech.
In January 2014, I remember reading on TechCrunch this random news. Right. You're like,
Deep what? That Google is spending a lot of money to buy something in London that I've never heard of.
that's working on artificial intelligence, question mark.
Right. This really illustrates how outside of mainstream tech AI was at the time.
Yeah. And then you dig in a little further and you're like, this company doesn't seem to have any products.
And it also doesn't even really say anything on its website about what deep mind is.
It says it is a, quote unquote, cutting edge artificial intelligence company.
Wait, did you look this up on the way back,
I did, I did. Oh, nice. To build general purpose learning algorithms for simulations,
e-commerce, and games. This is 2014. This does not compute, does not register.
Simulations, e-commerce, and games. It's kind of a random spattering of...
Exactly. It turns out, though, not only was that description of what deep mind was fairly
accurate. This company and this purchase of it by Google,
was the butterfly flapping its wings equivalent moment
that directly leads to open AI,
chat GPT, anthropic, and basically everything.
Certainly Gemini.
That we know, yeah, Gemini directly in the world of AI today.
And probably XAI, given Elon's involvement.
Yeah, of course, XAI.
In a weird way, it sort of leads to Tesla self-driving, too, Carpathie.
Yeah, definitely.
Okay, so what is the story here?
DeepMind was founded in 2010 by a neuroscience PhD named Demis Heshabis,
who previously started a video game company?
Oh yeah.
And a postdoc named Shane Legg at University College London and a third co-founder
who was one of Demas's friends from growing up, Mustafa Suleiman.
This was unlikely to say the least.
This would go on to produce a knight and Nobel Prize winner.
Yes.
So Demis, the CEO, was a childhood chess prodigy turned video game developer
who, when he was age 17 in 1994,
he had gotten accepted to the University of Cambridge,
but he was too young, and the university told him,
Hey, take a, you know, gap year, come back.
He decided that he was going to go work at a video game developer
at a video game studio called Bullfrog Productions for the year.
And while he's there, he created the game theme park, if you remember that.
It was like a theme park version of SimCity.
This was a big game.
This was very commercially successful.
Rollercoaster Tycoon would be sort of a clone of this that would have many, many sequels over the years.
Oh, I played a ton of that.
Yeah.
It sells 15 million copies.
copies in the mid-90s. Wow. Wild. Then after this, he goes to Cambridge, studies computer science
there. After Cambridge, he gets back into gaming, founds another game studio called Elixir.
That would ultimately fail. And then he decides, you know what, I'm going to go get my PhD in
neuroscience. And that is how Demis ends up at University College, London. There he meets
Shane Legg, who's there as a postdoc. Shane is a self-described at the time.
member of the lunatic fringe in the AI community in that he believes, this is 2008, 9, 10,
he believes that AI is going to get more and more and more powerful every year and that it will
become so powerful that it will become more intelligent than humans. And Shane is one of the
people who actually popularizes the term artificial general intelligence, AGI.
Oh, interesting.
Which, of course, lots of people talk about now and approximately zero people were afraid of that.
I mean, you had like the Nick Bostrom type folks, but very few people were thinking about
superintelligence or the singularity or anything like that.
For what it's worth, not Elon Musk.
He's not included in that list because Demis would be the one who tells Elon about this.
Yes. We'll get to it. So Demis and Shane hit it off. They pull in Mustafa, Demis's childhood friend, who is himself extremely intelligent. He had gone to the University of Oxford and then dropped out, I think, at age 19, to do other startup-y type stuff. So the three of them decide to start a company, Deep Mind. The name, of course, being a reference to deep learning, Jeff Hinton's work and everything coming out of the University of Toronto, and the goal that the three of these guys have of actually creating an intelligent
mind with deep learning.
Like, Javanilla and Alex aren't really thinking about this yet.
As we said, this is lunatic fringe type stuff.
Yes.
AlexNet, the cat paper, that whole world is about better classifying data.
Can we better sort into patterns?
It's a giant leap from there to say, oh, we're going to create intelligence.
Yes.
I think probably some people, almost certainly at Google, we're thinking, oh, we can create
narrow intelligence that'll be better than humans at certain tasks.
I mean, a calculator is better than humans at certain tasks.
Right.
But I don't think too many people were thinking, oh, this is going to be general intelligence smarter than humans.
Right.
So they decide on the tagline for the company is going to be solve intelligence and use it to solve everything else.
Ooh, I like it.
I like it.
Yeah, yeah.
They're good marketers, too, these guys.
So there's just one problem to do what they want to do.
Money.
Just say it.
Yeah, money. Money is the problem.
Right, right, right. Money is the problem for lots of reasons, but even more so than any other given startup in the 2010 era, it's not like they can just go spin up an AWS instance and like build an app and deploy it to the app store.
They want to build really, really, really, really, really big deep learning neural networks that requires Google size levels of compute.
Well, it's interesting.
And actually, they don't require that much funding yet.
The AI of the time was go grab a few GPUs.
We're not training giant LLMs.
That's the ambition eventually.
But right now what they just need to do is raise a few million bucks.
But who's going to give you a few million bucks when there's no business plan?
When you're just trying to solve intelligence, you need to find some lunatics.
It's a tough sell to VCs.
Except for the exact right.
As you say, they need to find some lunatics.
Oh, I chose my words carefully.
Yeah, we use the term lunatic in, uh...
It's endearing, ish.
Most endearing possible way here, given that they were all basically right.
So, in June 2010, Demis and Shane managed to get invited to the Singularity Summit in San Francisco, California.
Because they're not raising money for this in London.
Yeah, definitely not.
I think they tried for a couple months and learned that that was not going to be a viable path.
Yes.
The Summit, the Singularity Summit, Organized.
by Ray Kurzweil,
a future
Google employee,
I think,
chief futurist
and noted futurist,
Eliezer Yudkowski,
and Peter Thiel.
Yes.
So Demison and Shane
are excited
about getting this
invite.
This is probably
our one chance
to get funded.
But we probably
shouldn't just walk in
guns blaze it
and say,
Peter, can we pitch you?
Yeah.
So they finagle
their way into Demis getting to give a talk on stage at the summit.
Always the hack.
They're like, this is great.
This is going to be the hack.
The talk is going to be our pitch to Peter and Founders Fund.
Peter has just started Founders Fund at this point.
Obviously, a member of the PayPal Mafia, very wealthy.
I think he had a big Roth IRA at this point is the right way to frame it.
Big Roth IRA that he had invested in Facebook, first investor in Facebook.
He is the perfect target.
the architected the presentation at the summit to be a pitch directly to Peter, essentially, a
thinly veiled pitch. Shane has a quote in Parmy Olson's great book Supremacy that we used as a source
for a lot of this deep mind story. And Shane says, we needed someone crazy enough to fund an AGI company.
Somebody who had the resources not to sweat a few million and liked super ambitious stuff.
They also had to be massively contrarian because every professor that he would go talk to,
would certainly tell him
absolutely do not even think
about funding this.
That bend diagram sure
sounds a lot like Peter Thiel.
So they show up at the conference.
Demas is going to give the talk,
goes out on stage,
he looks out into the audience,
Peter is not there.
Turns out Peter wasn't actually
that involved in the conference.
He's a busy guy.
He's a co-founder,
co-organizer, but is a busy guy.
Yes.
Guys like, shoot,
oh, we missed our chance.
what are we going to do?
And then
Fortune turns in their favor.
They find out that Peter is hosting an after-party
that night at his house in San Francisco.
They get into the party.
Demis seeks out Peter,
and Demis is very, very, very smart.
As anybody who's ever listened to him talk
would immediately know,
he's like, rather than just pitching Peter head on,
I'm going to come about this obliquely.
He starts talking to Peter about
chess because he knows, as everybody does, that Peter Thiel loves chess.
And Demis had been the second highest ranked player in the world as a teenager in the under 14 category.
Good strategy. Great strategy. The man knows his chess moves. So Peter's like, hmm, I like you. You seem
smart. What do you do? And Demis explains he's got this AGI startup. They were actually here. He
gave a talk on stage as part of the conference. People are excited about this. And Peter said,
Okay. All right. Come back to Founders Fund tomorrow and give me the pitch. So they do. They make the pitch. It goes well. Founders Fund leads DeepMind's seed round of about $2 million. My hot times have changed for AI company seed rounds these days. Oh, yes. Imagine leading DeepMind's Seed round with less than $2 million check. And through Peter and Founders Fund, they get introduced.
Hey, Elon. You should meet this guy.
another member of the PayPal Mafia, Elon Musk.
Yes. So it's teed up in a pretty low-key way.
Hey, Elon, you should meet this guy. He's smart. He's thinking about artificial intelligence.
So Elon says, great, come over to SpaceX. I'll give you the tour of the place.
So Demas comes over for a launch and a tour of the factory. Of course, Demis thinks it's very cool,
but really he's trying to reorient the conversation over to artificial intelligence.
and I'll read this great excerpt from an article in The Guardian.
Musk told Hasabas his priority was getting to Mars as a backup planet in case something went wrong here.
I don't think he'd thought much about AI at this point.
Hasabas pointed out a flaw in his plan.
I said, what if AI was the thing that went wrong here?
Then being on Mars wouldn't help you because if we got there,
then it would obviously be easy for an AI to get there through our communication systems or whatever it was.
He hadn't thought about that.
So he sat there for a minute without saying anything, just sort of thinking.
Hmm, that's probably true.
Shortly after, Musk, too, became an investor in Deep Mind.
Yes.
Yes, yes.
I think it's crazy that Demis is sort of the one that woke Elon up to this idea of,
we might not be safe from the AI on Mars either.
Right, right.
I hadn't considered that.
So this is the first time the bit flips for Elon of,
we really need to figure out a safe, secure AI for the good of the people, that sort of seed
being planted in his head.
Yep.
Which, of course, is what Deep Mind's ambition is.
We are here doing research for the good of humanity like scientists in a peer-reviewed way.
Yep.
I think all that is true.
Also, in the intervening months to year after this meeting between Demis and Elon and
Elon investing in DeepMind, Elon also starts to get really, really excited and convinced about
the capabilities of AI in the near term, and specifically the capabilities of AI for Tesla.
Yes. Like with everything else in Elon's world, once the bit flips and he becomes interested,
he completely changes the way he views the world, completely sheds all the old ways and actions
that he was taking. And it's all about, what do I most do?
to embrace this new worldview that I have.
And other people have been working on for a while already by this point, AI driving cars.
Yep.
That sounds like it would be a pretty good idea for Tesla.
Does.
So Elon starts trying to recruit as many AI researchers as he possibly can
and machine vision and machine learning experts into Tesla.
And then AlexNet happens.
And man, Alex Nett's really, really, really good at identifying and classifying images and cat videos on YouTube and the YouTube recommender feed.
Well, is that really that different from a live feed of video from a car that's being driven and understanding what's going on there?
Can we process it in real time and look at differences between frames?
Perhaps controlling the car?
Not all that different.
So Elon's excitement, channeled initially through DeepMind and Demas, about AI and AI for Tesla, starts ratcheting up big time.
Yep.
Meanwhile, back in London, DeepMind is getting to work.
They're hiring researchers.
They're getting to work on models.
They're making some vague noises about products to their investors.
Maybe we could do something in shopping, maybe something in gaming, like the description on the website.
at the time of acquisition, said,
but mostly what they really, really want to do
is just build these models and work on intelligence.
And then, one day, in late 2013,
they get a call from Mark Zuckerberg.
He wants to buy the company.
Mark has woken up to everything that's going on at Google
after AlexNet
and what AI is doing for social media feed recommendations at YouTube.
the possibility of what it can do at Facebook and for Instagram.
He's gone out and recruited Jan Lacoon,
Jeff Hinton's old postdoc,
who's together with Jeff,
one of the sort of godfathers of AI and deep learning.
And really popularized the idea of convolutional neural networks,
the next hot thing in the field of AI at this point in time.
And so with Jan, they have created Fair Facebook AI research,
which is a Google brain rival within Facebook.
and remember who the first investor in Facebook was, who's still on the board.
Peter Thiel.
And is also the lead investor in DeepMind.
Where do you think Mark learned about DeepMind, Peter Thiel?
Was it? Do you know for sure that it was from Peter?
No, I don't know for sure, but like how else could Mark have learned about this startup in London?
I've got a great story of how Larry Page found out about it.
Oh, okay. Well, we'll get to that in one second.
So Mark calls and offers to buy the company.
And there are various rumors of how much Mark offered.
but according to Parmy Olson in her book Supremacy,
the reports are that it was up to $800 million,
company with no products and a long way from AGI.
That squares with what Cade Metz has in his book
that the founders would have made about twice as much money
from taking Facebook's offer versus taking Google's offer.
Yep. So, Demis, of course, takes this news to the investor group.
Which, by the way, is kind of against everything the company was founded on.
The whole aim of the company and what he's,
promise the team is that deep mind is going to stay independent, do research, publish in the
scientific community, we're not going to be sort of captured and told what to do by the whims
of a capitalist institution. Yep. So definitely some deal point negotiating that has to happen
with Mark and Facebook if this offer is going to come through. But Mark is so desperate at this
point. He is open to these very large deal point negotiations, such as Jan Lecun gets to stay in New York,
Jan Lacoon gets to stay operating his lab at NYU.
Jan Lacoon is a professor.
He's flexible on some things.
Turns out, Mark is not flexible on letting Demis keep control of DeepMind if he buys it.
Demis sort of argued for, we need to stay separate and carved out, and we need this
independent oversight board with his ability to intervene if the mission of DeepMind is no longer
being followed.
And Mark's like, no, you'll be a part of Facebook.
Yeah.
And you'll make a lot of money.
So, as this negotiation is going on, of course, the investors in deep mind get wind of this.
Elon finds out about what's going on.
He immediately calls up Demis and says, I will buy the company right now with Tesla stock.
This is late 2013, like early 2014.
Tesla's market cap is about $20 billion.
So Tesla stock from then to today is about a 70x run up.
Demis and Shane and Mustafa are like, wow, okay, there's a lot going on right now.
But to your point, they have the same issues with Elon and Tesla that they had with Mark.
Elon wants them to come in and work on autonomous driving for Tesla.
They don't want to work on autonomous driving.
Right. Or at least exclusively.
At least exclusively, yep.
So then Demis gets a third call.
from Larry Page.
Do you want my story
of how Larry knows about the company?
I absolutely want your story
of how Larry knows about the company.
All right.
So this is still early in DeepMind's life.
We haven't progressed all the way
to this acquisition point yet.
Apparently, Elon Musk is on a private jet
with Luke Nosek,
who's another member of the PayPal Mafia
and an angel investor in DeepMind.
And they're reading an email from Demis
with an update about a breakthrough
that they had where DeepMind AI
figured out a clever way to win
at the Atari game breakout.
Yes.
And the strategy it figured out
with no human training
was that you could bounce the ball
up around the edges of the bricks
and then without needing to intervene,
it could bounce around along the top
and win the game faster
without you needing to have a whole bunch of interactions
with the paddle down at the bottom.
They're watching this video
of how clever it is,
and flying with them on the same private plane
is Larry Page.
Of course, because,
Because Elon and Larry used to be very good friends.
Yes.
And Larry is like, wait, what are you watching?
What company is this?
And that's how he finds out.
Wow.
Yes.
Elon must have been so angry about all this.
And the crazy thing is this kinship between Larry and Demis is, I think, the reason why the deal gets done at Google.
Once the two of them get together, they are like peas in a pod.
Larry has always viewed Google as an AI company.
Demis, of course, views DeepMind so much as an AI company
that he doesn't even want to make any products
until they can get to AGI.
And Demis, in fact, we should share with listeners.
Demas told us this when we were talking to him to prep for this episode,
just felt like Larry got it.
Larry was completely on board with the mission of everything that DeepMind was doing.
And there's something else very convenient about Google.
they already have brain.
So Larry doesn't need Demis and Shane and Mustafa and DeepMind to come work on products within Google.
Right.
Brain is already working on products within Google.
Demis can really believe Larry when Larry says, nah, stay in London.
Keep working on intelligence.
Do what you're doing.
I don't need you to come work on products within Google.
Brain is like actively going and engaging with the product groups trying to figure out,
hey, how can we deploy neural nets into your product to make it better?
That's like their reason for being.
So they're happy to agree to this.
And it's working.
Brain and neural nets are getting integrated into search,
into ads, into Gmail, into everything.
It is the perfect home for DeepMind.
Home away from home, shall we say.
Yes.
And there's a third reason why Google's the perfect fit for DeepMind.
Infrastructure.
Google has all the compute infrastructure you could ever want,
right there on tap.
Yes, at least with CPUs.
so far. Yes. So, how's the deal actually happen? Well, after buying DNN research, Alan Eustace,
who David, you spoke with, right? Yep. Was Google's head of engineering at the time. He makes up his mind
that he wanted to hire all the best deep learning research talent that he possibly could,
and he had a clear path to do so. A few months earlier, Larry Page held a strategy meeting on an
island in the South Pacific in Cade Metz's book. It's an undisclosed island. Of course he did. Larry thought
that deep learning was going to completely change the whole industry. And so he tells his team,
this is a quote, let's really go big, which effectively gave Alan a blank check to go secure
all the best researchers that he possibly could. So in 2013, he decides I'm going to get on a
plane in December before the holidays and go meet DeepMind. Crazy story about this. Jeff Hinton,
who's at Google at the time, had a thing with his back where he couldn't sit down. He either
has to stand or lay. And so a long flight across the ocean is not doable. But
he needs to be there as a part of the diligence process. You have Jeff Hinton. You need to use him
to figure out if you're going to buy a deep learning company. And so Alan Eustace decides he's going to
charter a private jet and he's going to build this crazy custom harness rig so that
Jeff Hinton won't be sliding around when he's laying on the floor during takeoff and landing.
Wow. I was thinking the first part of this, I'm pretty sure Google has planes. They could just get into
Google Play. For whatever reason, this was a separate charter.
But it's not solvable just with a private plane.
You need also a harness.
Right.
And Alan is the guy who set the record for jumping out of the world's highest, was it a balloon?
I actually don't know.
The highest free fall jump that anyone has ever done even higher than that Red Bull stunt a few years before.
So he's very used to designing these custom rigs for airplanes.
He's like, oh, no problem.
You just need a bed and some straps.
I jumped out of the atmosphere in a scuba suit.
I think we'll be fine.
That is amazing.
So they fly to London, they do the diligence, they make the deal.
Demis has true kinship with Larry, and it's done.
550 million U.S. dollars.
There's an independent oversight board that is set up to make sure that the mission and goals of DeepMind are actually being followed.
And this is an asset that Google owns today that, again, I think is worth half a trillion dollars if it's independent.
Do you know what other member of the PayPal Mafia
gets put on the Ethics Board after the acquisition.
Reid Hoffman?
Read Hoffman.
Has to be, given the Open AI tie later.
We are going to come back to read in just a little bit here.
Yes.
So after the acquisition, it goes very well, very quickly.
Famously, the data center cooling thing happens
where DeepMind carved off some part of the team
to go and be an emissary to Google
and look for ways to use DeepMind.
And one of them is around data center cooling,
very quickly, July of 2016, Google announces a 40% reduction in the energy required to cool data centers. I mean, Google's got a lot of data centers. A 40% energy reduction. I actually talked with Jim Gao, who's a friend of the show, and actually led a big part of this project. And, I mean, it was just the most obvious application of neural networks inside of Google right away. Pays for itself.
Yeah, imagine that paid for the acquisition pretty quickly there.
Yes. David, should we talk about AlphaGo on this episode?
Yeah, yeah, yeah.
I watched the whole documentary that Google produced about it. It's awesome.
This is actually something that you would enjoy watching, even if you're not researching a podcast episode,
and you're just looking to pull something up and spend an hour or two. I highly recommend it.
It's on YouTube. It's the story of how DeepMind post-acquisition from Google trained a model
to beat the World Go champion at Go. And, I mean, everyone, the whole Go community coming in
thought there's no chance. This guy, Lee C. Dahl, is so good that there's no way that an AI could
possibly beat him. It's a five-game thing. It just won the first three games straight. I mean,
completely cleaned up. And with inventive, new creative moves that no human has played before.
That's sort of the big, crazy takeaway. There's a moment in one of the games right where it makes a
move of people like, is that a mistake? Like, that must have just been an error? Yeah, move 37. Yeah,
yeah. And then a hundred moves later, it plays out. That it was like completely genius.
humans are now learning from deep minds strategy of playing the game and discovering new
strategies. A fun thing for acquired listeners who are like, why is it Go? Go is so complicated.
Compared to chess, chess has 20 moves that you can make at the beginning of the game in any given
turn, and then midgame there's like 30 to 40 moves that you could make. Go on any given turn
has about 200. And so if you think commentatorily, the number of possible configurations of
the board is more than the number of atoms in the universe. That's a great Demis quote,
by the way. And so he says, even if you took all the computers in the world and ran them for a
million years as of 2017, that wouldn't be enough compute power to calculate all the possible
variations. So it's cool because it's a problem that you can't brute force. You have to do
something like neural networks. And there is this white space to be creative and explore. And so it's
served as this amazing breeding ground for watching a neural network be creative against a human.
Yeah. And of course, it's totally in with Demis' background and the DNA of the company of playing games.
You know, Demis was chess champion. And then after Go, then they play StarCraft, right?
Oh, really? I actually didn't know that.
Yeah, that was the next game that they tackle was StarCraft, you know, a real-time strategy game against an opponent.
And that'll come back up in a sec with another opponent.
here in OpenAI.
Yes, David, but before we talk about the creation of the other opponent,
should we think another one of our friends here required?
Yes, we should.
All right, listeners, we are here today to tell you about a new friend of the show.
We are very excited about Work OS.
Yes, if you're building software that is used in enterprises,
you've probably felt the pain of integrating things like SSO,
SCIM, or SCIM, or SCIM, permissions, audit logs,
and all the other features that are required by big customers.
And if you haven't felt this pain yet, just wait until you get your first big enterprise customer and trust us, you will.
Yes. WorkOS turns these potential deal blockers into simple drop-in APIs. And while WorkOS had great product market fit a few years ago with developers who just want to save on some headache, they really have become essential in the AI era.
Yeah. I was shocked when they sent over their latest customer list. Almost all the big AI companies use WorkOS today as the way that they've been able to rapidly scale revenue so fast. Companies like OpenAI, Anthropic, Cursor, Perplexity, Sierra, Replit, Vercell, hundreds of other AI startups all rely on Work OS as their auth solution.
So I called the founder to ask why, and he said it's basically two things. One, in the AI era, these companies scale so much.
faster, that they need things like authorization, authentication, and SSO quickly to become
enterprise ready and keep up with customer demand even early in life, unlike older SaaS companies
of yesteryear. And two, unlike that world, where you could bring your own little SaaS product
just for you and your little team, these AI products reach deep into your company's systems and
data to become the most effective. So IT departments are scrutinizing heavier than ever to make sure
that new products are compliant before they can adopt them. Yeah, it's this.
kind of like second order effect of the AI era, that the days of, oh, just swipe a credit card,
bring your own SaaS solution for your product team. You actually need to be enterprise ready a lot
sooner than you did before. Yeah. It's not just about picking up that big potential customer
for the revenue itself either. It's about doing it so your competitors don't. Enterprise readiness
has become so table stakes for companies no matter their stage. And Work OS is basically the weapon
of choice for the best software companies to shortcut this process and get back to focusing
on what makes their beer taste better, building the product itself.
Amen, amen.
So if you're ready to get started with just a few lines of code for Samil, Skim, Arbock, SSO,
authorization, authentication, and everything else to please IT admins and their checklists,
check out WorkOS.
It's the modern software platform to make all this happen.
That's WorkOS.com and just tell them that Ben and David sent you.
All right, David.
So what are the second order effects of Google buying Deuteman?
mind? Well, there's one person who is really, really, really upset about this, and maybe two
people, if you include Mark Zuckerberg, but Mark tends to play his cards a little closer to the
best. Of course, Elon Musk is very upset about this acquisition. When Google buys DeepMind
out from under him, Elon goes ballistic. As we said, Elon and Larry had always been very close.
And now here's Google, who Elon has already started to sour on a little bit as he's now trying to hire AI researchers.
And you've got Alan Eustace flying around the world sucking up all of the AI researchers into Google.
And Elon's invested in DeepMind, wanted to bring DeepMind into his own AI team at Tesla and gone out from under him.
So this leads to one of the most fateful dinners in Silicon Valley.
history. Organized in the summer of 2015 at the Rosewood Hotel on Sand Hill Road,
of course, where else would you do a dinner in Silicon Valley, but the Rosewood, by two of the
most leading figures in the valley at the time, Elon Musk and Sam Altman. Sam, of course,
being president of Y Combinator at the time. So what is the purpose of this
dinner, they are there to make a pitch to all of the AI researchers that Google, and to a certain
extent Facebook, have sucked up and basically created this duopoly status on.
Again, Google's business model and Facebook's business model, these feed recommenders or these
classifiers turn out to be unbelievably valuable so they can, it's funny in hindsight saying this,
pay tons of money to these people.
Tons of money, like millions of dollars.
Take them out of academia and put them into their dirty capitalist research labs inside the companies.
Selling advertising.
Yes.
How dirty could you be?
And the question in the pitch that Elon and Sam have for these researchers gathered at this dinner is,
what would it take to get you out of Google for you to leave?
And the answer, they go around the table from almost everybody is nothing.
You can't. Why would we leave?
We're getting paid way more money than we ever imagined.
Many of us get to keep our academic positions and affiliations.
And we get to hang out here at Google.
With each other.
With each other.
Iron sharpens iron.
These are some of the best minds in the world, getting to do cutting-edge research
with enormous amount of resources and hardware at their disposal.
It's amazing.
It's best infrastructure in the world.
We've got Jeff Dean here.
There is nothing you could tell us that would cause us to leave Google.
Except there's one person who is intrigued.
And to quote from an amazing Wired article at the time by Cade Metz, who would later write genius makers, right?
Yep, exactly.
Quote is, the trouble was so many of the people most qualified to solve these problems were already working for Google.
And no one at the dinner was quite sure.
that these thinkers could be lured into a new startup,
even if Musk and Maltman were behind it.
But one key player was at least open to the idea of jumping ship.
And then there's a quote from that key player,
I felt like there were risks involved.
But I also felt like it would be a very interesting thing to try.
It's the most Ilya quote of all time.
The most Ilya quote of all time,
because that person was Ilya Sutskiber.
Of course, of Alexnet and D&N research and Google.
and about to become founding chief scientist of open AI.
So the pitch that Elon and Sam are making to these researchers is,
let's start a new nonprofit AI research lab
where we can do all this work out in the open.
You can publish.
Free of the forces of Facebook and Google and independent of their control.
Yes, you don't have to work on products.
You can only work on research.
You can publish your work.
It will be open. It will be for the good of humanity. All of these incredible advances, this intelligence that we believe is to come, will be for the good of everyone, not just for Google and Facebook.
And for one of the researchers, it seemed too good to be true. So they basically weren't doing it because they didn't think anyone else would do it. It's sort of an activation energy problem where once Ilya said, okay, I'm in. And once he said, I'm in, by the way, Google came back with a big counter, something like double the offer.
And I think it was delivered from Jeff Dean personally, and Ilya said, nope, I'm doing this.
That was massive for getting the rest of the top researchers to go with him.
And it was nowhere near all of the top researchers who left Google to do this.
But it was enough.
It was a group of seven or so researchers who left Google and joined Elon and Sam and Greg Brockman from Stripe, who came over to create open AI.
Because that was the pitch.
We're all going to do this in the open.
And that's totally what it was.
It totally is what it was.
And the stated mission of OpenAI was to, quote, advance digital intelligence in the way that is most likely to benefit humanity as a whole, unconstrained by a need to generate financial return.
Which is fine as long as the thing that you need to fulfill your mission doesn't take tens of billions of dollars.
Yes.
So here's how they would fund it originally.
There was a billion dollars pledged.
Yes.
And that came from famously Elon Musk, Sam Altman, Reid Hoffman,
Jessica Livingston, who I think most people don't realize was part of that initial tranche,
and Peter Thiel.
Yep.
Founders Fund, of course, would go on to put massive amounts of money into Open AI itself later as well.
The funny thing is, it was later reported that a billion dollars was not actually collected,
only about 130 million of it
was actually collected to fund this nonprofit.
And for the first few years,
that was plenty for the type of research they were doing,
the type of compute they needed.
Most of that money was going to paying salaries to the researchers.
Not as much as they could make at Google and Facebook,
but still a million or $2 million for these folks.
Right.
And yeah, so that really worked until it really didn't.
Yeah.
So, David, what were they doing in the early days?
Well, in the first days, it was all hands on deck recruiting and hiring researchers.
And there was the initial crew that came over.
And then pretty quickly, after that in early 2016, they get a big, big win when Dario Amadei leaves Google, comes over, joins Ilya and crew at Open AI, dream team, you know, assembling here.
And was he on Google Brain before this?
He was on Google Brain.
Yep. And he, along with Ilya, would run large parts of Open AI for the next couple years
before, of course, leaving to start Anthropic. But we're still a couple of years away from
Anthropic, Claude, ChatGPT, Gemini, everything today. For at least the first year or two,
basically the plan at OpenAI is let's look at what's happening at DeepMind and show the research
community that we can do as a new lab do the same incredible things that they're doing and maybe
even do them better. Is that why it looks so game-like and game-focused? Yes. Yes. So they start
building models to play games. Famously, the big one that they do is Dota 2, Defense of the
Agents 2, the massively online battle arena video game. They're like, all right, well, deep mind you're
playing StarCraft. Well, we'll go play Dota 2. That's even more complex, more real time.
the emergent properties of Go, the game would devise unique strategies that you wouldn't see
humans trying. So it clearly wasn't humans coded their favorite strategies and rules in. It was
emergent. They did other things. They had a product called Universe, which was around training
computers to play thousands of games from Atari games to open world games like Grand Theft Auto.
They had something where they were teaching a model how to do a Rubik's Cube. And so it was a diverse set
of projects that didn't seem to coalesce around one of these is going to be the big thing.
Yeah. It was research stuff. It was what DeepMind was doing. Yeah. It was like a university
research. It was like DeepMind. And if you think back to Elon being an investor in DeepMind,
being really upset about Google acquiring it out from under him, makes sense. And I think Elon
deserves a lot of credit for having his name and his time attached to OpenA at the beginning.
lot of the big heavy hitter recruiting was Elon throwing his weight behind this. I'm willing to
take a chance. Absolutely. Okay, so that's what's going on over at OpenAI, doing a lot of deep mind
like stuff, a bunch of projects, not one single obvious big thing they're coalescing around.
It's not chat GPT time. Let's put it that way. Let's go back to Google, because last we sort of
checked in on them, yeah, they bought DeepMind, but they had their talent rated. And I don't
want you to get the wrong impression about where Google is sitting just because some people left
to go to OpenAI. So back in 2013, when Alex Krashefski arrives at Google with Jeff Hinton
and Ilius at Skeaver, he was shocked to discover that all their existing machine learning
models were running on CPUs. People had asked in the past for GPUs since machine learning
workloads were well suited to run in parallel, but Google's infrastructure team had pushed back
and said the added complexity in expanding and diversifying the fleet, let's keep things simple.
That doesn't seem important for us.
We're a CPU shop here.
Yes.
And so to quote from Genius Makers, in his first days at the company, he went out and bought a GPU machine, this is Alex, from a local electronics store, stuck it in the closet, down the hall from his desk, plugged it into the network, and started training his neural networks on this lone piece of hardware.
Just like he did in academia, except this time Google's paying for the electricity.
Obviously, one GPU is not sufficient, especially as more Googlers wanted to start using it to.
And Jeff Dean and Alan Eustace had also come to the conclusion that disbelief, while Amazing, had to be re-architected to run on GPUs and not CPUs.
So, spring of 2014 rolls around, Jeff Dean and John Giandra, who we haven't talked about this episode.
Yeah, J-G.
Yes, you might be wondering, wait, isn't that the Apple guy?
Yes, he went on to be Apple's head of AI, who at this point in time was at Google and oversaw Google Brain, 2014.
they sit down to make a plan
for how to actually
formally put GPUs
into the fleet
of Google's data centers
which is a big deal
it's a big change
but they're seeing
enough reactions to neural networks
that they know to do this
yeah after Alex Net is
just a matter of time
yeah so they settle on a plan
to order 40,000 GPUs
from Nvidia
yeah of course
who else you get to order them from
for a cost of $130 million
dollars
that's a big enough price
tag that the request gets elevated to Larry Page, who personally approves it, even though
finance wanted to kill it, because he goes, look, the future of Google is deep learning.
As an aside, let's look at Nvidia at the time. This is a giant, giant order. Their total
revenue was $4 billion. This is one order for $130 million. I mean, Nvidia is primarily a consumer
graphics card company at this point. Yes, and their market cap is $10 billion. It's
Almost like Google gave Nvidia a secret that, hey, not only does this work in research like
the ImageNet competition, but neural networks are valuable enough to us as a business to make
a hundred plus million dollar investment in right now, no questions asked.
We got to ask Jensen about this at some point.
This had to be a tell.
This had to really give Nvidia the confidence, oh, we should way forward invest on this being a giant thing
in the future. So all of Google wakes up to this idea. They start really putting it into their
products. Google Photos happened. Gmail starts offering typing suggestions. David, as you pointed out
earlier, Google's giant AdWords business started finding more ways to make more money with
deep learning. In particular, when they integrated it, they could start predicting what ads people
would click in the future. And so Google started spending hundreds of millions more on GPUs on top of that
$130 million, but very quickly paying it back from their ad system. So it became more and more
of a no-brainer to just buy as many GPUs as they possibly could. But once neural nets started to work,
anyone using them, especially at Google scale, kind of had this problem. Well, now we need to do
giant amounts of matrix multiplications. Anytime anybody wants to use one, the matrix multiplications
are effectively how you do that propagation through the layers of the neural network. So you sort
to have this problem. Yes, totally. There's the inefficiency of it. But then there's also the
business problem of, wait a minute, it looks like we're just going to be shipping hundreds of millions,
soon to be billions of dollars over to NVIDIA every year for the foreseeable future.
Right. So there's this amazing moment right after Google rolls out speech recognition,
their latest use case for neural nets, just on Nexus phones. Because again, they don't have the
infrastructure to support it on all Android phones. It becomes a super,
popular feature, and Jeff Dean does the math and figures out if people use this for, I don't know, call it three minutes a day, and we roll it out to all billion Android phones, we're going to need twice the number of data centers that we currently have across all of Google just to handle it.
Just for this feature. Yeah. There's a great quote where Jeff goes to Ers Holtzl and goes, we need another Google.
Or, David, as you were hinting at, the other option is we build a new type of chip customized for just our particular use case.
Yeah. Matrix multiplication, tensor multiplication, a tensor processing unit, you might say.
Ah, yes. Wouldn't that be nice? So conveniently, Jonathan Ross, who's an engineer at Google, has been spending his 20% time at this point in history,
working on an effort involving FPGAs. These are essentially expensive but programmable chips
that yield really fantastic results. So they decide to create a formal project to take that work,
combine it with some other existing work, and build a custom ASIC or an application-specific integrated
circuit. So enter, David, as you said, the tensor processing unit made just for neural
networks that is far more efficient from GPUs at the time with the trade-off that you can't
really use it for anything else. It's not good for graphics processing. It's not good for lots of
other GPU workloads, just matrix multiplication and just neural networks. But it would enable
Google to scale their data centers without having to double their entire footprint. So the big
idea behind the TPU, if you're trying to figure out what was the core insight, they use
reduced computational precision. So it would take numbers like 4,586.8272, and round it just to 4,586.
0.8, or maybe even just 4,586 with nothing after the decimal point. And this sounds kind of
counterintuitive at first. Why would you want less precise, rounded numbers for this complicated
math? The answer is efficiency. If you can do the heavy lifting in your software architecture,
or what's called quantization to account for it, you can store information as less precise numbers.
Then you can use the same amount of power and the same amount of memory and the same amount of
transistors on a chip to do far more calculations per second. So you can either spit out answers
faster or use bigger models. The whole thing is quite clever behind the TPU. The other thing that has to
happen with the TPU is it needs to happen now because it's very clear speech to text is a thing. It's
very clear some of these other use cases at Google. Yeah. Demand for all of this stuff that's coming
out of Google brain is through the roof immediately. Right. And we're not even two LLMs yet. It's just like
everyone sort of expects some of this, whether it's computer vision and photos or speech
recognition. It's just becoming a thing that we expect, and it's going to flip Google's
economics upside down if they don't have it. So the TPU was designed, verified, built, and
deployed into data centers in 15 months. Wow. It was not like a research project that
could just happen over several years. This was like a hair-on-fire problem that they launched
immediately. One very clever thing that they did was, A, they used the FPGAs as a stopgap.
So even though they were, like, too expensive on a unit basis, they could get them out as a test fleet and just make sure all the math worked before they actually had the A6 printed at, I don't know if it was a TSMC, but, you know, fabbed and ready.
The other thing they did is they fit the TPU into the form factor of a hard drive.
So it could actually slot into the existing server racks.
You just pop out a hard drive and you pop in a TPU without needing to do any physical re-architecture.
Wow.
That's amazing.
That's the most Googlingy infrastructure store.
since the cork boards.
Exactly.
Also, all of this didn't happen in Mountain View.
It was at a Google satellite office in Madison, Wisconsin.
Whoa.
Yes.
Why Madison, Wisconsin?
There was a particular professor out of the university,
and there was a lot of students that they could recruit from.
Wow.
Yeah, I mean, it was probably them or epic.
Where are you going to go work?
Yeah.
Wow.
They also then just kept this a secret.
Right.
Why would you tell anybody about this?
Because it's not.
like they're offering these in Google Cloud, at least at first, and why would you want to tell
the rest of the world what you're doing? So the whole thing was a complete secret for at least a
year before they announced it at Google I.O. So really crazy. The other thing to know about the
TPUs is they were done in time for the AlphaGo match. So that match ran on a single machine
with four TPUs in Google Cloud. And once that worked, obviously, that gave Google a little bit
of extra confidence to go really, really reproduction. So that's the TPU. V1.
by all accounts was not great.
They're on V7 or V8 now.
It's gotten much better.
TPUs and GPUs look a lot more similar than they used to
than they've sort of adopted features from each other.
But today, Google, it's estimated,
has 2 to 3 million TPUs for reference.
NVIDIA shipped, people don't know for sure,
somewhere around 4 million GPUs last year.
So people talk about AI chips,
like it's this just, oh, one-horse race with NVIDIA.
Google has like an almost NVIDIA scale
internal thing making their own ships at this point
for their own and for Google Cloud customers.
The TPU is a giant deal in AI
in a way that I think a lot of people don't realize.
Yep, this is one of the great ironies
and maddening things to OpenAI and Elon Musk
is that OpenAI gets founded in 2015
with the goal of, hey, let's shake all this talent
out of Google and level the playing field.
And Google just accelerates.
Right.
They also build TensorFlow.
that's the framework that Google Brain built
to enable researchers to build and train and deploy
machine learning models. And they built it in such a way
that it doesn't just have to run on TPUs. It's super portable
without any rewrites to run on GPUs or even CPUs
to. So this would replace the old distbelief system
and kind of be their internal and external
framework for enabling ML researchers going forward.
So somewhat paradoxically,
during these years after the founding of OpenA.
Yes, some amazing researchers are getting siphoned off from Google and Google Brain.
But Google Brain is also firing on all cylinders during this time frame.
Delivering on the business purposes for Google left and right.
Yes. And pushing the state of the art forward in so many areas.
And then in 2017, a paper gets published from eight researchers on the Google Brain team.
Kind of quietly.
These eight folks were obviously very very.
excited about the paper and what it described and the implications of it, and they thought it
would be very big. Google itself, uh, cool. This is like the next iteration of our language model
work. Great. Which is important to us. But are we sure this is the next Google? No. No.
There are a whole bunch of other things we're working on that seem more likely to be the next
Google. But this paper and its publication would actually be what gave OpenAI the opportunity
to build the next Google. To grab the ball and run with it and build the next Google. Because this
is the transformer paper. Okay, so where did the transformer come from? Like, what was the latest thing
that language models had been doing at Google? So coming out of the success of Franz Ox's work on
Google Transly and the improvements
that happened there. In like the late
2000s-ish, 2007? Yeah, mid
to late 2000s. They keep
iterating on Transly and
then once Jeff Hidden
comes on board and AlexNet happens
they switch over to a neural
network-based language model for
Translate. Which was
dramatically better and like
a big crazy cultural thing
because you've got these researchers parachuting
in, again led by Jeff Dean
saying, I'm pretty sure
our neural networks can do this way better than the classic methods that we've been using
for the last 10 years. What if we take the next several months and do a proof of concept?
They end up throwing away the entire old code base and just completely wholesale switching
to this neural network. There's actually this great New York Times magazine story that ran in
2016 about it. And I remember reading the whole thing with my jaw on the floor. Like,
wow, neural networks are a big effing deal. And this was the year before
the transformer paper one come out.
Before the transformer paper, yes.
So they do the rewrite of Google Translate,
make it based on recurrent neural networks,
which were state of the art at that point in time,
and it's a big improvement.
But as teams within Google Brain and Google Translate keep working on it,
there's some limitations.
And in particular, a big problem was that they, quote, unquote,
forgot things too quickly.
I don't know if it's exactly the right analogy,
but you might say in sort of like today's transformer,
or world speak, you might say that their context window was pretty short.
As these language models progressed through text, they needed to sort of remember everything
they had read so that when they need to change a word later or come up with the next word,
they could have a whole memory of the body of text to do that.
So one of the ways that Google tries to improve this is to use something called long,
short-term memory networks or LSTMs, as the acronym that people use for this. And basically,
what LSTMs do is they create a persistent or long short-term memory. You got to use your brain
a little bit here for the model so that it can keep context as it's going through a whole bunch of
steps. And people were pretty excited about LSTMs at first. People are thinking like, oh, LSTMs are
what are going to take language models and large language models mainstream.
Right.
And indeed, in 2016, they incorporated into Google Translate, these LSTMs.
It reduces the error rate by 60%.
Huge jump.
Yep.
The problem with LSTMs, though, they were effective, but they were very computationally intensive.
And they didn't parallelize that great.
All the efforts that are coming out of AlexNet and then the TPU project,
of paralyzation. This is the future. This is how we're going to make AI really work. LSTMs are a bit of a roadblock here. Yes. So a team within Google Brain starts searching for a better architecture that also has the attractive properties of LSTMs that it doesn't forget context too quickly, but can parallelize and scale better.
Take advantage of all these new architectures.
Yes. And a researcher named Jakob Oskarite had been toying around with the idea of
broadening the scope of quote-unquote attention in language processing. What if rather than
focusing on the immediate words, instead, what if you told the model, hey, pay attention to the
entire corpus of text, not just the next few words, look at the whole thing. And then based on that
entire context and giving your attention to the entire context, give me a prediction of what the next
translated word should be. Now, by the way, this is actually how professional human translators
translate text. You don't just go word by word. I actually took a translation class in college,
which was really fun. You read the whole thing of the original in the original language. You get
and understand the context of what the original work is, and then you go back and you start to
translate it with the entire context of the passage in mind.
So it would take a lot of computing power for the model to do this,
but it is extremely parallelizable.
So Jakub starts collaborating with a few other people on the brain team.
They get excited about this.
They decide that they're going to call this new technique,
the transformer.
Because one, that is literally what it's doing.
It's taking in a whole chunk of information, processing,
understanding it, and then transforming it,
and B, they also love Transformers as kids.
That's not why they named it the Transformer.
And it's taking in the giant corpus of text
and storing it in a compressed format, right?
Yeah.
I bring this up because that is exactly how you pitched
the Micro Kitchen conversation with Noam Shazir
in 2000, 2001, 17 years earlier.
Who is a co-author on this paper?
Yes. Well, so speaking of Noam Shazir,
he learns about this project, and he decides, hey, I've got some experience with this.
This sounds pretty cool. LSTMs definitely have problems. This could be promising.
I'm going to jump in and work on it with these guys. And it's a good thing he did, because before
Nome joined the project, they had a working implementation of the Transformer, but it wasn't actually
producing any better results than LSTMs. Nome joins the team, basically pulls a
Jeff Dean, rewrites the entire code base from scratch, and when he's done, the transformer
now crushes the LSTM-based Google Translate solution.
And it turns out that the bigger they make the model, the better the results get.
It seems to scale really, really, really well.
Stephen Levy wrote a piece in Wired about the history of this.
And there are all sorts of quotes from the other members of the team just littered all over
this piece with things like, Nome is a magician.
Nome is a wizard.
Nome took the idea and came back and said, it works now.
Yeah.
And you wonder why Nome and Jeff Dean are the ones together, working on the next
version of Gemini now.
Yes.
Nome and Jeff Dean are definitely two pieces in a pod here.
Yes.
So we talked to Greg Carrotto from Google Brain, one of the founders of Google Brain, and it was a
really interesting conversation because he,
underscored how elegant the transformer was. And he said it was so elegant that people's response was
often, this can't work. It's too simple. Transformers are barely a neural network architecture.
Right. It was another big change from the Alex Net, Jeff Hinton lineage neural networks.
Yeah. It actually has changed the way that I look at the world because he pointed out that in nature,
this is Greg, the way things usually work is the most energy efficient way they could work.
Almost from an evolution perspective, that the most simple, elegant solutions are the ones that
survive because they are the most efficient with their resources. And you can kind of port this
idea over to computer science too, that he said he's developed a pattern recognition inside of
the research lab to realize that you're probably onto the right solution when it's really simple
and really efficient versus a complex idea. It's very clever. I think it's very true. You know how when you
siter and you have a thorny problem and you debate and you whiteboard and you come
with all and then you're like oh my god oh my god it's so simple and that ends up being the right
answer yeah there's an elegance to the transformer yes and that other thing that you touched on there
this is the beginning of the modern AI just feed it more data the famous piece the bitter lesson
by rich Sutton wouldn't be published until 2019 for anyone who hasn't read it it's basically
we always think as AI research we're so smart
and our job is to come up with another great algorithm
but effectively in every field from language
to computer vision to chess
you just figure out a scalable architecture
and then the more data wins
just these infinitely scaling
more data, more compute, better results
yes and this is really the start
of when that starts to be like
oh we have found the scalable architecture
that will go at so far for
I don't know close to a decade
of just more data in, more energy, more compute, better results.
So the team and Noam, like, yo, this thing has a lot, a lot of potential.
This is more than better translate. We can really apply this.
Yeah, this is going to be more than better Google translate.
The rest of Google, though, definitely slower to wake up to the potential.
They build some stuff. Within a year, they build Burt, the large language model.
Yes, absolutely true. It is a third.
false narrative out there that Google did nothing with the transformer after the paper was published.
They actually did a lot. In fact, Bert was one of the first LLMs. Yes, they did a lot with transformer-based
large language models after the paper came out. What they didn't do was treated as a wholesale technology
platform change. Right. They were doing things like Bert and Mom, this other model. You know,
they could work it into search results quality. And I think that did meaningfully move the needle,
even though Google wasn't bragging about it and talking about it,
they got better at query comprehension.
They were working it into the core business
just like every other time Google Brain came up with something great.
Yeah.
So in perhaps one of the greatest decisions ever for value to humanity
and maybe one of the worst corporate decisions ever for Google,
Google allows this group of eight researchers to publish the paper.
Under the title, attention is all you need,
obviously a nod to the classic Beatle song of L.
love. As of today in 2025, this paper has been cited over 173,000 times in other academic papers,
making it currently the seventh most cited paper of the 21st century. And I think all of the other
papers above it on the list have been out much longer. Wow. And also, of course, within a couple
years, all eight authors of the Transformer paper had left Google to either start or join AI
startups, including Open AI.
Brutal. And of course, Noam starting Character AI, which would we call it a haquisition,
he would end up back at Google via some strange licensing and IP and hiring agreement
on the few billion dollars order. Very, very expensive mistake on Google's part.
is fair to say that 2017 begins the five-year period of Google not sufficiently seizing the
opportunity that they had created. With the Transformer, yes. So speaking of seizing opportunities,
what is going on at Open AI during this time? And does anyone think the Transformers a big deal
over there? Yes, yes, they did. But here's where history gets really, really crazy.
Right after Google publishes the Transformer paper, in September of 2017,
Elon gets really, really fed up with what's going on at Open AI.
There's like seven different strategies.
Are we doing video games?
Are we doing competitions?
What's the plan?
What is happening here?
As best as I can tell, all you're doing is just trying to copy deep mind.
Meanwhile, I'm here building SpaceX and Tesla.
Self-driving is becoming more and more clear as critical to the future of Tesla.
I need AI researchers here
and I need great AI advancements to come out
to help what we're doing at Tesla.
Open AI isn't cutting it.
So he makes an ultimatum to Sam and the rest of the OpenAI board.
He says, I'm happy to take full control of OpenAI
and we can merge this into Tesla.
I don't even know how that would be possible
to merge a nonprofit into Tesla.
But in Elon Land, if he takes over as CEO of OpenAI, it almost doesn't matter.
We're just treating it as if it's the same company anyway, just like we do with the deals with all of my companies.
Right. Or he's out completely, along with all of his funding.
And Sam and the rest of the board are like, no.
And as we know now, they're sort of calling capital into the business.
It's not like they actually got all the cash up front.
Right. So they're only $130 million-ish into the billion dollars of commitment.
they don't reach a resolution, and by early 2018, Elon is out, along with him the main source of Open AI's funding.
So either this is just a really, really, really bad misjudgment by Elon, or the sort of panic that this throws Open AI into is the catalyst that makes them reach for the transformer and say,
all right, we got to figure things out.
Necessities the mutter of invention.
Let's go for it.
It's true.
I don't know if during this personal tension between Elon and Sam
if they had already decided to go all in on Transformers or not.
Because the thing you very quickly get to,
if you decide transformers, language models,
we're going all in on that,
you do quickly realize you need a bunch of data,
you need a bunch of compute,
you need a bunch of energy,
and you need a bunch of capital.
And so if your biggest backer is walking away, the 3D chess move is, oh, we got to keep him because we're about to pivot the company and we need his capital for this big pivot we're doing.
The 4D chess is if he walks away, maybe I can turn it into a for-profit company and then raise money into it and eventually generate enough profits to fund this extremely expensive new direction we're going in.
I don't know which of those it was.
Yeah, I don't know either.
I suspect the truth is it's some of both.
Yes.
But either way, how nuts is it that A, these things happen at the same time, and B, the company wasn't burning that much cash, and then they decided to go all in on we need to do something so expensive that we need to be a for-profit company in order to actually achieve this mission because it's just going to require hundreds of billions of dollars for the far foreseeable future.
Yep. So in June of 2018, OpenAI releases a paper describing how they have taken the transformer
and developed a new approach of pre-training them on very large amounts of general text on the internet
and then fine-tuning that general pre-training to specific use cases. And they also announced that they have trained and run the first
proof-of-concept model of this approach, which they are calling GPT1, generatively pre-trained transformer
version 1, which we should say is right around the same time as Burt and right around the same
time as another large language model based on the transformer out of here in Seattle, the Allen
Institute. Yes, indeed. So it's not as if this is heretical and a secret, other AI labs, including
Google's own is doing it. But from the very beginning, OpenAI seemed to be taking this more
seriously given the cost of it would require betting the company if they continued down this path.
Yep. Or betting the nonprofit. Betting the entity. Yes. We're going to need some new terminology here.
Yes. So Elon's just walked out the door. Where are they going to get the money for this?
Sam turns to one of the other board members of Open AI, Reid Hoffman. Read just a year or so earlier
had sold LinkedIn to Microsoft,
and Reid is now on the board of Microsoft.
So Reed says, hey, why don't you come talk to Satya about this?
Do you know where he actually talks to Satya?
Oh, I do. Oh, I do.
In July of 2018, they set a meeting for Sam Altman and Satya and Adela
to sit down while they're both at the Allen & Company Sun Valley Conference in Sun Valley, Idaho.
It's perfect.
And while they're there, they hash out a deal for Microsoft.
to invest $1 billion into OpenAI in a combination of both cash and Azure cloud credits.
And in return, Microsoft will get access to OpenAI's technology,
get an exclusive license to OpenAI's technology for use in Microsoft's products.
And the way that they will do this is OpenAI, the nonprofit,
will create a captive for-profit entity called OpenAILP,
controlled by the nonprofit OpenAI. Inc.
And Microsoft will invest into the captive for-profit entity.
Reid Hoffman joins the board of this new structure along with Sam,
Ilya, Greg Brockman, Adam D'Angelo, and Tasha McCauley.
And thus the modern OpenAI for-profit, non-profit, question mark, is created.
The thing that's still being figured out, even today here in 2025, is created.
This is like the complete history of AI.
This is not just the Google AI episode.
Well, these things are totally inextricable.
And I was just going to say, this is the Google Part 3 episode.
Microsoft, they're back.
Microsoft is Google's mortal enemy.
Yes.
That in our first episode on the founding of Google and search,
and then in the second episode on Alphabet and all the products that they made,
the whole strategy at Google was always about Microsoft.
They finally beat them on every single whole.
front and here they are showing up again saying what was sate's line we just want to see them dance
i think the line that would come a couple years later is we want the world to know that we made
google dance oh man but this is all still pre chat gptt this is just sam lining up the financing
he needs for what appears to be a very expensive scaling exercise they're about to embark on
with GPT2 and onward.
Yep. And this is the right time to talk about why, from OpenAI's perspective, Microsoft is the absolute perfect partner.
It's not just that they have a lot of money.
Although that helps.
I mean, that helps. That helps a lot.
But more important than money, they have a really, really great public cloud.
Azure.
Yes.
OpenAI is not going to go buy a bunch of Nvidia GPUs and then build their own data center here at this point in 20.
2018, that's not the scale of company that they are. They need a cloud provider in order to actually do all the compute that they want to do. If they were back at Google and these researchers are doing it, great, then they have all the infrastructure. But OpenAI needs to tie themselves to someone with the infrastructure.
And there's basically only two non-Google options. They're both in Seattle. And hey, one of them in Microsoft is really interested, also has a lot of cash.
It seems like a great partnership.
That's true.
I wonder if they did talk to AWS at all about it.
Because I think, this is a crazy Easter egg,
I hesitate to say it out loud,
but I think AWS was actually in the very first investment
with Elon in OpenAI.
Oh, wow.
And I don't know if it was in the form of credits
or what the deal was,
but I'd seen it reported a couple places
that AWS actually was in that nonprofit round.
Yeah, in the nonprofit funding.
the donations to the early open AI.
Anyway, Microsoft's OpenAI, they end up tying up.
A match made in heaven.
Satya and Sam are on stage together talking about how this amazing partnership and marriage
has come together and they're off to model training.
Yeah.
And this paves the way for the GPT era of OpenAI.
But before we tell that story.
Yes.
Now is a great time to thank one of our favorite companies,
Shopify. Yes, and this is really fun because we have been friends and fans of Shopify for years.
We just had Toby on ACQ2 to talk about everything going on in AI and everything that has happened at Shopify in the six years now since we covered the company on Acquired.
It's been a pretty insane transformation for them.
Yeah. So back at their IPO, Shopify was the go-to platform for entrepreneurs and small businesses to get online.
What's happened since is that is still true, and Shopify has also become the world's leading commerce platform for enterprises of any size, period.
Yeah, so what's so cool about the company is how they've managed to scale without losing their soul.
Even though companies like Everlane and Vorey and even older established companies like Mattel are doing billions of revenue on Shopify, the company's mission is still the same as the day Toby founded it to create a world where more entrepreneurs exist.
Oh, yeah. Ben, you got to tell everyone your favorite enterprise brand that is on Shopify.
Oh, I'm saving that for next episode. I have a whole thing planned for episode two of this season.
Okay, okay, great. Anyway, the reason enterprises are now also using Shopify is simple,
because businesses of all sizes just sell more with Shopify. They built this incredible ecosystem
where you can sell everywhere. Obviously, your own site, that's always been true.
but now with Shopify you can easily sell on Instagram, YouTube, TikTok, Roblox,
Roku, ChatGPT, perplexity, anywhere.
Plus, with ShopPay, their accelerated checkout,
you get amazing conversion,
and it has a built-in user base of 200 million people
who have their payment information already stored with it.
Shopify is the ultimate example of not doing what doesn't make your beer taste better.
Even if you're a huge brand, you're not going to build a better
e-commerce platform for your product. But that is what Toby and Shopify's entire purpose is. So you should
use them. Yes. So whether you're just getting started or already at huge scale, head on over to
Shopify.com slash acquired. That's S-H-O-P-I-F-Y.com slash acquired. And just tell them that Ben and David
sent you. All right. So what will be? In G-P-T-2, is that what's being trained right here?
Yes. G-B-T-2. This was the first time I heard about it. Data scientists around Seattle.
We're talking about this cool.
Right.
So after the first Microsoft partnership, the first billion dollar investment, in 2019, OpenAI releases GPT2, which is still early, but very promising, that can do a lot of things.
A lot of things, but it required an enormous amount of creativity on your part.
You kind of had to be a developer to use it.
And if you were a consumer, there was a very heavy load put on you.
you had to go write a few paragraphs and then paste those few paragraphs into the language model,
and then it would suggest a way to finish what you were writing based on the source paragraphs.
But it wasn't interactive.
Yes, it was not a chat interface.
Yes.
There was no interface essentially for it.
It was an API.
But it can do things like obviously translate text.
I mean, Google's been doing that for a long time.
But GPT2, you could do stuff like make up a fake news headline and,
give it to GPT2 and it would write a whole article. You would read it and you'd be like,
uh, sounds like it was written by a bot. Yeah. But again, there was no front door to it for
normal people. You had to really be willing to wait in the muck to use this thing. So then,
the next year in June of 2020, GPT3 comes out. Still no front door, you know, user interface to
the model, but it's very good. GPT2 showed the promise of,
what was possible. GPT3, it's starting to be in the conversation of, can this thing pass the
Turing test. Oh, yeah. You have a hard time distinguishing between articles that GPT wrote and
articles that humans wrote. It's very good. And there starts to be a lot of hype around this thing.
And so even though consumers aren't really using it, the broader awareness is that there's something
interesting on the horizon. I think the number of AI pitch decks that VCs are seeing,
is starting to tick up around this time.
As is the NVIDIA stock price.
Yes.
So then in the next year, in the summer of 2021,
Microsoft releases GitHub copilot using GPP3.
This is the first, not just Microsoft product
that comes out with GPT baked into it,
but first productization.
Product anywhere.
Yeah, first productization of GPT.
Yes, I have any open AI technology.
Yeah, it's big. This starts a massive change in how software gets written in the world.
Slowly, then all at once. It's one of these things where at first, just a few software engineers and there was a lot of whispers of how cool is this? It makes me a little bit more efficient. And now you get all these comments. Like 75% of all companies code is written with AI.
Yep. So after that, Microsoft invests another $2 billion in Open AI, which seemed like a lot of money at the time. So that takes us to the end of 2021. There's an interesting.
kind of context shift that happens around here.
Yeah, the bottom falls out on tech stocks.
Crypto, the broader markets, really, everyone suddenly goes from risk on to risk off.
And part of it was war in Ukraine, but a lot of it was interest rates going up.
And Google gets hit really hard.
The high watermark was November 19th of 2021.
Google was right at $2 trillion of market cap.
About a year after that slide began, they were worth a trillion dollars, nearly a 50% drawdown.
Wow. So towards the end of 2022, leading up to the launch of ChatGPT, people I think are starting to realize Google's slow. They're slow to react to things. It feels like they're an old crusty company. Are they like the Microsoft of the 2000s where they haven't had a breakthrough product in a while?
People are not bright on the future of Google, and then chat GPT comes out.
Yeah. Wow. Which means if you were bullish on Google back then and contrarian, you could have invested at a trillion dollar market cap.
Which is interesting. Like in October of 21, the market was saying that the forthcoming AI wave will not be a strength for Google. Or maybe what it was saying is we don't even know anything about a forthcoming AI wave because people are talking about AI, but they've been talking about VR and they've been talking about crypto and they've been talking about all this.
frontier tech, and like, that's not the future at all. This company just feels slow and unadaptive.
And slow and unadaptive at that point in history, I think would have been a fair characterization.
They had an internal chatbot, right? Yes, they did. All right. So, before we talk about
chat GPT, Google had a chatbot. So, Nome Shazir, incredible engineer, re-architected the
transformer, made it work, one of the lead authors of the paper, storied career within Google, has all
of this sway, should have all of this
sway within the company. After the
transformer paper comes out,
he and the rest of the team are like,
guys, we can use this for
a lot more than Google Transly.
And in fact, the last paragraph of the paper...
Are you about to read the transformer paper?
Yes, I am. We are
excited about the future of attention-based
models and plan to apply them to
other tasks. We plan to
extend the transformer
to problems involving input and
output modalities other than
them text and to investigate large inputs and outputs such as images, audio, and video.
This is in the paper.
Wow.
Google obviously does not do any of that for quite a while.
Nome, though, immediately starts advocating to Google leadership.
Hey, I think this is going to be so big, the transformer, that we should actually consider
just throwing out the search index and the 10 Blue Links model and go all.
all in on transforming all of Google into one giant transformer model.
And then Nome actually goes ahead and builds a chatbot interface to a large transformer model.
Is this Lambda?
This is before Lambda, MENA is what he calls it.
And there is a chatbot in the late teens 2020 timeframe that Noma is built within Google.
that arguably is pretty close to ChatGPT.
Now, it doesn't have any of the post-training safety
that ChatGPT does, so it would go off the rails.
Yeah, someone told us that you could just ask it,
who should die, and it would come up with names for you
of people that should die.
It was not a shippable product.
It was a very raw, not safe, not post-trained chatbot and model.
Right.
But it exists.
within Google, and they didn't ship it.
And technically, not only did it not have post-training, it didn't have RLHF either, this
very core component of the models today, the reinforcement learning with human feedback,
that ChatGPT, I don't know if it had it in 3, but it did in 3.5, and it did for the launch
of ChatGPT.
Realistically, it wasn't launchable, even if it was an open AI thing because it was so bad,
but a company of Google's stature certainly could not take the risk.
So strategically, they have this working again.
against them. But aside from the strategy thing, there's two business model problems here.
One, if you're proposing drop the 10 blue links and just turn Google.com into a giant AI chatbot,
revenue drops when you provide direct answers to questions versus showing advertisers and letting
people click through to websites. That upsets the whole Apple cart. Obviously, they're thinking
about it now, but until 2021, that was an absolute non-starter to suggest something like that.
two, there were legal risks of sitting in between publishers and users.
I mean, Google at this point had spent decades fighting the public perception and court rulings
that they were disintermediating publishers from readers.
So there was like a very high bar internally, culturally to clear if you were going to do
something like this.
Even those info boxes that popped up that took until the 20 teens to make it happen,
those really were mostly on non-monetizable queries anyway.
So anytime that you were going to say, hey, Google's going to provide you an answer instead of 10 blue links, you had to have a bulletproof case for it.
Yeah.
And there was also a brand promise and trust issue, too. Consumers trusted Google so much.
For us, even today, you know, when I'm doing research for Acquired, we need to make sure we get something right.
I'm going to Google.
I look something up in Claude.
Yeah.
It gives me an answer.
I'm like, that's a really good answer.
and then I verify by searching Google that I can find those facts too
if I can't click through the sources on Cloud.
That's my workflow.
Which sort of sounds funny today, but it's important.
If you're going to propose replacing the 10 blue links with a chatbot,
you need to be really damn sure that it's going to be accurate.
Yes.
And in 2020, 2021, that was definitely not the case.
Arguably, still isn't the case today.
And there also wasn't a compelling
reason to do it because nobody was really asking for this product. Right. No knew and people in
Google knew that you could make a chatbot interface to a transformer-based LLM and that was a really
compelling product. The general public didn't know. Open AI didn't even really know. I mean,
GPT was out there. Do you know the story of the launch of chat GPT? Well, I think I do. I have it in my notes here.
All right. So they've got GPT 3.5.
it's becoming very, very useful.
Yeah, this is late 2020.
They've got 3.5.
But there's still this problem of how am I supposed to actually use it?
How does it productized?
And Sam just kind of says, we should make a chat bot.
That seems like a natural interface for this.
Can someone just make a chat?
And within like a week, internally...
Someone makes a chat.
They just turn calls to the chat GBT 3.5 API
into a product where you're just chatting with it.
And every time you kick off a chat message,
it just calls GPT 3.5 on the API.
And that turns out to be this magic product.
I don't think they expected it.
I mean, servers are tipping over.
They're working with Microsoft to try to get more compute.
They're cutting deals with Microsoft in real time
to try to get more investment,
to get more Azure credits,
or get advances on their Azure credits
in order to handle the incredible load
in November of 2022 that's coming in
of people wanting to use this thing.
they also just throw up a paywall randomly
because they thought that the business
was going to be an API business.
They thought that the projections
were all about how much revenue
they were going to do
through B-to-B licensing deals
and then they just realized
oh, there's all these consumers trying to use this,
put up a paywall to at least
dampen the most expensive use of this thing
so we can kind of offset the costs
or slow the rollout.
Right. This isn't Google Search,
you know, 89% gross margin stuff here.
Right.
So they end up having incredibly fast revenue takeoff just from the quick stripe paywall
that they threw up over a weekend to handle all the demand.
So to say that OpenAI had any idea what was coming would also be completely false.
They did not get that this would be the next big consumer product when they launched it.
Ben Thompson loves to call OpenAI the accidental consumer tech company, right?
Yes.
It was definitely accidental.
Now, there is actually another slightly different version of the motivation for launching the chat.
Is this the Dario?
Interface, yeah, the Dario and Anthropic version.
So Anthropic was working on what would become Claude, and rumors were out there and people
at Open AI got wind of like, oh, hey, Anthropic and Dario are working on a chat interface.
We should probably do one, too.
and if we're going to do one,
we should probably launch it before they launch theirs.
So I think that had something to do with the timing.
But again, I don't think anybody, including Open AI, realized
what was going to happen, which is, Ben, you alluded to it,
but to give the actual numbers.
On November 30th, 2020,
basically Thanksgiving.
OpenAI launches a research preview
of an interface to the new GPT3.5
called chat GPT. That morning on the 30th, Sam Altman tweets, today we launched chat GPT. Try talking
with it here. And then a link to chat.com. Within a week, less than a week, actually, it gets
one million users. By the end of the year, so, you know, one month later, December 31st, 2022,
it has 30 million users. By the end of the next month, by the end of January 23, so two,
months after launch, it crosses 100 million registered users, the fastest product in history
to hit that milestone. Completely insane. Completely insane. Before we talk about what that unleashes within
Google, which is the famous Code Red, to rewind a little bit back to Noam and the chat bot within
Google, Mina, Google does keep working on Mina. They,
develop it into something called Lambda,
which is also a chatbot, also internal.
I think it was a language model.
At this point in time, they still differentiated
between the underlying model brand name
and the application name.
Yes, Lambda was the model,
and then there also was a chat interface to Lambda
that was internal for Google use only.
Nome is still advocating to leadership,
we got to release this thing.
He leaves, in 2020,
and found a chatbot company, Character AI,
that still exists to this day,
and they raise a lot of money, as you would expect.
And then Google ultimately in 2024,
after ChatGPT launches, pays $2.7 billion, I think,
to do a licensing deal with Character AI,
the net of which Noam comes back to Google.
Yeah, I think Larry and Sergei were like,
if we're going to compete seriously,
obviously, we kind of need
them back and blank check to go
get him. Yeah. So
throughout 2021, 2022, Google's
working on the Lambda
model and then the chat interface to
it. In May of
2022, they do release something
that is available to the public called
AI Test Kitchen, which is a
AI product
test area where people can play around
with Google's internal AI
products, including the Lambda
chat interface. Yep. And
all fairness predates chat GPT.
Do you know what they do to Nerf chat so that it doesn't go too far off the rails?
This is amazing.
No.
For the version of Lambda chat that is in AI Test Kitchen, they stop all conversations after five turns.
So you can only have five turns of conversation with the chat bot.
And then it's just, and we're done for today.
Thank you.
Goodbye.
Oh, wow.
And the reason they did that was for safety of like, you know, if the more turns you had with it,
the more likely it would start to go off the rails.
And honestly, it was a fair concern.
I mean, this thing was not for public consumption.
And if you remember back a few years before, Microsoft released Tay, which was this crazy racist chat bot.
Yeah, they launched it as a Twitter bot, right?
And it was going off the rails on Twitter.
This was in 2016, I think.
Right.
Maximal impact of badness.
Yeah.
And so despite Google, all the way back in 2017, Sundar declared we are an AI-first company,
is being understandably very cautious
in real public AI launches,
especially on consumer-facing things.
Yep.
And as far as anyone else is concerned
before ChatGPT,
they are an AI-first company
and they're launching all this amazing AI stuff.
It's just within the vector
of their existing products.
Right.
So Chat-GPD comes out,
becomes the fastest product in history
to 100 million users.
It is immediately obvious
to Sundar, Larry, Sergei,
all of Google leadership
that this is an existential threat to Google.
ChatTPT is a better user experience
to do the same job function that Google Search does.
And to underscore this,
so if you didn't know it in November of 22,
you sure knew it by February of 23
because good old Microsoft,
our biggest, scariest enemy...
Oh, yeah.
announces a new Bing powered by OpenAI,
and Satya has a quote,
It's a new day for search.
The race starts today.
There's an announcement of a new AI-powered search page.
He says, we want to rethink what search was meant to be in the first place.
In fact, Google's success in the initial days came by reimagining what could be done in search.
And I think the AI era we're entering gets us to think about it.
This is the worst possible thing that could happen to Google that now Microsoft can actually challenge Google on their own turf, intent on the Internet,
with a legitimately different, better, differentiated product vector,
not what Bing was trying to do, copycat.
This is the full leapfrog, and they have the technology partnership to do it.
Or so everybody thinks at the moment.
Oh, my God. Terrifying.
This is when Satya says the quote in an interview around this launch with Bing.
I want people to know that we made Google dance.
Oh, boy.
Well, hey, if you come at the king,
You'd best not miss.
Right.
And this big launch kind of misses.
Yes.
So what happens in Google?
December 2020, even before the big launch, but after the chat GPT moment,
Sundar issues a code red within the company.
And what does that mean?
Up until this point, Google and Sundar and Larry and everyone had been thinking about AI
as a sustaining innovation in Clay Christensen's turn.
this is great for Google. This is great for our products. Look at all these amazing things that we're doing.
It further entrenches incumbents.
It further is entrenching our lead in all of our already leading products.
We can deploy more capital in a predictable way to either drive down costs or make our product experiences that much better than any startup could make.
Or monetize that much better. All the things. Once Chat CheapT comes out,
on a dime overnight,
AI shifts from being a sustaining innovation
to a disruptive innovation.
It is now an existential threat.
And many of Google's strengths
from the last 10, 15, 20 years
of all the AI work that's happened in the company
are now liabilities.
They have a lot of existing castles to protect.
That's right.
They have to run everything through a lot of filters
before they can decide
if it's a good idea to go try to out open AI, open AI.
Yeah. So this code read that Sundar issues to the company is actually a huge moment because what it means and what he says is we need to build and ship real native AI products ASAP.
This is actually what you need to do, the textbook response to a disruptive innovation as the incumbent.
you need to not bury your head in the sand, and you need to say, okay, we need to actually go build and ship products that are comparable to these disruptive innovators.
And you need to be laser operationally in all the details to try and figure out where is it that the new product is actually cannibalizing our old product and where is it that the new product can be complementary and just lean in to all the ways in which you can be complementary.
in all the different little scenarios.
And really what they've been trying to do,
this ballet from 2022 onward,
is protect the growth of search
while also creating the best AI experiences they can.
And so it's very clever the way that they do AI overviews
for some but not all queries.
And they have AI mode for some, but not all users.
And then they have Gemini, the full AI app,
but they're not redirecting Google.com to Gemini.
it's this like very delicate dance of protecting the existing franchise
while also building a hopefully non-cannibalizing as much as we can, new franchise.
Yeah. And you see them really going hard and I think building leading products
in non-search cannibalizing categories like video.
Right. V-O-3 or Nanobanana. These are things that don't in any way cannibalize the existing
franchise. They, in fact, use some of Google Strength, all the YouTube training data,
and stuff like that.
Yeah.
So what happens next?
As you might expect,
it gets worse before it gets better.
Code Red goes out December 2020.
BARD, baby.
Launch Bard.
Oh, boy.
Well, even before that, January 23,
when OpenAI hits 100 million registered users for ChatGPT,
Microsoft announces they are investing another $10 billion in OpenAI
and says that they now own 40,
99% of the for-profit entity. Incredible in and of itself. But then now think about this from the Google lens of Microsoft. Our enemy, they now arguably own. Obviously, in retrospect here, they don't own Open AI. But it seems at the time like, oh my God, Microsoft might now own Open AI, which is our first true existential threat in our history as a company.
Not great, Bob. So then February 2023, the Bing integration launches Satya has the quote,
about wanting to make Google Dance. Meanwhile, Google is scrambling internally to launch AI products
as fast as possible. So the first thing they do is they take the Lambda model and the chatbot
interface to it. They rebrand it as barred. They ship that publicly. And they release it immediately.
February 2023, ship it publicly. Available GA to anyone. Which maybe was the right move,
but God, it was a bad product. It was really bad. I didn't know the term at the time.
time our LHF, but it was clear it was missing a component of some magic that ChatGPT had,
this reinforcement learning with human feedback where you could really tune the appropriateness,
the tone, the voice, the sort of correctness of the responses. It just wasn't there.
Yep. So to make matters worse, in the launch video for Bard, a video. This is a choreographed,
pre-recorded video where they're showing conversations with Bard, Bard gives an inaccurate
factual response to one of the queries that they include in the video.
This is one of the worst keynotes in history.
After the Bard launch and this keynote, Google's stock drops 8% on that day.
And then, like we were saying, once the actual product comes out, it becomes clear.
It's just not good.
Yep.
And it pretty quickly becomes clear.
It's not just that the chatbot isn't good.
It's the model isn't good.
So in May, they replace Lambda with a new model from the brain team called POM.
It's a little bit better, but it's still clearly behind not only GPT 3.5, but in March of
2023, OpenAI comes out with GPT4, which is even better.
You can access that now through chat GPT.
And here is where Sundar makes two really, really big decisions.
Number one, he says, we cannot have.
two AI teams within Google anymore were merging brain and deep mind into one entity called Google
DeepMind. Which is a giant deal. This is in full violation of the original deal terms of
bringing DeepMind in. Yep. And the way he makes it work is he says, Demis, you are now
CEO of the AI division of Google, Google DeepMind. This is all hands on deck.
and you and DeepMind are going to lead the charge, you're going to integrate with Google Brain,
and we need to change all of the past 10 years of culture around building and shipping AI products within Google.
To further illustrate this, when Alphabet became Alphabet, they had all these separate companies,
but things that were really core to Google, like YouTube actually stayed a part of Google,
DeepMind was its own company. That's how separate this was.
was. They're working on their own models. In fact, those models are predicated on reinforcement learning.
That was the big thing that DeepMind had been working on the whole time. And so reading in between
the lines, it's Sundar looking at his two AI labs and going, look, I know you two two don't actually
get along that well. But look, I don't care that you had different charters before. I am taking
the responsibility of Google Brain and giving it to DeepMind, and DeepMind is absorbing the Google
brain team. I think that's what you should sort of read into it because as you look at where the
models went from here, they kind of came from DeepMind. Yep. There's a little bit of interesting
backstory to this too. So Mustafa, Suleiman, the third co-founder of DeepMind, at some point before
this, he became like the head of Google AI policy or something. He had already shifted over
to Brain and to Google. He stayed there for a little while, and then he ended up getting
close with who else, Reid Hoffman.
Remember, Reed is on the ethics board for DeepMind, and Mustafa and Reed leave and go found
inflection AI, which fast forward now into 2024 after the absolute insanity that goes
down at OpenAI in Thanksgiving, 2023 when Sam Altman gets fired over the weekend during Thanksgiving
and then brought back by Monday when all the team threatened to quit and go to Microsoft.
Open Eye loves Thanksgiving. Can't wait for this year.
They love Thanksgiving. Yeah. Gosh.
After all that, which certainly strains the Microsoft relationship,
remember, again, Reed is on the board of Microsoft.
Microsoft does one of these acquisition-type deals with inflection AI
and brings Mustafa in as the head of AI for Microsoft.
Crazy. Wild, right? Just wild. Crazy turn of events. Okay, so that first big decision that Sundar makes is unifying deep mind and brain. That was huge. Equally big, he says, I want you guys to go make a new model and we're just going to have one model. That is going to be the model for all of Google internally, for all of our AI products externally.
It's going to be called Gemini, no more different models, no more different teams, just one model for everything.
This is also a huge deal.
It's a giant deal, and it's twofold.
It's push and its pull.
It's saying, hey, if anyone's got a need for an AI model, you've got to start using Gemini.
But two, it's actually kind of the Google Plus thing, where they go to every team and they start saying, Gemini as our future, you need to start looking for ways to integrate Gemini into your product.
Yes, I'm so glad you brought up Google Plus.
This came up with a few folks I spoke to in the research.
Obviously, this is all playing out real time.
But the point a lot of people at Google made is
the Jev and I situation is very different than the Google Plus situation.
This is a technical thing, A, which has always been Google's wheelhouse.
But B, even more importantly, this is the rational business thing to do
in the age of these huge models.
Even for a company like Google, there are massive.
massive scaling laws to models.
The more data you put in, the better it's going to get, the better all the outputs are going to be.
And because of scaling laws, you need your models to be as big as possible in order to have the best performance possible.
If you're trying to maintain multiple models within a company, you're repeating multiple huge costs to maintain huge models.
You definitely don't want to do that.
You need to centralize on just one model.
Yeah, it's interesting.
There's also something to read into where at first it was the Gemini model underneath the Bard product.
Bard was still the consumer name.
Then at some point they said, no, we're just calling it all Gemini, and Gemini became the user-facing name also.
This pulls in my quintessence from the alphabet episode.
I know it's a little bit woo-woo, but with Google saying we're actually going to name the consumer service the name of the AI model.
they're sort of admitted to themselves, this product is nothing but technology.
There isn't productiness to do on top of it. It's just like Gmail. Gmail was technology.
It was fast search. It was lots of storage. It was use it in the web. The productiness wasn't
particular the way that like Instagram was all about the product. Gemini the model, Gemini
the chatbot says, we're just exposing our amazing breakthrough technology to you all and you get to
interface directly with it. Anthropologically looking from afar, it kind of feels like it's
that principle at work. I totally agree. I think it's actually a really important branding point
and sort of rallying point to Google and Google culture to do this. Right. All right. So this is all
the stuff going on in Google, 23-ish in AI. Before we catch up to the present, I have a whole other
branch of alphabet that has been a real bright spot for AI. Can I go there? Can I take this
off ramp, if you will? Can you take the wheel, so to speak? May I take the wheel? May I investigate
another bet? Yeah. Please tell us the Waymo story. Awesome. So we got to rewind back all the way to
2004, the DARPA Grand Challenge, which was created as a way to spur research into autonomous
ground robots for military use.
And actually, what it did for our purposes here today
is create the seed talent
for the entire self-driving car revolution
20 years later.
So the competition itself is really cool.
There is a 132-mile race course.
Now, mind you, this is 2004.
In the Mojave Desert that the cars have to race on,
it is a dirt road.
No humans are allowed to be in or interact with the cars.
They are monitored 100% remotely,
and the winner gets $1 million.
$1 million, which was a break from policy.
Normally, these are grants, not prize money,
so this needs to be authorized by an act of Congress.
The $1 million eventually felt comical,
so the second year they raised the pot to $2 million.
It's crazy thinking about what these researchers are worth today
that that was the prize for the whole thing.
So the first year in 2004 went fine.
There were some amazing tech demonstrations
on these really tight budgets, but ultimately
zero of the 100
registered teams finish the race.
But the next year, in 2005,
was the real special year.
The progress that the entire industry
made in those first 12 months
from what they learned is totally
insane. Of the 23
finalists that were entering the competition,
22 of them
made it past the spot
where the furthest team
the year before had made it.
The amount that the field advanced
in that one year is
insane. Not only that, five of those teams actually finished all 132 miles. Two of them were from
Carnegie Mellon, and one was from Stanford, led by a name that all of you will now recognize,
Sebastian Thrun. Mm-hmm. Indeed. This is Sebastian's origin story before Google. Now, as we said,
Sebastian was kind enough to help us with Pratt for this episode, but I actually learned most of this
from watching a 20-year-old Nova documentary
that is available on Amazon Prime Video
thanks to Brett Taylor
for giving us the tip
on where to find this documentary.
Yes, the hot research tip.
So what was special about this Stanford team?
Well, one, there's a huge problem
with noisy data
that comes out of all of these sensors.
You know, it's in a car in the desert
getting rocked around,
it's in the heat, it's in the sun.
So common wisdom, and what Carnegie Mellon did,
was to do as much as you possibly can
on the hardware to mitigate that.
So things like custom rigging and gimbals and giant springs to stabilize the sensors,
Cardigee Mellon would essentially buy a Hummer and rip it apart and rebuild it from the wheels
up. We're talking like welding and real construction on a car. The Stanford team did the
exact opposite. They viewed any new piece of hardware as something that could fail. And so in
order to mitigate risks on race day, they used all commodity cameras and sensors that they just
mounted on a nearly unmodified Volkswagen. So they only innovated in software,
and they figured they would just kind of come up with clever algorithms
to help them clean up the messy data later.
Very Googly, right?
Very Googly.
The second thing they did was an early use of machine learning
to combine multiple sensors.
They mounted laser hardware on the roof,
just like what other teams were doing,
and this is the way that you can measure texture and depth
of what is right in front of you.
And the data, it's super precise,
but you can't drive very fast
because you don't really know much about what's far away,
since it's this fixed field of view, it's very narrow.
Essentially, you can't answer that question of
how fast can I drive, or is there a turn coming up?
So on top of that, the way they solved it was
they also mounted a regular video camera.
That camera can see a pretty wide field of view,
just like the human eye,
and it can see all the way to the horizon,
just like the humid eye.
And crucially, it could see color.
So what it would do, this is like really clever.
They would use a machine learning algorithm,
in real time, in 2005,
this computer is like sitting in the middle of the car.
They would overlay the data from the lasers on top
onto the camera feed.
And from the lasers, you would know
if the area right in front of the car was okay to drive or not,
then the algorithm would look up in the frames
coming off the camera,
overlaid what color that safe area was,
and then extrapolate by looking further ahead
at other parts of the video frame
to see where that safe area extended to.
So you can figure it.
out your safe path through the desert.
That's awesome.
It's so awesome.
I'm imagining like a Dell PC sitting in the middle of this car in 2005.
It's not far off in the email that we send out.
We'll share some photos of it.
It could then drive faster with more confidence and it knew when turns were coming up.
Again, this is real-time onboard the camera.
2005 is wild on that tech.
So ultimately, both of these bets worked and the Stanford team won in super dramatic fashion.
they actually passed one of the Carnegie Mellon teams autonomously through the desert.
It's like this big dramatic moment in the documentary.
So you would kind of think, so then Sebastian goes to Google and builds Waymo.
No.
As we talked about earlier, he does join Google through that crazy,
please don't raise money from Benchmark and Sequoia and we'll just hire you instead.
But he goes and works on Street View and Project Ground Truth and co-founds Google X.
David, as you were alluding to earlier, this Project
chauffeur that would become Waymo is the first project inside Google X.
And I think the story, right, is that Larry came to Sebastian and was like,
yo, that's self-driving car stuff. Like, do it. And Sebastian was like, no, come on. That was
a DARPA challenge. And he was like, no, no, no, you should do it. He's like, no, no, that won't be
safe. There's people running around cities. I'm not just going to put multi-ton killer robots
on roads and go and potentially harm people. And Larry finally comes to him and says,
why? What is the technical reason that this is impossible? And Sebastian goes home, has a sleep on it,
and he comes in the next morning, and he goes, I realize what it was. I'm just afraid.
Such a good moment. So they start. He's like, there's not a technical reason. As long as we can take
all the right precautions and hold a very high bar on safety, let's get to work. So Larry then goes,
great, I'll give you a benchmark. So that way you know if you're succeeding.
He comes up with these 10 stretches of road in California that he thinks will be very difficult to drive.
It's about 1,000 miles, and the team starts calling it the Lowry 1000.
And it includes driving to Tahoe, Lombard Street in San Francisco, Highway 1 to Los Angeles, the Bay Bridge.
This is the bogey.
Yep.
If you can autonomously drive these stretches of road, pretty good indication that you can probably do anything.
Yep.
So they start the project in 2009 within 18 months.
This tiny team, I think they hired, I don't know, it's like a dozen people or something, they've driven thousands of miles autonomously, and they managed to succeed in the full Larry 1000 within 18 months.
Totally unreal how fast they did it. And then also totally unreal, how long it takes after that to productize and create the waymo that we know today.
Right. It's like the first 99% and then the second 99% that takes 10%.
10 years.
Self-driving is one of these really tricky types of problems where it's surprisingly
easy to get started, even though it seems like it would be an impossible thing.
But then there's edge cases everywhere, weather, road conditions, other drivers, novel
road layouts, night driving.
So it takes this massive amount of work for a production system to actually happen.
So then the question is, what business do we build?
What is the product here?
And there was what Sebastian wanted, which was highway assist, sort of the lowest stakes,
most realistic, let's make a better cruise control. There's what Eric Schmidt wanted, which is crazy. He proposed, oh, let's just go buy Tesla. And that'll be our starting place. And then we'll just put all of our self-driving equipment on all the cars. David, do you know what it would have cost to buy Tesla at the time? I think at the time that negotiations were taking place between Elon and Larry and Google, this was in the depths of the Model S production scaling woes.
I think Google could have bought the company for $5 billion.
That's what I remember.
It was $3 billion.
$3 billion.
Oh, my goodness.
Obviously, that didn't happen, but what a crazy alternative history that could have been.
Right.
I mean, I think if that had happened, DeepMind would not have gone down in the same way,
and probably Open AI would not have gotten founded.
Ooh, that's probably right.
I think that is obviously unprovable.
Right.
The counterfactuals that we always come up with on this show, you can't know.
Yeah.
seems more likely than not to me
that at a minimum, open AI would not exist.
Right. So then there was what Larry wanted to do,
option three, build robotaxies.
Yeah.
And ultimately, that is at least right now
what they would end up doing.
So we could do a whole episode about this journey,
but we will just hit some of the major points for the sake of time.
The big thing to keep in mind here,
neither Google nor the public really knew
if self-driving was something that could happen in the next two years from any given point
or take another 10. And just to illustrate it, for the first five years of project chauffeur,
it did not use deep learning at all. They did the Larry 1000 without any deep learning and then
went another three and a half years. Wow, that's crazy. Yeah. Yeah, totally illustrates you never
know how far away the end goal is. And this is a field that comes from,
the only way progress happens is through these series of breakthroughs,
and you don't know, A, how far the next breakthrough is.
Because at any given time, there's lots of promising things in the field,
most of which don't work out.
And then B, when there is a breakthrough, actually how much lift that will give you
over existing methods.
So anytime people are forecasting, oh, and AI, we're going to be able to do
X, Y, Z, and X years, it's a complete fool's errand.
Even the experts don't know.
Here are the big milestones.
2013, they started using convolutional neural nets.
They could identify objects.
They got much better perception.
capabilities. This 2013-2014 period is when Google found religion around deep learning. So this is
like right after the 40,000 GPUs rolled out. So they've actually got some hardware to start
doing this on now. 2016, they've seen enough technology proof that they think let's commercialize
this. We can actually spin this out into a company. So Waymo becomes its own subsidiary inside
of Alphabet. It's no longer a part of Google X anymore. 2017, obviously the Transformer comes out.
they incorporate some learnings from the transformer, especially around prediction and planning.
March of 2020, they raise $3.2 billion from folks like Silver Lake Canada Pension and Investment Board
Mubatala and Dresen Horowitz. And of course, the biggest check, I think, Alphabet, and I think
they're always the biggest check because Alphabet is still the majority owner, even after a bunch more
fundraisers. In October of 2020, they launched the first public commercial, no human behind the
driver seat thing in Phoenix. It's the first in the world. This is 11 years after succeeding
in the Larry 1000. And this is nuts. I had given up at this point. I was like, that's cute
that Waymo and all these other companies are trying to do self-driving. Seems like it's never
going to happen. And then they actually were doing a large volume of rides safely with consumers
and charging money for it in Phoenix. Then they bring it to San Francisco where for me and lots of people
in San Francisco. It is a huge part of life in the city here now. It's amazing. Yeah. Every time I'm down,
I love taking them. They're launching in Seattle soon. I'm pumped. Interestingly, they don't make
the hardware, so they use a Jaguar vehicle that from what I can tell is only in Waymo's. Like,
I don't know if anybody else drives that Jaguar or if you can buy it, but they're working on a
sort of van next. They have some next generation hardware. For anyone who hasn't taken it, it's an Uber,
but with no driver.
And that launched in June of 24.
Along the way there, they raised their quote-unquote series B, another $2.5 billion.
Then after the San Francisco rollout, they raised their quote-unquote series C, $5.6 billion.
This year in January, they were reportedly doing more in gross bookings than Lyft in San Francisco.
Wow.
I totally believe it.
I mean, it is the number one option in San Francisco that I and everybody I know, too,
always goes to for right healing. It's like try to get a Waymo. If there's not a Waymo available
anytime soon, then go down the stack. Like we're living in the future and how quickly we fail to
appreciate it. Yeah. And what's cool, I think for people who it hasn't come to their city and is not
part of their lives yet, it's not just that it's a cool experience to not have a driver behind
it. Pretty quickly that just fades. It's actually a different experience. So if I need to go somewhere
with my older daughter. I don't mind hailing a Waymo,
bringing the car seat, installing the car seat in the Waymo, and driving with my daughter.
And she loves it. We call it a robot car. And she's like, a robot car. I'm so excited.
Huh.
I would never do that with an Uber.
That's interesting.
To my dog. Whatever I need to go with my dog. Like, it's super awkward to hail an Uber and be like,
hey, I got my dog. Can the dog come in it? Not a big deal with a Waymo.
And then when you're in town. Yeah, we can actually have sensitive conversations in the car.
You can have phone calls.
It really is a different experience.
Yeah, that's so true.
Yeah, so may as well catch up to today.
They're operating in five cities, Phoenix, San Francisco, L.A., Austin, and Atlanta.
They have hundreds of thousands of paid rides every week.
They've now driven over 100 million miles with no human behind the wheel, growing at 2 million every week.
There's over 10 million paid rides across 2,000 vehicles in the fleet.
They're going to be opening a bunch more cities in the U.S. next year.
They're launching in Tokyo.
their first international city slowly and then all at once.
I mean, that's kind of the lesson here.
The technology, they really continued with that multi-sensor approach all the way from
the DARPA Grand Challenge, camera, LIDAR, they added radar, and actually they use audio
sensing as well.
And their approach is basically any data that we can gather is better because that makes
it safer.
So they have 13 cameras, four LIDAR, six radar, and the array of external microphones.
This is obviously way more expensive of a solution.
than what Tesla is just doing with cameras.
But Waymo's party line is they believe it is the only path to full autonomy
to hit the safety bar and regulatory bar that they're aiming for.
It seems like a really big line in the sand for them
any time you talk to somebody in that organization.
Yeah, and look, as a regular user of both products,
happy owner and driver of a Model Y in addition to regular Waybo user,
at least with the current instantiation of full self-driving on my Tesla,
vastly different products.
full self-driving on my Model Y is great.
I use it all the time on the freeway,
but I would never not pay attention.
Whereas every time I get in a Waymo,
it's almost like Google Search, right?
It's like I just trust that,
oh, this is going to be completely and totally safe,
and I'm sitting in the backseat, and I can totally tune out.
I think I trust my Model Y, FSD, more than you do,
but I get what you're saying.
And frankly, regulatoryly,
you are required to still pay attention in Tesla and not in the Waymo.
The safety thing is super real,
though. I mean, if you look at the numbers, over a million motor vehicle crashes cause
fatalities every year, or there's over a million fatalities. In the U.S. alone, over 40,000 deaths
occur per year. So if you break that down, that's 120 every day. That's like a giant cause of
death. Yes. The study that Waymo just released last month showed that they have 91% fewer crashes
with serious injuries or worse
compared to the average human driver
even controlled for the fact
that Waymo's right now
are only driving on city surface streets.
So they controlled it apples to apples
with human driving data
and it's a 91% reduction
in those serious, either fatality
or serious injury things.
Why aren't we all talking about this
all the time every day?
This is going to completely change the world
in a giant cause of death.
Yeah.
So while we're in Waymo land,
what do you think about doing some quick analysis?
Great. Because I've been scratching my head here of, what is this business? And then I promise we'll go back to the rest of Google AI and catch up to today. It is super expensive to operate, especially at early scale. The training is high. The inference is high. The hardware is high, et cetera, et cetera, et cetera. Also, the operations are expensive. Yes. And in fact, they're experimenting some cities. They actually outsource the operations. So the fleet is managed by, there's a rental car.
company in Texas that manages it, or they've partnered, I believe, with Lyft and with Uber and
different. So they're trying all sorts of O&O versus partnership models to operate it.
Yeah. And the operations are like, these are electric cars. They need to be charged. They need to be
cleaned. They need to be returned to depots. They need to be checked out. They need to have
sensors replaced. So the question is, what is the potential market opportunity? How big could
this business be? And there's a few different ways you could try to quantify it. One
total market size thing you could do is try to sum the entire automaker market cap today. And
that would be $2.5 trillion globally if you include Tesla or $1.3 trillion without. But Waymo's not
really making cars. So that's probably the wrong way to slice it. You could look at all the
ride sharing companies today, which might be a better comp because that's the business that Waymo is
actually in today. That's on the order of $300 billion, most of which is Uber. Yep. That's
addressable market cap today with ride sharing. Waymo's ambitions, though, are bigger than that.
They want to be in the cars that you own. They want to be in long-haul trucking. So they believe they can
grow the share of transportation because there's blind people that could own a car. There's elderly
people who could get where they need to go on their own without having a driver, that sort of thing.
So the most squishy, but I think the most interesting way to look at it is what is the value from all of
the reduction in accidents. Because that's really what they're doing. It's a product to replace
accidents with non-accidents. I think that's viable. But again, I would say as a regular user of
the product, it is a different and expanding product to human ride share. So your argument is
whatever number I come up with for reducing accidents, it's still a bigger market than that because
there's additional value created in the product experience itself. Yeah. Scoping just to ride share,
now that we have Waymo in San Francisco, I use Waymo in scenarios where I would never use an Uber or a Lyft.
Yeah, makes sense. So here's the data we have. The CDC released a report saying deaths from crashes in 2022 in the U.S. resulted in $470 billion in total costs, including medical costs and the cost estimates for lives lost, which is crazy that the CDC has some way of putting the costs on human life, but they do. So if you reduce crash,
is 10x, which is what Waymo seems to be saying in their data, at least for the serious crashes,
that's over $420 billion a year in total costs that we would save as a nation. Now, it's not
totally apples to apples, I recognize this, but that cost savings is more than Google does
today in revenue in their entire business. You could see a path to a Google-sized opportunity
for Waymo as a standalone company just through this analysis, as long as they figure out a way
to get costs down to the point where they can run this as a large and profitable business.
Yeah, it is an incredible 20-plus-year success story within Google.
The way I want to close it is the investment so far actually hasn't been that large.
When you consider this opportunity, they have burned somewhere in the neighborhood of $10 to $15 billion.
That's sort of why I was listing all the investments to get to this point.
Jump change compared to foundational models.
Dude, also, let's just keep it scoped in this sector.
That's one year of Uber's profits.
Wow. Seems like a good bet.
I used to think this was like some wild goose chase.
It now looks really, really smart.
Yep. Totally agree.
Also, that cost $10 to $15 billion is the profits that Google made last month.
Google.
Well, speaking of Google.
Google. Should we catch us up to today with Google AI? Yes. So I think where you were is the Gemini
launch. So Sundar makes these two decrees mid-20203. One, we're merging brain and deep mind
into one team for AI within Google. And two, we're going to standardize on one model,
the future Gemini and deep mind slash brain team. You go build it and then
everybody in Google, you're going to use it.
Not to mention, apparently Sergey Brin is like now back as an employee working on Gemini.
Yes.
Employee number.
Got his badge back.
Yeah, got his badge back.
So, once Sundar makes these decisions, Jeff Dean and Oriol Vinalis from Brain go over and team up with the
DeepMind team and they start working on Gemini.
I'm a believer now, by the way.
You got Jeff Dean working on it.
I'm in.
If you got Jeff Dean on it, it's probably going to work.
If you weren't a believer yet, wait until I'm going to tell you next.
Once they get Noam back, when they do the deal with character AI, bring him back into the fold.
Noam joins the Gemini team, and Jeff and Noam are the two co-technical leads for Gemini now.
Let's go.
Let's go.
So they actually announced this very quickly at the Google I.O. keynote in May 2023.
they announce Gemini, they announce the plans,
they also launch AI overviews in search,
first as a labs product,
and then later that becomes just standard
for everybody using Google search.
Which is crazy, by the way.
The number of Google searches that happen
is unfathomably large.
I'm sure there's a number for it,
but just think about,
that's about the highest level of computing scale that exists
other than high bandwidth things like streaming.
But just think about the instances of Google,
searches that happen, they are running an LLM inference on all of those, or at least as many as they're
willing to show AI overviews on, which I'm sure is not every query, but many.
A subset, but still a large, large number of Google search. I mean, I see them all the time.
Yep.
This is really Google immediately deciding to operate at AI speed. I mean, chat GPT happened in
November 30th, 2022. We're now in May 2023. All of these decisions have been made. All of these changes have
happened and they're announcing things at I.O.
And they're really flexing
the infrastructure that they've got. I mean, the fact
that they can go like, oh, yeah, sure, let's do inference on every
query. We're Google. We can handle it.
So a key part
of this new Gemini model
that they announce in May
2023 is it's going to be
multimodal. Again, this is one model
for everything. Text, images, video,
audio, one model. They
release it for
early public access
in December, 2020,
So also crazy, six months, they build it, they train it, they release it.
That is amazing.
Wild.
February 2024, they launched Gemini 1.5 with a 1 million token context window,
much, much larger context window than any other model on the market.
Which enables all sorts of new use cases.
There's all these people who were like, oh, I tried to use AI before,
but it couldn't handle my XYZ use case.
Now they can.
Yep.
The next year, February 2025, they released Gemini 2.0. March of 2025, one month later, they launched Gemini 2.5 Pro in experimental mode, and then that goes GA in June.
This is like Nvidia pace, how often they're shipping.
Yeah, seriously. And also, in March of 2025, they launch AI mode. So you can now switch over on Google.com to chatbot mode.
And they're split testing, auto-opting some people into AI mode to see what the response is.
This is the golden goose.
Yeah, the elephant is tap dancing here.
Yep.
Then there's all the other AI products that they launched.
So Notebook LM comes out during this period, AI-generated podcasts.
Which, does that sound like us to you?
It feels a little trained.
The number of texts that we got when that came out of, this must be trained on acquiring.
I do know that a bunch of folks on the notebook LM team are acquired fans, so I don't know if they trained on us.
And then there's the video, the image stuff, V-O-3, Nano Banana, Genie 3 that just came out recently.
This is insane.
This is a world builder based on prompts and videos.
Yeah.
You haven't actually used it yet, right?
You watch that hype video.
Yeah, I watched the video.
I haven't actually used it.
Yeah.
I mean, if it does that, that's unbelievable.
It's a real-time generative world build.
World builder. Yeah, you look right and it invents stuff to your right. I mean, you combine that with like a Vision Pro hardware. You're just living in a fantasy land. So they announced there are now 450 million monthly users of Gemini. Now, that includes everybody who's accessing nanobanana. Yeah, I can't believe this stat. This is insane. Even with recently being number one in the app store, it still feels hard to believe. Google's saying it, so it must be true. But I just wonder.
what are they counting as use cases of the Gemini app?
Right.
Certainly, everybody who's using Nanobanana is using Gemini.
But is it counting AI overviews or is it counting AI mode or is it counting something
where I'm like accidentally used, like, Meta said that crazy high number of people using
meta AI.
Right, right, right.
That was complete garbage.
That was people searching Instagram who accidentally hit a llama model that made some things happen
and they were like, oh, go away.
I actually am just looking for a user.
Is it really $450 million or is it $450 million?
or is it $450 million?
Yeah.
Good question.
Either way, going from zero
is crazy impressive
in the amount of time that they have done.
Especially given revenues
at an all-time high.
They seem to so far be,
at least in this squishy early phase,
able to figure out how to keep the core business going
while doing well as a competitor
in the cutting edge of AI.
Yeah, and to foreshadow a little bit
to we're going to do a bull and bear
here in a minute.
as we talked about in our alphabet episode, Google does have a history of navigating platform shifts incredibly well in the transition of mobile.
It's true. Definitely a rockier start here in the AI platform shift. Much rockier. But hey, look, I mean, if you were to lay out a recipe for how to respond, given the rocky start, be hard to come up with a much better slate of things than what they've done over the last two years.
Yeah. All right. Should I give us the snapshot of the business today?
Give us the snapshot of the business today. Oh, yeah. Also, by the way, the federal government decided they were a monopoly and then decided not to do anything about it because of AI.
Yeah. So between the time when we shipped our Alphabet episode and here with our Google AI episode, or our part two and part three, for those who prefer simpler naming schemes, yeah, there was a U.S. versus Google.
antitrust case, the judge first ruled that Google was a monopoly in internet search and then
did not come up with any material remedies. I mean, there are some, but I would call them immaterial.
They did not need to spin off Chrome, and they did not need to stop sending tens of billions of
dollars to Apple and others. In other words, yes, Google is a monopoly, and the cost of doing anything
about that would have too many downstream consequences on the ecosystem, so we're just going to let
them keep doing what they're doing. And one of the reasons that the judge cited of why they
weren't going to really take these actions is because of the race in AI. That because tens of billions
of dollars have funding have gone into companies like OpenAI and Anthropic and perplexity,
Google essentially has this new war to fight and we're going to leave it to the free market to do
its thing where it creates viable competition on its own and we're not going to hamstring Google.
Personally, I think this argument is a little bit silly.
I mean, none of these AI companies are generating net income.
And just because they've raised a huge amount of money, it doesn't mean that will last forever.
They'll all burn through their existing cash in a pretty short period of time.
And if the spigots ever dry up, Google doesn't have any self-sustaining competition right now, whether in their old search business or in AI.
It is all dependent on people believing that the opportunity is so large that they keep pouring tens of billions of dollars into these competitors.
Yep. Plenty of other folks have made the sort of glib comment, but there's merit to it of, hey, as flat-footed as Google was when chat GPT happened, if the outcome of this is they avoid a Microsoft-level distraction and damage to their business from a U.S. Federal Court monopoly judgment, worth it?
Well, there's a funny meme here that you could draw. You know that meme of someone pushing the domino and it knocking over some big,
wall later. Yeah. There's the domino of Ilya leaving Google to start open AI. And the downstream
effect is Google is not broken up. Yeah, right. Exactly. It actually saves Google. It actually
saves Google. It's totally wild. Totally wild. All right. So here's the business today. Over the last
12 months, Google has generated $370 billion in revenue. On the earnings side, they've generated
$140 billion over the last 12 months, which is more profit than any other tech company. And the only
company in the world with more earnings is Saudi Aramco. Let's not forget, Google is the best business
ever. And we also made the point at the end of the alphabet episode, even in the midst of all of this
AI era and everything that's happened over the last 10 years, the last five years, Google's
core business has continued to grow 5X since the end of our Alphabet episode in 2015, 2016.
Yeah.
Market cap, Google Surge passed their old peak of $2 trillion and just hit that $3 trillion mark earlier this month.
They're the fourth most valuable company in the world behind Nvidia, Microsoft, and Apple.
It's just crazy.
On their balance sheet, actually, I think this is pretty interesting.
I normally don't look at balance sheet as a part of this exercise, but it's useful.
And here's why, in this case, they have $95 billion in cash and marketable securities.
And I was about to stop there and make the point, wow, look how much cash and resources they have.
I'm actually surprised it's not more.
So it used to be $140 billion in 2021.
And over the last four years, they've massively shift from this mode of accumulating cash to deploying cash.
And a huge part of that has been the CAPEX of the AI data center buildout.
So they're very much playing offense in the way that Meta, my,
Microsoft and Amazon are in deploying that CAPEX. But the thing that I can't quite figure out is the
largest part of that was actually buybacks and they started paying a dividend. So if you're not
a finance person, the way to read into that is, yes, we still need a lot of cash for investing in the
future of AI and data centers, but we still actually had way more cash than we needed and we
decided to distribute that to shareholders. That's crazy. Best business of all time.
right? That illustrates what a crazy business, their core search ads business is, if they're saying
the most capital intense race in business history is happening right now, we intend to win it.
Yeah. And we have tons of extra cash lying around on top of what we think, plus a safety cushion
for investing in that CapEx race. Yeah. Yes. Wow. So there are two businesses that are worth
looking at here. One is Gemini to try to figure out what's happening there. And two is a brief
history of Google Cloud. I want to tell you the cloud numbers today, but it's probably worth actually
understanding how did we get here on Cloud. First on Gemini, because this is Google, and they have,
I think, the most obfuscated financials of any of the companies we've studied. They anger me the
most in being able to hide the ball in their financial statements. Of course, we don't know
Gemini's specific revenue. What we do know is there are over 150 million paying subscribers
to the Google One bundle. Most of that is on a very low tier. It's on like the $5 a month,
$10 a month. The AI stuff kicks in on the $20 a month tier where you get the premium AI features,
but I think that's a very small fraction of the $150 million today. I think that's what I'm on.
But two things to note. One, it's growing quickly, that $150 million is growing almost 50% year
over year, but two is Google has a subscription bundle that 150 million people are subscribed
to. And so I've kind of had it in my head that AI doesn't have a future as a business model
that people pay money for, that it has to be ad-supported like search. But, hey, that's
not nothing. That's like a... That's almost half of America. I mean, how many subscribers
does Netflix have? Netflix is in the hundreds of millions. Yeah. Spotify is now a quarter-billion.
billion, something like that.
We now live in a world where there are real scaled consumer subscription services.
I owe this insight to Shashir Maroto.
We chatted actually last night because I named dropped him in the last episode and then he heard it.
And so we reached out and we talked.
And that's made me do a 180.
I used to think if you're going to charge for something, your total addressable market shrunk by 90 to 99%.
But he kind of has this point that if you build a really compelling bundle and Google has the
digital assets to build a compelling bundle. Oh, my goodness, YouTube premium, NFL Sunday
ticket. Yes, stuff in the Play Store, YouTube music, all the Google One storage stuff.
They could put AI in that bundle and figure out through clever bundle economics a way to
make a paid AI product that actually reaches a huge number of paying subscribers.
Totally. So we really can't figure out how much money Gemini makes right now. Probably not profitable
anyway. So what's the point of even analyzing it?
Yeah. But okay, tell us the cloud story. So we intentionally did not include cloud in our
alphabet episode. Google Part 2, effectively. Google Part 2, yes. Because it is a new product and now
very successful one within Google that was started during the same time period as all the other
ones that we talked about during Google Part 2. But it's so strategic for AI.
Yes.
It is a lot more strategic now, in hindsight, than it looked when they launched it.
So just quick background on it, it started as Google App Engine.
It was a way in 2008 for people to quickly spin up a backend for a web or soon after a mobile app.
It was a platform as a service, so you had to do things in this very narrow, Googly way.
It was very opinionated.
You had to use this SDK.
You had to write it in Python or Java.
You had to deploy exactly the way they wanted you to deploy.
It was not a thing where they would say, hey, developer, you can do anything you want.
Just use our infrastructure.
It was opinionated.
Super different than what AWS was doing at the time and what they're still doing today, which the whole world eventually realized was right, which is cloud should be infrastructure as a service.
Even Microsoft pivoted Azure to this reasonably quickly, where it was like, you want some storage, we got storage for you.
You want a VM?
We got a VM for you.
You want some compute?
You want a database.
We got to.
fundamental building blocks. So eventually, Google launches their own infrastructure as a service in 2012, took four years. They launched Google Compute Engine, that they would later rebrand Google Cloud Platform. That's the name of the business today. The knock on Google is that they could never figure out how to possibly interface with the enterprise. Their core business, they made really great products for people to use, that they loved polishing. They made them all as self-service possible. And then the way they made money was from advertisers.
let's be honest, there's no other choice but to use Google search.
Right. It didn't necessarily need to have a great enterprise experience for their advertising
customers because they were going to come anyway. Right. So they've got this self-serve experience.
Meanwhile, the cloud is a knife fight. These are commodities. All about the enterprise. It's the lowest
possible price, and it's all about enterprise relationships and clever ways to bundle and being
able to deliver a full solution. You say solution. I hear gross margin.
Yes.
But yes, so Google out of their natural habitat in this domain.
And early on, they didn't want to give away any crown jewels.
They viewed their infrastructure as this is our secret thing.
We don't want to let anybody else use it.
And the best software tools that we have on it that we've written for ourselves, like Bigtable or Borg, how we run Google or disbelief, these are not services that we're making available on Google Cloud.
Yep.
These are competitive advantages.
Yes.
And then they hired the former president of...
Oracle. Thomas Kiryan. Yes. And everything kind of changed. So 2017, two years before he comes in,
they had $4 billion in revenue 10 years into running this business.
2018 is their first very clever strategic decision. They launched Kubernetes. The big insight here
is if we make it more portable for developers to move their applications to other clouds,
the world is kind of wanting multi-cloud here. Right.
We're the third place player.
We don't have anything to lose.
Yes.
So we can offer this tool
and kind of counterposition
against AWS and Azure.
We shift the developer paradigm
to use these containers.
They orchestrate on our platform
and then, you know,
we have a great service
to manage it for you.
It was very smart.
So this kind of becomes
one of the pillars of their strategy
is you want multi-cloud.
We're going to make that easy
and you can, sure,
choose AWS or Azure too.
It's going to be great.
So David, as you said,
the former president of Oracle,
Thomas Curion,
is hired.
in late 2018. You couldn't ask for a better person who understands the needs of the enterprise
than the former president of Oracle. This shows up in revenue growth right away. In 2020,
they crossed 13 billion in revenue, which was nearly tripling in three years. They hired like
10,000 people into the go-to-market organization. I'm not exaggerating that, and that's on a base of
150 people when he came in, most of which were seated in California, not
not regionally distributed throughout the world.
The funniest thing is, Google kind of was a cloud company all along.
They had the best engineers building this amazing infrastructure.
Right.
They had the products.
They had the infrastructure.
They just didn't have the go-to-market organization.
Right.
And the productization was all like Googly.
It was like for us, for engineers, they didn't really build things that led enterprises
build the way that they wanted to build.
This all changes.
2022, they hit 26 billion in revenue.
2023, they're like a real viable third cloud. They also flip to profitability in 2023. And today,
they're over $50 billion in annual revenue run rate. It's growing 30% year over year.
They're the fastest growing of the major cloud providers, 5X and 5 years. And it's really three
things. It's finding religion on how to actually serve the enterprise. It's leaning into this
multi-cloud strategy and actually giving enterprise developers what they want. And three,
AI has been such a good tailwind for all hyperscalers because these workloads all need to run in the cloud because it's giant amounts of data and giant's amount of compute and energy. But in Google Cloud, you can use TPUs, which they make a ton of, and everyone else is desperately begging NVIDIA for allocations to GPUs. So if you're willing to not use Kuda and build on Google Stack, they have an abundant amount of TPUs for you.
This is why we saved Cloud for this episode.
There are two aspects of Google Cloud that I don't think they first saw back when they started the business with App Engine, but are hugely strategically important to Google today.
One is just simply that cloud is the distribution mechanism for AI.
So if you want to play an AI today, you either need to have a great application, a great model,
a great chip or a great cloud. Google is trying to have all four of those. Yes. There is no other
company that has, I think, more than one. I think that's the right call. Think about the big
AI players. Nvidia. Chips. Kind of has a cloud, but not really. They just have chips. And they
have the best chips and the chips everyone wants, but chips. And then you just look around the rest of
the big tech companies. Meta right now, only an application. They're
completely out of the race for the frontier models at the moment. We'll see what they're hiring spree yields. You look at Amazon. Infrastructure. They have application. Maybe. I don't actually know if Amazon.com, I'm sure it benefits from LLMs in a bunch of ways.
Mainly it's cloud. Yes, cloud and cloud leader. Microsoft? Cloud. It's just cloud, right? They make some models, but... I mean, they've got applications, but yeah, cloud. Cloud. Apple? Nothing. Nothing.
AMD, just chips.
Yep.
Open AI model.
Anthropic model.
Yep.
Yep.
These companies don't have
their own data centers.
They are like making noise
about making their own chips,
but not really.
And certainly not at scale.
Google has scale data center,
scale chips,
scale usage of model.
I mean,
even just from Google.com queries
now on AI overviews.
And scale applications.
Yes.
Yeah.
They have all of the pillars.
of AI. And I don't think any other company has more than one. And they have the very most
net income dollars to lose. Right. So then there's the chip side specifically of this. If Google
didn't have a cloud, it wouldn't have a chip business. It would only have an internal chip business.
The only way that external companies, users, developers, model researchers could use TPUs would be
if Google had a cloud to deliver them because there's no way in hell that Amazon or Microsoft
are going to put TPUs from Google in their clouds.
We'll see.
We'll see, I guess.
I think within a year it might happen.
There are rumors already that some neoclouds in the coming months are going to have TPUs.
Interesting.
Nothing announced, but TPUs are likely going to be available in neocloud soon, which is an interesting thing.
Why would Google do that?
Are they trying to build an NVIDIA-type business where they make money selling chips?
I don't think so.
I think it's more that they're trying to build an E.
ecosystem around their chips the way that Kuta does. And you're only going to credibly be
able to do that if your chips are accessible and anywhere that someone's running their
existing workloads. Yep. It'd be very interesting if it happens. And, you know, look,
you may be right. Maybe there will be TPUs in AWS or Azure someday. But I don't think they
would have been able to start there. If Google didn't have a cloud and there weren't any way for
developers to use TPUs and start wanting TPUs.
Would Amazon or Microsoft be like, eh, you know, all right, Google, we'll take some of your TPUs,
even though no developer out there uses them.
Right.
All right.
Well, with that, let's move into analysis.
I think we need to do Bull and Bear on this one.
You have to this time.
Got to bring that back.
For these episodes in the present, it seems like we need to paint the possible futures.
Yes.
Bringing back Bull and Bear.
I love it.
Then we'll do Playbook, Powers, Quintessenceans, bring it home.
Perfect. All right. So here's my set of bull cases. Google has distribution to basically all humans as the front door to the internet. They can funnel that however they want. You've seen it with AI overviews. You've seen it with AI mode. Even though lots of people use chat GPT for lots of things, Google's traffic, I assume, is still essentially an all-time high, and it's a default behavior.
Yep. Powerful.
So that is a bet on implementation that Google figures out how to execute and build a great business out of AI, but it is still theirs to lose.
Yeah. And they've got a viable product. It's not clear to me that Gemini is any worse than Open AI or Anthropics products.
No, I completely agree. This is a value creation, value capture thing. The value creation is there in spades.
The value capture mechanism is still TBD. Yeah. Google's old value capture mechanism,
is one of the best in history.
So that's the issue at hand.
Let's not get confused
that it's not like a good experience.
It's a great experience.
Yeah, yeah.
Okay, so we've talked about the fact
that Google has all the capabilities
to win an AI, and it's not even close.
Foundational model, chips,
hyperscaler, all this with self-sustaining funding.
I mean, that's the other crazy thing
is you look at the clouds have self-sustaining funding,
and Vidia has self-sustaining funding.
None of the model makers have self-sustaining funding.
funding. So they're all dependent on external capital. Yeah. Google is the only model maker who has
self-sustaining funding. Yes. Isn't that crazy? Yeah. Basically, all the other large scale
usage foundational model companies are effectively startups. Yes. And Google's is funded by a money
funnel so large that they're giving extra dollars back to shareholders for fun. Yeah. Again,
we're in the whole case. Well, when you put it that way, yeah. A thing we didn't mention, Google has
incredibly fat pipes connecting all of their data centers.
After the dot-com crash in 2000,
Google bought all that dark fiber for pennies on the dollar,
and they've been activating it over the last decade.
They now have their own private backhaul network
between data centers.
No one has infrastructure like this.
Yep.
Not to mention, that serves YouTube.
They're fat pipes.
Which in and of itself is its own bulkcase for Google in the future.
That's a great point.
Yeah, Ben Thompson had a big article about this yesterday at the time of recording.
Yeah, that was like a mega bull case that Ben Thompson published this week, that it was an interesting point.
A text-based internet is kind of the old internet. It's the first instantiation of the internet because we didn't have much bandwidth.
The user experience that is actually compelling is video.
High-resolution video everywhere all the time.
We already live in the YouTube internet.
Right. And not only can they train models.
on really the only scale source of UGC media
across long form and short form.
But they also have that as the number two search engine,
this massive destination site.
So they previewed things like you'll be able to buy AI labeled
or AI determined things that show up in videos.
And if they wanted to,
they could just go label every single product in every single video
and make it all instantly shoppable.
It doesn't require any human work to do it.
They could just do it.
and then run their standard ads model on it.
That was a mind-expanding piece that Ben published yesterday,
or I guess if you're listening to this, a few weeks ago about that.
And then there's also all the video AI applications that they've been building like flow and V-O.
What is that going to do for generating videos for YouTube that will increase engagement and add dollars for YouTube?
Yep.
Going to work real well.
Yep.
they still have an insane talent bench,
even though they've bled talent here and there
and lost people.
They have also shown they're willing
to spend billions for the right people
and retain them.
Unit economics.
Let's talk about unit economics of chips.
Everyone is paying Nvidia
75, 80% gross margins,
implying something like a 4 or 5x markup
on what it costs to make the chips.
A lot of people refer to this as the Jensen tax
or the NVIDIA tax.
You can call it that,
you can call it good business,
about pricing power. You could call it scarcity of supply, whatever you want. But that is true.
Anyone who doesn't make their own chips is paying a giant, giant premium to Ambidia. Google has to
still pay some margin to their chip hardware partner, Broadcom, that handles a lot of the work to
actually make the chip interface with TSM. I have heard that Broadcom has something like a 50%
margin when working with Google on the TPU versus NVIDIA's 80%. But that's still a huge,
difference to play with. A 50% gross margin from your supplier or an 80% gross margin from your
supplier is the difference between a 2x markup and a 5x markup. Yeah, I guess that's right.
When you frame it that way, it's actually a giant difference of the impact to your cost.
So you might wonder appropriately, well, are chips actually the big part of the cost of the cost
of like the total cost of ownership of running one of these data centers or training one of
these models. Chips are the main driver of the cost. They depreciate very quickly. I mean,
this is at best a five-year depreciation because of how fast we are pushing the limits of what we can do
with chips, the needs of next generation models, how fast TSM is able to produce. Yeah. I mean,
even that is ambitious, right? If you think you're going to get five years of depreciation on AI
chips, five years ago, we were still two years away from chat GPT. Right. Or I think
about what Jensen said at, we were at GTC this year, he was talking about Blackwell, and he said
something about Hopper, and he was like, eh, you don't want Hopper. My sales guys are going to hate me,
but like, you really don't want Hopper at this point. I mean, these were the H-100s. This was the hot
chip just when we were doing our most recent Nvidia episode. Yes, things move quickly. Yes.
So, I've seen estimates that over half the cost of running an AI data center is the chips and the
associated depreciation. The human cost, that R&D, is actually a pretty high amount because hiring
these AI researchers and all the software engineering is meaningful. Call it 25 to 33%. The power is actually a
very small part. It's like 2 to 6%. So when you're thinking about the economics of doing what Google's
doing, it's actually incredibly sensitive to how much margin are you paying your supplier in the chips
because it's the biggest cost driver of the whole thing. So I was sanity checking some of this with
Gavin Baker, who's the partner at Atrides Management to prep for this episode. He's like a great
public equities investor who's studied this space for a long time. We actually interviewed him
at the NVIDIA GTC pregame show. And he pointed out, normally, like in historical technology
eras, it hasn't been that important to be the low-cost producer. Google didn't win because they
were the lowest-cost search engine. Apple didn't win because they were the lowest cost. You know,
that's not what makes people win. But this era might have
actually be different because these AI companies don't have 80% margins the way that we're
used to in the technology business, or at least in the software business. At best, these AI
companies look like 50% gross margins. So Google being definitively the low-cost provider of
tokens because they operate all their own infrastructure and because they have access to
low-markup hardware, it actually makes a giant difference and might mean that they are the winner
in producing tokens for the world.
Very compelling build case there.
That's a weirdly winding analytical bull case,
but it's kind of the,
if you want to really get down to it,
they produce tokens.
Yep.
I've got one more bullet point
to add to the bulkcase for Google here.
Everything that we talked about
in part two,
the alphabet episode,
all of the other products within Google,
Gmail, maps, docs, Chrome, Android.
that is all personalized data about you that Google owns that they can use to create personalized AI products for you that nobody else has.
Another great point. So really the question to close out the bullcase is, is AI a good business to be in compared to search?
Search is a great business to be in. So far, AI is not. But in the abstract, again, we're in the bull case, so I'll give you this. It should be. With traditional web search,
You type in two to three words.
That's the average query length.
And I was talking to Bill Gross, and he pointed out that in AI chat, you're often typing 20 plus words.
So there should be an ad model that emerges.
And ad rates should actually be dramatically higher because you have perfect precision.
Right.
You have even more intent.
Yes.
You know the crap out of what that user wants.
So you can really decide to target them with the ad or not.
And AI should be very good at targeting with the ad.
So it's all about figuring out the user interface, the mix of paid versus not, exactly what
this ad model is.
But in theory, even though we don't really know what the product looks like now, it should
actually lend itself very well to monetization.
And since AI is such a amazing, transformative experience, all these interactions that
were happening in the real world or weren't happening at all, like answers to questions
and being on time spent, is now happening in these.
So it seems like the pie is actually bigger for digital interactions than it was in the search era.
So again, monetization should kind of increase because the pie increases there.
And then you've got the bulkcase of Waymo could be its own Google-sized business.
I was just thinking that.
Yeah, that's scoping all of this to a replacement to the search market,
Waymo and potentially other applications of AI beyond the traditional search market.
could add to that. Right. And then there's the like galaxy brain bulk case, which is if Google actually
creates AGI, none of this even matters anymore. And like, of course it's the most valuable thing.
That feels out of the scope for an acquired episode. It's disconnected. Yes, agree. Bearcase.
So far, this is all fun to talk about, but then the product shape of AI has not lent itself well to ads.
So despite more value creation, there's way less value capture. Google makes something like
$400-ish dollars per user per year, just based on some napkin math in the U.S., that's a free service
that everyone uses, and they make $400-ish dollars a year. Who's going to pay $400 a year for access
to AI? It's a very thin slice of the population. Some people certainly will, but not every person in
America. Some people will pay $10 million, but right. So if you're only looking at the game on the field
today, I don't see the immediate path to value capture. And think about when Google launched in
1998, it was only two years before they had AdWords. They figured out an amazing value capture
mechanism instantly. Very quickly. Yep. Another bare case. Think back to Google launch in
1998. It was immediately, obviously, the superior product. Yes. Definitely not the case today.
No. There's four, five great products. Google's dedicated AI offerings in chatbot was initially the immediately obviously inferior product. And now it's arguably on par with several others.
Right. They own 90% of the search market. I don't know what they own of the AI market, but it ain't 90%, is it 25%. I don't know. But at steady state, it probably will be something like 25, maybe up to 50%, but this is going to be a market with several big players in it.
So even if they monetized each user, as great as they monetize it in search, they're just
going to own way less of them.
Yep.
Or at least it certainly seems that way right now.
Yes.
AI might take away the majority of the use cases of search.
And even if it doesn't, I bet it takes away a lot of the highest value ones.
If I'm planning a trip, I'm planning that in AI.
I'm no longer searching on Google for things that are going to land Expedia ads in my face.
Or health, another.
Oh, my gosh. Huge vertical. Hey, I think I might have something that reminds me of misothelioma. Is it that or not?
Right. Oh, where are you going to put the lawyer ads? Maybe you put them there. Maybe it's just an ad product thing. But these are very high value queries.
Former searches that those feel like some of the first things that are getting siphoned off to AI. Yep. Any other bear cases?
I think the only other bear case I would add is that they have the added challenge now of being the incumbent this time around. And people,
and the ecosystem isn't necessarily rooting for them
in the way that people were rooting for Google
when they were a startup.
And in the way that people were still rooting for Google
in the mobile transition,
I think the startups have more of the hearts and minds these days.
Right.
So I don't think that's quantifiable,
but is just going to make it all a little harder path to row
this time around.
Yep. You're right.
They had this incredible PR.
and public love tailwind the first time around.
Yep.
And part of that's systemic, too.
Like, all of tech and all of big tech is just generally more out of favor with the country
in the world now than it was 10 or 15 years ago.
There's more important.
It's just big infrastructure.
It's not underdogs anymore.
Yep.
And that affects the opening eyes and the anthropics and the startups, too, but I think to a lesser
degree.
Yeah, they had to start behaving like big tech companies really early in their life compared
to Google.
I mean, Google gave a Playboy interview during their quiet period of their IPO.
Times have changed.
Well, I mean, given all the drama at OpenAI, I don't know that I characterize them as acting like a mature company.
Fair, fair.
Company, entity, whatever they are.
Yes.
Yeah.
But point taken.
Well, I worked most of my playbook into the story itself.
So you want to do power?
Yeah, great.
Let's move on and do power.
Hamilton, Helmers, seven powers.
of Google here in the AI era. And the seven powers are scale economies, network economies,
counter positioning, switching costs, branding, quartered resource, and process power.
And the question is, which of these enables a business to achieve persistent differential
returns? What entitles them to make greater profits than their nearest competitor sustainably?
Normally we would do this on the business all up. I think for this episode we should try to
scope it to AI products. Yes, agreed.
usage of Gemini, AI mode, and AI overviews versus the competitive set of anthropic open
AI, perplexity, grok, meta-a-I, etc. Scale economies, for sure. Even more so in AI than
traditionally in tech. Yeah, they're just way better. I mean, look, they're amortizing the cost
of model training across every Google search. I'm sure it's some super distilled down model that's
actually happening for AI overviews. But think about how many inference tokens are generated for
the other model companies and how many inference tokens are generated by Gemini. They just are
amortizing that fixed training costs over a giant, giant amount of inference. I saw some
crazy chart. We'll send it out to email subscribers. In April of 24, Google was processing 10 trillion
tokens across all their services. In April of 25, that was almost 500 trillion. Wow. That's a 50x
increase in one year of the number of tokens that they're vending out across Google services through
inference. And between April of 25 and June 25, it went from a little under 500 trillion to a little
under one quadrillion tokens, technically 980 trillion, but they are now, because it's later in the summer,
definitely sending out maybe even multiple quadrillion tokens. Wow. Wow. So among all the other
obvious scale economies things of amortizing all the costs of their hardware, they are amortizing
the cost of training runs over a massive amount of value creation. Yeah. Scale economies must be
the biggest one.
I find switching costs
to be relatively low.
I use Gemini for some stuff,
then it's really easy to switch away.
That probably stops being the case
when it's personal AI
to the point that you're talking about
integrating with your calendar
and your mail and all that stuff.
Yeah, the switching costs
have not really come out
yet in AI products,
although I expect they will.
Yes.
They have within the enterprise, for sure.
Yep.
Network economies,
I don't think,
if anyone else is a Gemini user, it makes it better for me because they are sucking up the
whole internet, whether anyone's participating or not. Yep, I agree. I'm sure AI companies will develop
network economies over time. I can think of ways it could work, but yeah, right now, no. And
arguably for the foundational model companies, can't think of obvious reasons right now.
Where does Hamilton put distribution? Because that's a thing that they have right now that no one
else has. Despite ChatGBTGPT having the Kleenex brand, Google distribution is still unbelievable.
Is that a cornered resource? Cornered resource, I guess. Yeah. Definitely of that.
Yeah, Google Search is a cornered resource, for sure. Certainly don't have counter positioning.
They're getting counter positioned. Yeah. I don't think they have process power,
unless they were like coming up with the next transformer reliably, but I don't think we're necessarily
seeing that. There's great research being done at a bunch of different labs. Branding they have.
Yeah. Branding is a funny one, right? Well, I was going to say it's a little bit to my bare case point about they're the incumbent.
It cuts both ways, but I think it's net positive. Yeah, probably. For most people, they trust Google.
Yeah, they probably don't trust these. Who knows AI companies, but I trust Google. I bet that's actually stronger than any downsides, as long as they're willing to still release stuff on the cutting edge.
So to sum it up, it's scale economies is the biggest one. It's branding, and it's a cornered
resource. And potential for switching costs in the future. Yep, sounds right to me. But it's telling
that it's not all of them. In search, it was like very obviously all of them, or most of them.
Yep. Quite telling. Well, I'll tell you, after hours and hours, spending multiple months
learning about this company, my contests when I boil it all down, is just that this is the most
fascinating example of the innovator's dilemma ever. I mean, Larry and Sergei control the company.
They have been quoted repeatedly saying that they would rather go bankrupt than lose at AI.
Will they really? If AI isn't as good a business as search, and it kind of feels like,
of course it will be. Of course it has to be. It's just because of the sheer amount of value
creation. But if it's not, and they're choosing between two outcomes, one is fulfilling our mission
of organizing the world's information
and making it universally accessible
and useful
and having the most profitable tech company in the world,
which one wins.
Because if it's just the mission,
they should be way more aggressive on AI mode
than they are right now
and full flip over to Gemini.
It's a really hard needle of thread.
I'm actually very impressed
at how they're managing
to currently protect the core franchise,
but it might be one of these things
where it's being eroded away at the foundation in a way that just somehow isn't showing up
in the financials yet. I don't know. Yep. I totally agree. And in fact, perhaps influenced by
you, I think my quintessence is a version of that too. I think if you look at all the big tech
companies, Google, as unlikely as it seems, given how things started, is probably doing the best
job of trying to thread the needle with AI right now. And that is incredibly commendable
to Sundar and their leadership. They are making hard decisions like we're unifying deep mind
and brain, we're consolidating and standardizing on one model, and we're going to ship this stuff
real fast, while at the same time not making rash decisions. It's hard.
Rapid but not rash, you know.
Yes.
And obviously we're still in early innings of all this going on,
and we'll see in 10 years where it all ends up.
Yeah.
Being tasked with being the steward of a mission
and the steward of a franchise with public company shareholders
is a hard dual mission,
and Sundar and the company is handling it remarkably well,
especially given where they were five years ago.
Yep.
And I think this will be one of the most fascinating examples in history to watch it play out.
Totally agree.
Well, thus concludes our Google series for now.
Yes.
All right, let's do some carve-outs.
All right, let's do some carve-outs.
Well, first off, we have a very, very fun announcement to share with you all.
The NFL called us.
We're going to the Super Bowl, baby.
Acquired is going to the Super Bowl.
this is so cool.
It's the craziest thing ever.
The NFL is hosting a innovation summit
the week at the Super Bowl,
the Friday before Super Bowl Sunday.
The Super Bowl is going to be in San Francisco
this year in February.
And so it's only natural coming back to San Francisco
with the Super Bowl that the NFL should do
an innovation summit.
Yep.
And we're going to host it.
That's right.
So the Friday before,
there's going to be some great on-stableness
interviews and programming. Most of you, you know, we can't fit millions of people in a tidy
auditorium in San Francisco the week of the Super Bowl when every other venue has tons of stuff too.
So there will be an opportunity to watch that streaming online. And as we get closer to that date
in February, we will make sure that you all know a way that you can tune in and watch the
emceeing, interviewing, and festivities at hand Super Bowl week.
It's going to be an incredible, incredible day.
leading up to an incredible Sunday.
Yes.
Well, speaking of sport, my carve-out is I finally went and saw F-1.
It is great.
I highly recommend anyone go see it, whether you're an F-1 fan or not.
It is just beautiful cinema.
Amazing.
Did you see it in the theater?
I did see in the theater, yeah.
Wow.
I unfortunately missed the IMAX window, but it was great.
It's my first time being in a movie theater in a while.
And whether you watch it at home or whether you're watching the theater,
I recommend the theater.
But it's going to be a great surround sound experience,
wherever you are. I haven't been to the movie theater since the Ares Tour, which I think is just
more about the current state of my family life with two young children. Yes. My second one,
so if you're going to laugh, is the Travel Pro suitcase. Ah, this is the brand that pilots and flight
attendants use, right? Maybe, I think I've seen some of them use it. Usually they use something
higher end, like a Briggs and Riley or a Toomey or, you know, Travel Pro is not the most
high-end suitcase. But I bought two really big ones for some international travel that we were doing
with my two-year-old toddler. And I must say they're robust. The wheels glide really well. They're
really smooth. They have all the features you would want. They're soft shell so you can like really
jam it full of stuff. But it's also a thick amount of protection. So even if you do jam it full
of stuff, it's probably not going to break. This is approximately the most budget suitcase you could
buy. I mean, I'm looking at the big honkin international check bag version. It's $416 on Amazon right
now. I've seen it cheaper. They have great sales pretty often. Everything about this suitcase
checked lots of boxes for me. And I completely thought I would be the person buying the Ramoa suitcase
or the something very high end. And this is just perfect. So I think I may be investing in more
travel pro suitcases. More travel pro. Nice.
Nice. Well, I mean, hey, look, for family travel, you don't want nice stuff. Yeah. I mean, I bought it thinking, like, I'll just get something crappy for this trip. But it's been great. I don't understand why I wouldn't have a full lineup of travel pro gear. So this is my, like, budget pick gone right that I highly recommend for all of you.
I love how acquired is turning into the wire cutter here. That's it for me today. Great. All right. I have two carve outs. I have one carve out and then I have a update in my on
going Google carve-out saga. But first, my actual carve-out. It is the glue guys podcast.
Oh, it's great. Those guys are awesome. So great. Our buddy Ravi Gupta, partner at Sequoia,
and his buddies, Shane Badiye, the former basketball player and Alex Smith, the former
quarterback for the 49ers and the Kansas City Chiefs and the Redskins. Their dynamic is so great.
They have so much fun. Half of their episodes like us are just them. And then half of their
episodes are with guests. Ben and I, we went on it a couple weeks ago. That was really fun. When we were
on it, we were talking about this dynamic of some episodes do better than others and pressure for
episodes and whatnot. And the guys brought up this interview they did with a guy named Wright Thompson.
And they said, like, this is an episode. It's got like 5,000 listens. Nobody's listened to it.
It's so good. And the mentality that we have about it is not that we're embarrassed that nobody
listen to it, it's that we feel sorry for the people who have not yet listened to it because
it's so good. I was like, that is the way to think about. That's great. Your episodes. So here you
are. You're giving everyone the gift of... I'm giving everyone the gift because I then, I was like,
all right, well, I got to go listen to this episode. Ray Thompson, I didn't know anything about
him before. I probably read his work in magazines over the years without realizing it. He's the
coolest dude. He has the same accent as Bill Gurley. So listening to him sounds like listening to
Like, if Bill Gurley, instead of being a VC only wrote about sports and basically dedicated his whole life to understanding the mentality and psychology of athletes and coaches, it's so cool. It's so cool. It's a great episode.
Highly, highly, highly, highly recommend.
All right. Legitimately, I'm queuing that up right now.
Great. That's my carve out. And then my ongoing family video gaming saga.
Yeah. In Google Part 1, I said I was debating between the Switch 2 and the Steam Deck.
That's right. First, you got the Steam Deck because you decided your daughter actually wasn't old enough to play video games with you, so you just got the thing for you.
The update was, I went with the Steam Deck for that reason. I thought if it's just for me, it would be more ideal.
I have an update.
You also got a Switch?
No, not yet.
Okay, okay.
But the most incredible thing happened.
my daughter noticed this device that appeared in our house that dad plays every now and then
and we were on vacation and I was playing the steam deck and she was like what's that well let me
tell you and I was playing I've been playing this really cool indie old school style RPG called
sea of stars it's like a chrono trigger style super Nintendo style RPG I'm playing it when
my daughter comes up she's like can I watch you play and I'm like hell
Well, yeah, you can watch me play.
I get to play video games and you sit here and snuggle with me.
And, like, you know, amazing.
I get to play video games and call it parenting.
Then it gets even better.
Probably, like, two weeks ago, we're playing.
And she's like, hey, dad, can I try?
I'm like, absolutely you can try.
I hand her the Steam deck.
And it was the most incredible experience,
one of the most incredible experiences I've had as a parent
because she doesn't know how to play video games.
and I'm watching her learn how to, like, use a joystick and hit the button.
Supervised learning.
Yeah, yeah, yeah.
Supervised learning.
I'm telling her what to do.
And then within two or three nights, she got it.
She doesn't even know how to read yet, but she figured it out.
It's awesome.
I'm watching her in real time.
And so now the last week, it's turned to mostly she's playing.
And I'm, like, helping her asking questions of like, well, what do you think you should do here?
Like, you know, should you go here?
I think this is the goal.
I think this is where it's so, so fun.
So I think I might actually pretty soon, her birthday's coming up, end up getting a switch
so that we can play, you know, together on the switch.
Right.
But unintentionally, the Steam Deck was the gateway drug for my soon-to-be four-year-old daughter.
That's awesome.
There you go.
Parent of the year right there, getting to play video games.
Oh, honey, I got it.
I'll take it.
Oh, yeah, I got it.
All right, well, listeners, we have lots of thank yous to make for this.
episode, we talked to so many folks who were instrumental in helping put it together. First,
a thank you to our partners this season. JPMorgan payments, trusted, reliable payments infrastructure
for your business no matter the scale. That's jpmorgon.com slash acquired. Century, the best way
to monitor for issues in your software and fix them before users get mad. That's century.i.o slash
acquired. WorkOS, the best way to make your app enterprise ready, starting with single sign-on
in just a few lines of code, WorkOS.com. And Shopify, the best place to sell online.
whether you're a large enterprise or just a founder with a big idea,
Shopify.com slash Acquired.
The links are all in the show notes.
As always, all of our sources for this episode are linked in the show notes.
Yes.
First, Stephen Levy at Wired and his great classic book on Google in The Plex,
which has been an amazing source for all three of our Google episodes.
Definitely go by the book and read that.
Also to Parmy Olson at Bloomberg for her book, Supremacy,
about Deep Mind and Open AI,
which was a main source for this episode.
And I guess also to Cade Mets, right?
For Genius Makers, yeah.
Yeah.
Great book.
Our research thank yous.
Max Ross, Liz Reed,
Josh Woodward, Greg Carrado,
Sebastian Thrun, Anna Patterson,
Brett Taylor, Clay Beauvoir,
Demis Hasabas, Thomas Currian,
Sundar Pachahi.
A special thank you to Nick Fox,
who is the only person we spoke to
for all three Google episodes for research.
We got the hat trick.
To Arvin Navaratnam at Worley Partners
for his great write-up on Alphabet
linked in the show notes.
To Jonathan Ross, original team member
on the TPU,
and today the founder and CEO of GROC,
that's GROC with a Q,
making chips for inference.
To the Waymo folks,
Dmitri Dogglov and Suzanne Filion,
to Gavin Baker from Atreides Management,
to M.G. Seigler,
writer at Spyglass.
is just one of my favorite technology writers and pundance.
O.G. TechCrunch writer.
That's right.
To Ben Idelson for being a great thought partner on this episode
and his excellent recent episode on the Step Change podcast
on the history of data centers.
I highly recommend it if you haven't listened already.
It's only episode three for them of the entire podcast
and they're already getting, I don't know, 30, 40,000 listens on it.
I mean, this thing is taken off.
Amazing.
Dude, that's way better than we were doing on episode three.
It's way better than we were doing.
And if you like Acquired, you will love the Step Change podcast, and Ben is a dear friend.
So, highly recommend checking it out.
To Corai Kovachaloo from the DeepMind team, building the core Gemini models.
To Shashir Maroda, the CEO of Grammarly, formerly ran product at YouTube.
To Jim Gow, the CEO of Fadra and former DeepMind team member,
Chathin Putagunta, partner at Benchmark, Dwar Cash Patel,
for helping me think through some of my conclusions to draw,
and to Brian Lawrence from Oak Cliff Capital,
for helping me think about the economics of AI data centers.
If you like this episode, go check out our episode
on the early history of Google and the 2010s
with our alphabet episode, and of course,
our series on Microsoft and NVIDIA.
After this episode, go check out ACQ2 with Toby Lutka,
the founder and CEO of Shopify,
and come talk about it with us in the Slack
at Acquire.fm slash Slack.
And don't forget, our 10th anniversary celebration of Acquired.
We are going to do a open Zoom call, an LP call, just like the days of your, with anyone, listeners, come join us on Zoom.
It's going to be on October 20th at 4 p.m. Pacific time. Details are in the show notes.
And with that, listeners, we'll see you next time.
We'll see you next time.
Who got the truth? Is it you? Is it you? Is it you? Who got the truth now?
Thank you.